Schedule

YOU CAN ATTEND DAY-1 OR DAY-2 OR BOTH! AND SELECTED WORKSHOPS. CREDIT CARD PAYMENT AVAILABLE.

Connecting the Dots, Creating Data Centric Business Value (Dutch spoken)

How does one move from a technology centric to an integrated data strategy? And how do we improve the data literacy of the stakeholders and ensure that the data products can be used effectively?
Read more

It’s no longer about convincing management of the value of using data to create business value. The real question has now shifted to ensuring that the value is delivered in a sustainable way. Too many organizations still fail to actually get value from their data initiatives. What are the key elements that need to be put in place to ensure success? How do you move from a technology-centric to an integrated data strategy? How do we improve the data literacy of the stakeholders and ensure that the data products can be used effectively? With the regularity of the clock, we introduce new concepts such as data fabric and data mesh, where the question remains to what extent they solve problems or introduce new problems.

 You will learn:

  • Which aspects are really  relevant to get value from your data
  • How to increase  the “data literacy” of  your  employees
  • How more accurate data can contribute to better algorithms
  • Why it is important to look not only at internal but also at external data
  • Which technological solutions are needed to set up good data management
  • Which data governance model can lead to the best results.
Read less

The transition to the new pension system - Data Management approach and tooling (Dutch spoken)

With the arrival of te new pension system in The Netherlands, APG is facing a serious challenge: How to convert the pension rights of millions of participants to the new pension scheme?
Read more

With the arrival of te new pension system in The Netherlands, APG is facing a serious challenge: How to convert the pension rights of millions of participants to the new pension scheme? Datamanagement plays an important role in this transition. Arjen Bouman offers a look behind the scenes at this huge operation and shares his experiences and learnings that he has experienced during this process.

Read less

Enterprise Semantic Data Management (Dutch spoken)

Data lineage - the traceability of data to its meaning and the reason for which the data is used - is becoming a critical success factor. Marco Brattinga takes you into the world of enterprise semantic data management as a means to help tackle this important issue.
Read more

The interest in the meaning of data is increasing. Data lineage – the traceability of data to its meaning and the reason for which the data is used – is becoming a critical success factor. Additionally, the increasing variety of data calls for a grip on the individual data sources. The lack of available data specialists makes it necessary to make available knowledge explicit. The introduction of a distributed data architecture provides the final push to “clean up the attic of data”.

The processing of data is therefore not only a logistical challenge, but also requires a reliable approach to map the meaning of data. This approach goes beyond the traditional description of the structure of the data warehouse: a semantic approach is required.

This semantic approach takes the problem space as the starting point for the description: the domain for which data is acquired. An accurate analysis and model of the domain is the basis for a translation to a model of the data itself as it manifests in the solution space. The result can be seen as a knowledge graph: a network of connected (linked) data, including the definition of this data and the lineage to the basis for this data in legislation, compliance guidelines and company definitions.

Such an approach is not only relevant for the data warehouse: the result is an explicit, unambiguous recording of the knowledge about the relevant data in an organization. Marco Brattinga takes you into the world of enterprise semantic data management through the following topics:

  • The relevance of semantics for the data warehouse
  • The knowledge graph: linking data by and with metadata
  • The problem space versus the solution space
  • Semantic modeling and data lineage
  • The importance of an augmented data catalog
  • Best-practices to implement data lineage.
Read less

Responsible Data Science (Dutch spoken)

How do you embed 'responsible' data science and ethics in your projects and applications? Lecturers at Hogeschool Rotterdam, Tanja Ubert en Gabriella Obispa, will share which questions you should ask yourself and offer an approach to responsible data science.
Read more

As data scientists, our impact on the world is growing more significant every day. But what are the concrete steps we can make to become more responsible data scientists? In this session, we introduce you to the realm of responsible data science and how to embed ethics into your technology. Tanja Ubert and Gabriella Obispa will share their vision on how we need to include responsibility into our work with data. What questions do we need to ask, what responsibility do we, as specialists, have to take on collecting, using and implementing data solutions in our world?

  • Why responsible data science?
  • A practical entry point to embedding ethics into your data team and organization
  • Value sensitive design – values in your technology innovations
  • Workshop self-assessment current projects, the data ethics framework: how to start implementing transparency, accountability and fairness: design the first version of a (redefined) guideline to start towards responsible data science
  • Ethical dialogue: share insights in groups – the start of a community.
Read less

Simplified Data Architecture: Data Warehouse Automation with Datavault Builder

In this session Erik Franssen and Petr Beles will show how to design the simplified architecture by focusing on data modelling with Data Vault and automating the data engineering process with Datavault Builder.
Read more

Data architectures are becoming increasingly complex due to the need to serve many purposes: multiple personas, ranging from operational data users to data scientists need to have access to a variety of managed, governed data and demand real-time, self-service reporting and analytics. Applying principles while designing data architectures will help simplify the development and usage of those architectures by developers and end users. We apply the following principles:

  1. Data and the meaning of data are managed separately;
  2. Storage and Compute are implemented and managed separately;
  3. Requirements determine the needed capabilities and technological solutions, in that order;
  4. Automate data engineering as much as possible.

In this session we will show how Connected Data Group and 2150 Datavault Builder work together in designing the simplified architecture by focusing on data modelling with Data Vault and automating the data engineering process with Datavault Builder.

During this session you will learn:

  • Principles of the simplified architecture
    • Simplify complexity
    • Separating data from metadata
    • Eliminating data replication
    • Automating data engineering
  • Data Vault Builder
    • Relevance
    • Overview of the development process
    • Layers
    • USP’s
    • 5 minute demo
  • Q&A

How to use the full benefits of Data Vault? Data Vault is the modeling approach to become agile in Data Warehousing. The Data Vault approach is unbeatable, especially when the technical implementation is abstracted through automation. Datavault Builder has combined its Data Vault driven Data Warehouse approach with a standardized development process that allows to scale and allocate development resources flexible.

Quickly develop your own Data Warehouse. Rely on the visual element of Datavault Builder to facilitate the collaboration between business users and IT for fully accepted and sustainable project outcomes. Immediately lay the foundation for new reports or integrate new sources of data in an agile way. Deliver new requirements and features with fully automated deployment. Agile Data Warehouse development and CI/CD become a reality.

Read less

DataOps - best practices and lessons learned (Dutch spoken)

In practice, DataOps is not as common for data & analytics as DevOps is for software engineering. For the latter, Development and Operations are jointly responsible for developing a system, deploying it and maintaining the system. With the aim of delivering faster, being more agile and creating maximum business value. This is where DataOps is the same as DevOps: the objective is similar. But ‘How’ we do this, differs considerably.
Read more

Having the right data in the right place at the right time with the right quality, is becoming increasingly important for supporting business decisions, optimizing, automating and powering AI models. Just like with software development, you want to deliver new functionalities with premium quality much faster. You don’t want to make new data, new insights, new AI models available to the user every month, but when it is ready for deployment. That is what DataOps can achieve in theory. But in practice one faces serious challenges that make it a lot more difficult to effectuate the DataOps process in an organization. For example, how to deal with development sandboxes and representative test data across systems.

In this session Niels Naglé en Vincent Goris will show what DataOps is and that it is not just DevOps for data. They will discuss the unique challenges, solutions for these challenges and their lessons learned.

  • How does DataOps relate to DevOps and what are the differences?
  • A roadmap to implement DataOps in your organization
  • The effect on your teams and organization
  • The importance of metadata, the data catalog and automation
  • The challenges and practical solutions.
Read less

Data Minimization: A New Design principle for Data Architectures (Dutch spoken)

The amount of data that is being copied each day is enormous and growing. Data minimization means looking for ways to limit or stop this. Less copying offers many advantages for any organisation, but what does it mean for your data architecture and how do you go about it? [Video introduction (Dutch)]
Read more

We’ve all seen studies that showed the enormous amounts of data that are created on this planet every day. However, a large part of this data is not new but copied data. In existing data architectures, such as data warehouses, a lot of copying is taking place. But modern architectures, such as data lakes and data hubs, also rely heavily on copying data. This rampant copying must be reduced. We don’t always think about it, but copying data has many disadvantages, including higher data latency, complex forms of data synchronization, more complex data security and data privacy, higher development and maintenance costs, and degraded data quality. It is time to apply the data minimization principle when designing new data architectures. This means that the aim is to minimize copied data. In other words, users gain more access to original data and move from data-by-delivery to data-on-demand. The latter corresponds to what has happened in the movie industry: from collecting videos at a store to video-on-demand. In short, data minimization means that we are going to ‘Netflix’ our data.

  • The effect of data minimization on data warehouses, data lakes, and data hubs
  • The network becomes the database
  • Use of translytical databases, analytical databases, and data virtualization to apply data minimization
  • Focus on business rules and not on data storage
  • Examples of applying data minimization to existing data architectures.

[Video introduction]

Read less

Cutting Data Fabric and Mesh to Measure

What can the two new kids on the block, data architecture frameworks Data Mesh and Data Fabric offer? Dr. Barry Devlin explains how they relate to existing patterns and when should you consider these frameworks. [Video introduction]
Read more

The data warehouse is over thirty years old. The data lake just turned ten. So, is it time for something new? In fact, two new patterns have recently emerged—data fabric and data mesh—promising to revolutionise the delivery of BI and analytics.

Data fabric focuses on the automation of data delivery and discovery using artificial intelligence and active metadata. Data mesh has a very novel take on today’s problems, suggesting we must take a domain driven approach to development to eliminate centralised bottlenecks. Each approach has its supporters and detractors, but who is right? More importantly, should you be planning to replace your existing systems with one or the other?

In this session, Dr. Barry Devlin will explore what data fabric and mesh are, what they offer, and how they differ. We will compare them to existing patterns, such as data warehouse and data lake, data hub and even data lakehouse, using the Digital Information Systems Architecture (DISA) as a base. This will allow us to clearly see their strengths and weaknesses and understand when and how you might choose to move to one or the other.

What You Will Learn:

  • Why we are seeing new patterns emerge
  • What are data fabric and data mesh and how they differ
  • Why you would want to use them
  • What the roadblocks are to each
  • Under what circumstances would you use them and where would you start.

[Video introduction]

Read less

Remote Data Modelstorming with BEAM: Lessons learnt from 2 years of data modeling training and consulting online

Learn about the use of collaboration platforms and how to benefit from these in remote data modellling and requirements gathering activities.
Read more

In this interactive session Lawrence Corr shares his thoughts and experiences on using visual collaboration platforms such as Miro and MURAL for gathering BI data requirements remotely with BEAM (Business Event Analysis and Modeling) for designing star schemas. Learn how visual thinking, narrative, a simple script with 7Ws and lots of real and digital Post-it ™ notes can get your stakeholders thinking dimensionally and capturing their own data requirements with agility in-person and at a distance.

Attendees will have the opportunity to vote visually on a virtual whiteboard and should have their smartphones ready to send Lawrence some digital notes to play the ‘7W game’ using the Post-it app.

This session will cover:

  • Using BEAM (Business Event Analysis and Modeling) remotely to discover key business activity and define rich dimensional data sets
  • Playing the 7W game as an icebreaker and introduction to BEAM
  • Comparisons of key virtual whiteboard features in Miro, MURAL, InVision Freehand and LucidSpark
  • Hybrid modelstorming – starting in-person with real Post-its, capturing work digitally and completing in the cloud
  • BEAM modelstorming templates available in Miro and MURAL which you can using straight away
Read less

Driving measurable value in Established Industries with Traditional Machine Learning

There is no one size fits all when it comes to machine learing, supervised or unsupervised and deep learning technologies. But what are the use cases for these, so when do you use traditional machine learning and when do you apply the latest techniques? [Video introduction]
Read more

Developing a machine learning strategy designed to maximize business value in the age of Deep Learning

Deep Learning is so dominant in some discussions of AI and machine learning that many organizations feel that they need to try to keep up with the latest trends. But does it offer the best path for your organization? What is this technology all about and why should both executives and practitioners understand its history?

All business leaders know that they have to embrace analytics or be left behind. However, technology changes so rapidly that it is difficult to know who to hire, which technologies to embrace, and how to proceed. The truth is that traditional machine learning techniques are a better fit for more organizations than chasing after the latest trends. The hyped techniques are popular for a reason so leaders with a responsibility for analytics need to have a high-level understanding of them.

Learning objectives

  • Learn what makes Deep Learning so powerful and what are its limitations
  • Understand why for many use cases traditional machine learning continues to be a much better option
  • Learn the use cases in established industries where machine learning is driving measurable value
  • Learn the industries and use cases where Deep Learning has made recent revolutionary progress and why
  • Discuss the implications of these approaches for hiring and managing your analytics teams
  • Learn how to maximize the value of your analytics portfolio by choosing the right projects and assigning the ideal resources.

 [Video introduction]

Read less

Building a Business-Driven Roadmap for Modern Cloud Data Architecture

What are the steps you need to take to align the data strategy and roadmap with the business? John O'Brien will take you through a four step approach covering data architecture components, patterns, integration components and also modern data infrastructure on AWS, Azure, and GCP. [Video introduction]
Read more

Companies rely on modern cloud data architectures to transform their organizations into the agile analytics-driven cultures needed to be competitive and resilient. The modern cloud reference architecture applies data architecture principles into cloud platforms with current database and analytics technologies. However, many organizations quickly get in over their head without a carefully prioritized and actionable roadmap aligned with business initiatives and priorities. Building such a roadmap follows a step-by-step process that produces a valuable communication tool for everyone to deliver together.

This session will cover the four significant steps to align the data strategy and roadmap with the business. We’ll start with translating business strategy into data and analytics strategies with the Enterprise Analytics Capabilities Framework. This is followed with a logical modern cloud reference data architecture that can leverage agile architecture techniques for implementation as a modern data infrastructure on any cloud, hybrid or multi-cloud environment. This will provide the basis for drilling deeper into architecture patterns and developing proficiency with DataOps and MLOps.

This session will cover:

  • How to identify and translate business priorities into analytic capabilities
  • How the Enterprise Analytics Capabilities Framework guides architecture roadmaps
  • Modern data architecture components: data lake, DW, data hubs, and sandboxes
  • Modern architecture patterns: polyglot persistence, data lakehouse, data fabric, data mesh
  • Modern integration architecture components: ingestion, data pipelines, event streaming
  • Modern data infrastructure on AWS, Azure, and GCP.

[Video introduction]

Read less

Creating a Predictable and Mature BI Value Stream with Data Automation

How to benefit from Data (warehouse) Automation using WhereScape in a Data Vault 2.0 environment.
Read more

Do you want to generate more value out of your data with less effort and cost?

This presentation will help you to reduce your time to market and increase your development efficiency. Erik discusses projects he has been involved in and explains how he was able to accelerate and streamline them using WhereScape. His main focus will be on a Data Vault 2.0 implementation he was involved in at a large bank.

WhereScape Data Automation software accelerates the design, build, documentation and management of complex data ecosystems. It automates repetitive manual tasks such as hand coding and enables developers can produce architectures in a fraction of the time, without human error.

Read less

Openness and ownership - the balancing act of enterprise data

Governance, compliance and data privacy increasingly seem to conflict with matters like time-to-market, agility and openness. In this session Donald Farmer offers some approaches to this growing challenge. [Video-introduction]
Read more

The role of data in business processes has never been more critical. But as we develop new technologies and new skills it feels like we meet new dilemmas at every turn. Concerns about governance and compliance seem to conflict with demands for agility and collaboration. The expanding scope of the data we work with brings new ethical concerns to light. 

So, are we doomed to a constant struggle for control of our data assets? I don’t think so. In this session, I’ll sketch out a provocative, but hopefully useful idea – that we have confused ownership and accountability, governance and compliance, openness and collaboration. We’ll look at some potentially new approaches, which aim to resolve some of the complex puzzles of enterprise data.

  • Getting to know your enterprise data – do you really know what you have?
  • Why would anyone share enterprise data?
  • Security, privacy, governance, compliance – the essential differences
  • The process of data sharing
  • Catalogs vs Warehouses vs Lakes
  • Roles and responsibilities in data ownership.

[Video-introduction]

Read less

Concept Modelling - An Angst-Free Framework for Engaging your Executives

We all realise the importance of Concept Modelling, but how do you involve executive management in it, and how do you keep them engaged? And do you even need to involve executive management? [Video introduction]
Read more

We have all heard “This is the golden age of data” and “Data is the new oil” but that does not necessarily mean your senior executives are anxious to participate in Conceptual Data Modelling / Concept Modelling. The speaker recently had an interesting exception to the reluctance of senior executives to participate in data modelling. Led by the Chief Strategy Officer, a group of C-level executives and other senior leaders at a mid-size financial institution asked Alec to facilitate three days of Concept Modelling sessions.

Fundamentally, a Concept Model is all about improving communication among various stakeholders, but the communication often gets lost – in the clouds, in the weeds, or somewhere off to the side. This is bad enough in any modelling session, but is completely unacceptable when working at the C-level. Drawing on forty years of successful consulting and modelling experience, this presentation will illustrate core techniques and necessary behaviors to keep even your senior executives involved and engaged,

Key points in the presentation include:

  • What got the executives interested in the first place
  • How we prepared for and structured the sessions
  • How we communicated with the executives before, during, and after the sessions
  • An angst-free framework for developing definitions
  • The evolution of the Concept Model evolved, and the crucial findings
  • The executives’ reaction during the retrospective.

[Video introduction]

Read less

Profiting with Practical Supervised Machine Learning

Practical half day seminar on getting started with Supervised machine learning by Keith McCormick. How do you decide when to use Supervised versus Unsupervised machine learning? How to properly prepare data for different kinds of supervised models and how can data preparation be automated in parallel with the model if deployment is to succeed? [Video introduction]
Read more

Regression, decision trees, neural networks—along with many other supervised learning techniques—provide powerful predictive insights. Once built, the models can produce key indicators to optimize the allocation of organizational resources.
New users of these established techniques are often impressed with how easy it all seems to be. Modeling software to build these models is widely available but often results in disappointing results. Many fail to even recognize that proper problem definition was the problem. They likely conclude that the data was not capable of better performance.
The deployment phase includes proper model interpretation and looking for clues that the model will perform well on unseen data. Although the predictive power of these machine-learning models can be very impressive, there is no benefit unless they inform value-focused actions. Models must be deployed in an automated fashion to continually support decision-making for residual impact. The instructor will show how to interpret supervised models with an eye toward decisioning automation.

The seminar
In this half-day seminar, Keith McCormick will overview the two most important and foundational techniques in supervised machine learning, and explain why 70-80% or more of everyday problems faced in established industries can be addressed with one particular machine learning strategy. The focus will be on highly practical techniques for maximizing your results whether you are brand new to predictive analytics or you’ve made some attempts but have been disappointed in the results so far. Veteran users of these techniques will also benefit because a comparison will be made between these traditional techniques and some features of newer techniques. We will explore that while tempting, the newer techniques are rarely the best fit except in a handful of niche application areas that many organizations will not face (at least not in the short term). Participants will leave with specific ideas to apply to their current and future projects.

Learning Objectives

  • When to apply supervised or unsupervised modeling methods
  • Options for inserting machine learning into the decision making of your organization
  • How to use multiple models for value estimation and classification
  • How to properly prepare data for different kinds of supervised models
  • Interpret model coefficients and output to translate across platforms and languages, including the widely used Predictive Modeling Markup Language (PMML)
  • Explore the pros and cons of “black box” models including ensembles
  • How data preparation must be automated in parallel with the model if deployment is to succeed
  • Compare model accuracy scores to model propensity scores that drive decisions at deployment.

 

Who is it for?

  • Analytic Practitioners
  • Data Scientists
  • IT Professionals
  • Technology Planners
  • Consultants; Business Analysts
  • Analytic Project Leaders.

 

Course Description

1. How to choose the best machine learning strategy

  • How supervised learning compares to other options
  • The reality and the hype regarding machine learning
  • What are the classic traditional machine learning techniques?
  • The two main types of supervised machine learning

2. Decision Trees: Still the best choice for many everyday challenges

  • Exploring and interpreting insights with a completed decision tree model
  • A brief primer on the various types of decision tree algorithms
  • Strategic considerations and advantages of decision trees
  • Deployment and bringing you ML models into production

3. Introducing the CART decision tree

  • CART under the hood
  • Processing various variable types with CART
  • Understanding pruning
  • How CART handles missing data with “surrogates”
  • The “Roshoman effect” in machine learning

4. Additional Supervised Techniques

  • Comparing linear regression to neural networks
  • How to embrace the benefits of neural networks without actually using them
  • Regression trees: how to use decision trees to address regression problems
  • What are “ensemble” methods and why are they so popular?
  • Keeping your solutions practical and transparent

[Video introduction]

Read less

Data and Analytics as a Line of Business

Practical workshop by Donald Farmer on how to leverage on your analytics efforts and developing this into a line of business. [Video introduction]
Read more

By the end of this workshop your team will have a sound understanding of how data and analytics can expand, enhance and strengthen your business and your relationships with clients. You’ll have some practical guidelines for strategy, messaging and design which can get you started on your own analytics journey.

 

Learning objectives

  • The nature of data as asset
  • The importance of analytics to realising that value
  • Ways in which data and analytics can be developed as a line of business
  • Simple models for data and analytics as a line of business
  • The value of benchmarking
  • How to encourage and support internal and external communities
  • How to be strategically more agile when creating data and analytics lines of business
  • Addressing ethics and governance concerns.

 

Course Description

1. Introduction: Data as a resource, analytics as a differentiator

We believe that data without analytics is a wasted resource; analytics without action is a wasted effort. We review the value of data to software companies and the potential for analytics as a new line of business.

 2. Case studies

Real-world examples of software companies who have developed analytic products and services using a gameplan methodology.
 
3. Three simple models to get you started

Although there are many ways in which you can leverage data as a resource and analytics as an offering, we have found three to be relatively easy and effective to start with. We’ll review the components and technologies of each, with some guidelines for success and pitfalls to avoid.

  • User dashboards
  • Predictive analytics and alerting
  • Benchmarking and associative analytics

4. Communities of practice and tools of choice

When you introduce analytics as a line of business, users and their social interactions, whether in the office or online, will be critical to your success. We show how communities of practice develop around the tools we choose – and we describe how to ensure your tool is chosen.

 5. Governance and privacy

In any discussion of data and analytics today, concerns about privacy and compliance always come to the surface. We’ll introduce the subject with enough detail for you take the first, important, practical steps to being well governed for today’s regulatory environment.

 6. Narratives and gameplans

These are simple tools for mapping and aligning strategy. However, although simple, they offer subtle and effective capabilities for planning features and releases and for aligning teams such as marketing and management around a vision.

 

Who’s it for? 

  • CIOs, CTOs, analytics leaders and data management leaders
  • Data scientists, and data analysts.

[Video introduction]

Read less

The Data-Process Connection

Alec Sharp illustrates the many ways concept models (conceptual data models) support business process change and business analysis. [Video introduction]
Read more

Whether you call it a conceptual data model, a domain map, a business object model, or even a “thing model,” a concept model is invaluable to process and architecture initiatives. Why? Because processes, capabilities, and solutions act on “things” – Settle Claim, Register Unit, Resolve Service Issue, and so on. Those things are usually “entities” or “objects” in the concept model, and clarity on “what is one of these things?” contributes immensely to clarity on what the corresponding processes are.
After introducing methods to get people, even C-level executives, engaged in concept modelling, we’ll introduce and get practice with guidelines to ensure proper naming and definition of entities/concepts/business objects. We’ll also see that success depends on recognising that a concept model is a description of a business, not a description of a database. Another key – don’t call it a data model!
Drawing on almost forty years of successful modelling, on projects of every size and type, this session introduces proven techniques backed up with current, real-life examples.

Topics include:

  • Concept modelling essentials – things, facts about things, and the policies and rules governing things
  • “Guerrilla modelling” – how to get started on concept modelling without anyone realising it
  • Naming conventions and graphic guidelines – ensuring correctness, consistency, and readability
  • Concept models as a starting point for process discovery
  • Practical examples of concept modelling supporting process work, architecture work, and commercial software selection.

[Video introduction]

Read less

DataOps for Better and Faster Analytics

In this half day virtual seminar John O'Brien will explain the benefits of the DataOps methodology to improve on your Analytics. [Video introduction]
Read more

Adopting the DataOps Methodology is helping agile teams deliver data and analytics faster and more manageable in modern data infrastructure and ecosystems. DataOps is critical for companies to become resilient with data and analytics delivery in a volatile and uncertain global business environment. Going beyond DevOps for continuous deployments, DataOps leverages principles from other disciplines to evolve data engineering and management.

Companies need data and analytics more than ever to be agile and competitive in today’s fast-changing environment. DataOps can be an enterprise-wide initiative or an independent agile delivery team working to improve how they deliver data analytics for their customer. Gaining traction takes time and ongoing support.

This seminar will cover:

  • The challenges in current data environments and IT
  • What DataOps is and how it differs from other approaches
  • Which principles and technologies to focus on initially
  • How to adopt DataOps to speed analytics development and delivery
  • How to continuously engineer, deploy, and operationalize data pipelines with automation and monitoring
  • Setting expectations and planning for DataOps maturity.

 

Course Description

1. Understanding why we need to change

  • How business Analytics has changed from diagnostic to predictive
  • How data sources are increasing
  • The impact of data integration on Data Management
  • Changes in IT development methodologies and organizations
  • Supporting new data products
  • How DataOps is emerging as the next era
  • Reviewing the Agile Manifesto
  • Important aspect of DevOps
  • Review statistical process control for DataOps
  • How DataOps can embed Data Quality and Data Governance
  • Defining DataOps and the DataOps Manifesto
  • Comparing DevOps to DataOps

2. Making DataOps Work

The 7 key concepts to focus on for DataOps

  • How Connectors can make a difference
  • How engineered data pipelines will work
  • How “data drift” will impact data work
  • Set up repositories for Data Governance and Data Quality
  • The role of data hubs and MDM
  • How to set up measurements correctly
  • Leveraging DataOps Platform instrumentation

The 2 key processes to focus on for DataOps

  • Components needed to deliver on business ideation
  • Building data and Analytics deliverables with DataOps

3. Managing DataOps: defining Metrics and Maturity Models

  • Defining Metrics for Data and Analytics delivery
  • Key DataOps metrics
  • How to leverage reusability metrics
  • Reviewing metrics for process improvement
  • Maturity stage of DataOps adoption
  • CMMI-based Maturity Model
  • IBM Maturity Model.

[Video introduction]

Read less

Profiting with Practical Supervised Machine Learning

Practical half day seminar on getting started with Supervised machine learning by Keith McCormick. How do you decide when to use Supervised versus Unsupervised machine learning? How to properly prepare data for different kinds of supervised models and how can data preparation be automated in parallel with the model if deployment is to succeed? [Video introduction]
Read more

Regression, decision trees, neural networks—along with many other supervised learning techniques—provide powerful predictive insights. Once built, the models can produce key indicators to optimize the allocation of organizational resources.
New users of these established techniques are often impressed with how easy it all seems to be. Modeling software to build these models is widely available but often results in disappointing results. Many fail to even recognize that proper problem definition was the problem. They likely conclude that the data was not capable of better performance.
The deployment phase includes proper model interpretation and looking for clues that the model will perform well on unseen data. Although the predictive power of these machine-learning models can be very impressive, there is no benefit unless they inform value-focused actions. Models must be deployed in an automated fashion to continually support decision-making for residual impact. The instructor will show how to interpret supervised models with an eye toward decisioning automation.

The seminar
In this half-day seminar, Keith McCormick will overview the two most important and foundational techniques in supervised machine learning, and explain why 70-80% or more of everyday problems faced in established industries can be addressed with one particular machine learning strategy. The focus will be on highly practical techniques for maximizing your results whether you are brand new to predictive analytics or you’ve made some attempts but have been disappointed in the results so far. Veteran users of these techniques will also benefit because a comparison will be made between these traditional techniques and some features of newer techniques. We will explore that while tempting, the newer techniques are rarely the best fit except in a handful of niche application areas that many organizations will not face (at least not in the short term). Participants will leave with specific ideas to apply to their current and future projects.

Learning Objectives

  • When to apply supervised or unsupervised modeling methods
  • Options for inserting machine learning into the decision making of your organization
  • How to use multiple models for value estimation and classification
  • How to properly prepare data for different kinds of supervised models
  • Interpret model coefficients and output to translate across platforms and languages, including the widely used Predictive Modeling Markup Language (PMML)
  • Explore the pros and cons of “black box” models including ensembles
  • How data preparation must be automated in parallel with the model if deployment is to succeed
  • Compare model accuracy scores to model propensity scores that drive decisions at deployment.

 

Who is it for?

  • Analytic Practitioners
  • Data Scientists
  • IT Professionals
  • Technology Planners
  • Consultants; Business Analysts
  • Analytic Project Leaders.

 

Course Description

1. How to choose the best machine learning strategy

  • How supervised learning compares to other options
  • The reality and the hype regarding machine learning
  • What are the classic traditional machine learning techniques?
  • The two main types of supervised machine learning

2. Decision Trees: Still the best choice for many everyday challenges

  • Exploring and interpreting insights with a completed decision tree model
  • A brief primer on the various types of decision tree algorithms
  • Strategic considerations and advantages of decision trees
  • Deployment and bringing you ML models into production

3. Introducing the CART decision tree

  • CART under the hood
  • Processing various variable types with CART
  • Understanding pruning
  • How CART handles missing data with “surrogates”
  • The “Roshoman effect” in machine learning

4. Additional Supervised Techniques

  • Comparing linear regression to neural networks
  • How to embrace the benefits of neural networks without actually using them
  • Regression trees: how to use decision trees to address regression problems
  • What are “ensemble” methods and why are they so popular?
  • Keeping your solutions practical and transparent

[Video introduction]

Read less

Data and Analytics as a Line of Business

Practical workshop by Donald Farmer on how to leverage on your analytics efforts and developing this into a line of business. [Video introduction]
Read more

By the end of this workshop your team will have a sound understanding of how data and analytics can expand, enhance and strengthen your business and your relationships with clients. You’ll have some practical guidelines for strategy, messaging and design which can get you started on your own analytics journey.

 

Learning objectives

  • The nature of data as asset
  • The importance of analytics to realising that value
  • Ways in which data and analytics can be developed as a line of business
  • Simple models for data and analytics as a line of business
  • The value of benchmarking
  • How to encourage and support internal and external communities
  • How to be strategically more agile when creating data and analytics lines of business
  • Addressing ethics and governance concerns.

 

Course Description

1. Introduction: Data as a resource, analytics as a differentiator

We believe that data without analytics is a wasted resource; analytics without action is a wasted effort. We review the value of data to software companies and the potential for analytics as a new line of business.

 2. Case studies

Real-world examples of software companies who have developed analytic products and services using a gameplan methodology.
 
3. Three simple models to get you started

Although there are many ways in which you can leverage data as a resource and analytics as an offering, we have found three to be relatively easy and effective to start with. We’ll review the components and technologies of each, with some guidelines for success and pitfalls to avoid.

  • User dashboards
  • Predictive analytics and alerting
  • Benchmarking and associative analytics

4. Communities of practice and tools of choice

When you introduce analytics as a line of business, users and their social interactions, whether in the office or online, will be critical to your success. We show how communities of practice develop around the tools we choose – and we describe how to ensure your tool is chosen.

 5. Governance and privacy

In any discussion of data and analytics today, concerns about privacy and compliance always come to the surface. We’ll introduce the subject with enough detail for you take the first, important, practical steps to being well governed for today’s regulatory environment.

 6. Narratives and gameplans

These are simple tools for mapping and aligning strategy. However, although simple, they offer subtle and effective capabilities for planning features and releases and for aligning teams such as marketing and management around a vision.

 

Who’s it for? 

  • CIOs, CTOs, analytics leaders and data management leaders
  • Data scientists, and data analysts.

[Video introduction]

Read less

The Data-Process Connection

Alec Sharp illustrates the many ways concept models (conceptual data models) support business process change and business analysis. [Video introduction]
Read more

Whether you call it a conceptual data model, a domain map, a business object model, or even a “thing model,” a concept model is invaluable to process and architecture initiatives. Why? Because processes, capabilities, and solutions act on “things” – Settle Claim, Register Unit, Resolve Service Issue, and so on. Those things are usually “entities” or “objects” in the concept model, and clarity on “what is one of these things?” contributes immensely to clarity on what the corresponding processes are.
After introducing methods to get people, even C-level executives, engaged in concept modelling, we’ll introduce and get practice with guidelines to ensure proper naming and definition of entities/concepts/business objects. We’ll also see that success depends on recognising that a concept model is a description of a business, not a description of a database. Another key – don’t call it a data model!
Drawing on almost forty years of successful modelling, on projects of every size and type, this session introduces proven techniques backed up with current, real-life examples.

Topics include:

  • Concept modelling essentials – things, facts about things, and the policies and rules governing things
  • “Guerrilla modelling” – how to get started on concept modelling without anyone realising it
  • Naming conventions and graphic guidelines – ensuring correctness, consistency, and readability
  • Concept models as a starting point for process discovery
  • Practical examples of concept modelling supporting process work, architecture work, and commercial software selection.

[Video introduction]

Read less

DataOps for Better and Faster Analytics

In this half day virtual seminar John O'Brien will explain the benefits of the DataOps methodology to improve on your Analytics. [Video introduction]
Read more

Adopting the DataOps Methodology is helping agile teams deliver data and analytics faster and more manageable in modern data infrastructure and ecosystems. DataOps is critical for companies to become resilient with data and analytics delivery in a volatile and uncertain global business environment. Going beyond DevOps for continuous deployments, DataOps leverages principles from other disciplines to evolve data engineering and management.

Companies need data and analytics more than ever to be agile and competitive in today’s fast-changing environment. DataOps can be an enterprise-wide initiative or an independent agile delivery team working to improve how they deliver data analytics for their customer. Gaining traction takes time and ongoing support.

This seminar will cover:

  • The challenges in current data environments and IT
  • What DataOps is and how it differs from other approaches
  • Which principles and technologies to focus on initially
  • How to adopt DataOps to speed analytics development and delivery
  • How to continuously engineer, deploy, and operationalize data pipelines with automation and monitoring
  • Setting expectations and planning for DataOps maturity.

 

Course Description

1. Understanding why we need to change

  • How business Analytics has changed from diagnostic to predictive
  • How data sources are increasing
  • The impact of data integration on Data Management
  • Changes in IT development methodologies and organizations
  • Supporting new data products
  • How DataOps is emerging as the next era
  • Reviewing the Agile Manifesto
  • Important aspect of DevOps
  • Review statistical process control for DataOps
  • How DataOps can embed Data Quality and Data Governance
  • Defining DataOps and the DataOps Manifesto
  • Comparing DevOps to DataOps

2. Making DataOps Work

The 7 key concepts to focus on for DataOps

  • How Connectors can make a difference
  • How engineered data pipelines will work
  • How “data drift” will impact data work
  • Set up repositories for Data Governance and Data Quality
  • The role of data hubs and MDM
  • How to set up measurements correctly
  • Leveraging DataOps Platform instrumentation

The 2 key processes to focus on for DataOps

  • Components needed to deliver on business ideation
  • Building data and Analytics deliverables with DataOps

3. Managing DataOps: defining Metrics and Maturity Models

  • Defining Metrics for Data and Analytics delivery
  • Key DataOps metrics
  • How to leverage reusability metrics
  • Reviewing metrics for process improvement
  • Maturity stage of DataOps adoption
  • CMMI-based Maturity Model
  • IBM Maturity Model.

[Video introduction]

Read less

Profiting with Practical Supervised Machine Learning

Practical half day seminar on getting started with Supervised machine learning by Keith McCormick. How do you decide when to use Supervised versus Unsupervised machine learning? How to properly prepare data for different kinds of supervised models and how can data preparation be automated in parallel with the model if deployment is to succeed? [Video introduction]
Read more

Regression, decision trees, neural networks—along with many other supervised learning techniques—provide powerful predictive insights. Once built, the models can produce key indicators to optimize the allocation of organizational resources.
New users of these established techniques are often impressed with how easy it all seems to be. Modeling software to build these models is widely available but often results in disappointing results. Many fail to even recognize that proper problem definition was the problem. They likely conclude that the data was not capable of better performance.
The deployment phase includes proper model interpretation and looking for clues that the model will perform well on unseen data. Although the predictive power of these machine-learning models can be very impressive, there is no benefit unless they inform value-focused actions. Models must be deployed in an automated fashion to continually support decision-making for residual impact. The instructor will show how to interpret supervised models with an eye toward decisioning automation.

The seminar
In this half-day seminar, Keith McCormick will overview the two most important and foundational techniques in supervised machine learning, and explain why 70-80% or more of everyday problems faced in established industries can be addressed with one particular machine learning strategy. The focus will be on highly practical techniques for maximizing your results whether you are brand new to predictive analytics or you’ve made some attempts but have been disappointed in the results so far. Veteran users of these techniques will also benefit because a comparison will be made between these traditional techniques and some features of newer techniques. We will explore that while tempting, the newer techniques are rarely the best fit except in a handful of niche application areas that many organizations will not face (at least not in the short term). Participants will leave with specific ideas to apply to their current and future projects.

Learning Objectives

  • When to apply supervised or unsupervised modeling methods
  • Options for inserting machine learning into the decision making of your organization
  • How to use multiple models for value estimation and classification
  • How to properly prepare data for different kinds of supervised models
  • Interpret model coefficients and output to translate across platforms and languages, including the widely used Predictive Modeling Markup Language (PMML)
  • Explore the pros and cons of “black box” models including ensembles
  • How data preparation must be automated in parallel with the model if deployment is to succeed
  • Compare model accuracy scores to model propensity scores that drive decisions at deployment.

 

Who is it for?

  • Analytic Practitioners
  • Data Scientists
  • IT Professionals
  • Technology Planners
  • Consultants; Business Analysts
  • Analytic Project Leaders.

 

Course Description

1. How to choose the best machine learning strategy

  • How supervised learning compares to other options
  • The reality and the hype regarding machine learning
  • What are the classic traditional machine learning techniques?
  • The two main types of supervised machine learning

2. Decision Trees: Still the best choice for many everyday challenges

  • Exploring and interpreting insights with a completed decision tree model
  • A brief primer on the various types of decision tree algorithms
  • Strategic considerations and advantages of decision trees
  • Deployment and bringing you ML models into production

3. Introducing the CART decision tree

  • CART under the hood
  • Processing various variable types with CART
  • Understanding pruning
  • How CART handles missing data with “surrogates”
  • The “Roshoman effect” in machine learning

4. Additional Supervised Techniques

  • Comparing linear regression to neural networks
  • How to embrace the benefits of neural networks without actually using them
  • Regression trees: how to use decision trees to address regression problems
  • What are “ensemble” methods and why are they so popular?
  • Keeping your solutions practical and transparent

[Video introduction]

Read less

Data and Analytics as a Line of Business

Practical workshop by Donald Farmer on how to leverage on your analytics efforts and developing this into a line of business. [Video introduction]
Read more

By the end of this workshop your team will have a sound understanding of how data and analytics can expand, enhance and strengthen your business and your relationships with clients. You’ll have some practical guidelines for strategy, messaging and design which can get you started on your own analytics journey.

 

Learning objectives

  • The nature of data as asset
  • The importance of analytics to realising that value
  • Ways in which data and analytics can be developed as a line of business
  • Simple models for data and analytics as a line of business
  • The value of benchmarking
  • How to encourage and support internal and external communities
  • How to be strategically more agile when creating data and analytics lines of business
  • Addressing ethics and governance concerns.

 

Course Description

1. Introduction: Data as a resource, analytics as a differentiator

We believe that data without analytics is a wasted resource; analytics without action is a wasted effort. We review the value of data to software companies and the potential for analytics as a new line of business.

 2. Case studies

Real-world examples of software companies who have developed analytic products and services using a gameplan methodology.
 
3. Three simple models to get you started

Although there are many ways in which you can leverage data as a resource and analytics as an offering, we have found three to be relatively easy and effective to start with. We’ll review the components and technologies of each, with some guidelines for success and pitfalls to avoid.

  • User dashboards
  • Predictive analytics and alerting
  • Benchmarking and associative analytics

4. Communities of practice and tools of choice

When you introduce analytics as a line of business, users and their social interactions, whether in the office or online, will be critical to your success. We show how communities of practice develop around the tools we choose – and we describe how to ensure your tool is chosen.

 5. Governance and privacy

In any discussion of data and analytics today, concerns about privacy and compliance always come to the surface. We’ll introduce the subject with enough detail for you take the first, important, practical steps to being well governed for today’s regulatory environment.

 6. Narratives and gameplans

These are simple tools for mapping and aligning strategy. However, although simple, they offer subtle and effective capabilities for planning features and releases and for aligning teams such as marketing and management around a vision.

 

Who’s it for? 

  • CIOs, CTOs, analytics leaders and data management leaders
  • Data scientists, and data analysts.

[Video introduction]

Read less

The Data-Process Connection

Alec Sharp illustrates the many ways concept models (conceptual data models) support business process change and business analysis. [Video introduction]
Read more

Whether you call it a conceptual data model, a domain map, a business object model, or even a “thing model,” a concept model is invaluable to process and architecture initiatives. Why? Because processes, capabilities, and solutions act on “things” – Settle Claim, Register Unit, Resolve Service Issue, and so on. Those things are usually “entities” or “objects” in the concept model, and clarity on “what is one of these things?” contributes immensely to clarity on what the corresponding processes are.
After introducing methods to get people, even C-level executives, engaged in concept modelling, we’ll introduce and get practice with guidelines to ensure proper naming and definition of entities/concepts/business objects. We’ll also see that success depends on recognising that a concept model is a description of a business, not a description of a database. Another key – don’t call it a data model!
Drawing on almost forty years of successful modelling, on projects of every size and type, this session introduces proven techniques backed up with current, real-life examples.

Topics include:

  • Concept modelling essentials – things, facts about things, and the policies and rules governing things
  • “Guerrilla modelling” – how to get started on concept modelling without anyone realising it
  • Naming conventions and graphic guidelines – ensuring correctness, consistency, and readability
  • Concept models as a starting point for process discovery
  • Practical examples of concept modelling supporting process work, architecture work, and commercial software selection.

[Video introduction]

Read less

DataOps for Better and Faster Analytics

In this half day virtual seminar John O'Brien will explain the benefits of the DataOps methodology to improve on your Analytics. [Video introduction]
Read more

Adopting the DataOps Methodology is helping agile teams deliver data and analytics faster and more manageable in modern data infrastructure and ecosystems. DataOps is critical for companies to become resilient with data and analytics delivery in a volatile and uncertain global business environment. Going beyond DevOps for continuous deployments, DataOps leverages principles from other disciplines to evolve data engineering and management.

Companies need data and analytics more than ever to be agile and competitive in today’s fast-changing environment. DataOps can be an enterprise-wide initiative or an independent agile delivery team working to improve how they deliver data analytics for their customer. Gaining traction takes time and ongoing support.

This seminar will cover:

  • The challenges in current data environments and IT
  • What DataOps is and how it differs from other approaches
  • Which principles and technologies to focus on initially
  • How to adopt DataOps to speed analytics development and delivery
  • How to continuously engineer, deploy, and operationalize data pipelines with automation and monitoring
  • Setting expectations and planning for DataOps maturity.

 

Course Description

1. Understanding why we need to change

  • How business Analytics has changed from diagnostic to predictive
  • How data sources are increasing
  • The impact of data integration on Data Management
  • Changes in IT development methodologies and organizations
  • Supporting new data products
  • How DataOps is emerging as the next era
  • Reviewing the Agile Manifesto
  • Important aspect of DevOps
  • Review statistical process control for DataOps
  • How DataOps can embed Data Quality and Data Governance
  • Defining DataOps and the DataOps Manifesto
  • Comparing DevOps to DataOps

2. Making DataOps Work

The 7 key concepts to focus on for DataOps

  • How Connectors can make a difference
  • How engineered data pipelines will work
  • How “data drift” will impact data work
  • Set up repositories for Data Governance and Data Quality
  • The role of data hubs and MDM
  • How to set up measurements correctly
  • Leveraging DataOps Platform instrumentation

The 2 key processes to focus on for DataOps

  • Components needed to deliver on business ideation
  • Building data and Analytics deliverables with DataOps

3. Managing DataOps: defining Metrics and Maturity Models

  • Defining Metrics for Data and Analytics delivery
  • Key DataOps metrics
  • How to leverage reusability metrics
  • Reviewing metrics for process improvement
  • Maturity stage of DataOps adoption
  • CMMI-based Maturity Model
  • IBM Maturity Model.

[Video introduction]

Read less

Profiting with Practical Supervised Machine Learning

Practical half day seminar on getting started with Supervised machine learning by Keith McCormick. How do you decide when to use Supervised versus Unsupervised machine learning? How to properly prepare data for different kinds of supervised models and how can data preparation be automated in parallel with the model if deployment is to succeed? [Video introduction]
Read more

Regression, decision trees, neural networks—along with many other supervised learning techniques—provide powerful predictive insights. Once built, the models can produce key indicators to optimize the allocation of organizational resources.
New users of these established techniques are often impressed with how easy it all seems to be. Modeling software to build these models is widely available but often results in disappointing results. Many fail to even recognize that proper problem definition was the problem. They likely conclude that the data was not capable of better performance.
The deployment phase includes proper model interpretation and looking for clues that the model will perform well on unseen data. Although the predictive power of these machine-learning models can be very impressive, there is no benefit unless they inform value-focused actions. Models must be deployed in an automated fashion to continually support decision-making for residual impact. The instructor will show how to interpret supervised models with an eye toward decisioning automation.

The seminar
In this half-day seminar, Keith McCormick will overview the two most important and foundational techniques in supervised machine learning, and explain why 70-80% or more of everyday problems faced in established industries can be addressed with one particular machine learning strategy. The focus will be on highly practical techniques for maximizing your results whether you are brand new to predictive analytics or you’ve made some attempts but have been disappointed in the results so far. Veteran users of these techniques will also benefit because a comparison will be made between these traditional techniques and some features of newer techniques. We will explore that while tempting, the newer techniques are rarely the best fit except in a handful of niche application areas that many organizations will not face (at least not in the short term). Participants will leave with specific ideas to apply to their current and future projects.

Learning Objectives

  • When to apply supervised or unsupervised modeling methods
  • Options for inserting machine learning into the decision making of your organization
  • How to use multiple models for value estimation and classification
  • How to properly prepare data for different kinds of supervised models
  • Interpret model coefficients and output to translate across platforms and languages, including the widely used Predictive Modeling Markup Language (PMML)
  • Explore the pros and cons of “black box” models including ensembles
  • How data preparation must be automated in parallel with the model if deployment is to succeed
  • Compare model accuracy scores to model propensity scores that drive decisions at deployment.

 

Who is it for?

  • Analytic Practitioners
  • Data Scientists
  • IT Professionals
  • Technology Planners
  • Consultants; Business Analysts
  • Analytic Project Leaders.

 

Course Description

1. How to choose the best machine learning strategy

  • How supervised learning compares to other options
  • The reality and the hype regarding machine learning
  • What are the classic traditional machine learning techniques?
  • The two main types of supervised machine learning

2. Decision Trees: Still the best choice for many everyday challenges

  • Exploring and interpreting insights with a completed decision tree model
  • A brief primer on the various types of decision tree algorithms
  • Strategic considerations and advantages of decision trees
  • Deployment and bringing you ML models into production

3. Introducing the CART decision tree

  • CART under the hood
  • Processing various variable types with CART
  • Understanding pruning
  • How CART handles missing data with “surrogates”
  • The “Roshoman effect” in machine learning

4. Additional Supervised Techniques

  • Comparing linear regression to neural networks
  • How to embrace the benefits of neural networks without actually using them
  • Regression trees: how to use decision trees to address regression problems
  • What are “ensemble” methods and why are they so popular?
  • Keeping your solutions practical and transparent

[Video introduction]

Read less

Data and Analytics as a Line of Business

Practical workshop by Donald Farmer on how to leverage on your analytics efforts and developing this into a line of business. [Video introduction]
Read more

By the end of this workshop your team will have a sound understanding of how data and analytics can expand, enhance and strengthen your business and your relationships with clients. You’ll have some practical guidelines for strategy, messaging and design which can get you started on your own analytics journey.

 

Learning objectives

  • The nature of data as asset
  • The importance of analytics to realising that value
  • Ways in which data and analytics can be developed as a line of business
  • Simple models for data and analytics as a line of business
  • The value of benchmarking
  • How to encourage and support internal and external communities
  • How to be strategically more agile when creating data and analytics lines of business
  • Addressing ethics and governance concerns.

 

Course Description

1. Introduction: Data as a resource, analytics as a differentiator

We believe that data without analytics is a wasted resource; analytics without action is a wasted effort. We review the value of data to software companies and the potential for analytics as a new line of business.

 2. Case studies

Real-world examples of software companies who have developed analytic products and services using a gameplan methodology.
 
3. Three simple models to get you started

Although there are many ways in which you can leverage data as a resource and analytics as an offering, we have found three to be relatively easy and effective to start with. We’ll review the components and technologies of each, with some guidelines for success and pitfalls to avoid.

  • User dashboards
  • Predictive analytics and alerting
  • Benchmarking and associative analytics

4. Communities of practice and tools of choice

When you introduce analytics as a line of business, users and their social interactions, whether in the office or online, will be critical to your success. We show how communities of practice develop around the tools we choose – and we describe how to ensure your tool is chosen.

 5. Governance and privacy

In any discussion of data and analytics today, concerns about privacy and compliance always come to the surface. We’ll introduce the subject with enough detail for you take the first, important, practical steps to being well governed for today’s regulatory environment.

 6. Narratives and gameplans

These are simple tools for mapping and aligning strategy. However, although simple, they offer subtle and effective capabilities for planning features and releases and for aligning teams such as marketing and management around a vision.

 

Who’s it for? 

  • CIOs, CTOs, analytics leaders and data management leaders
  • Data scientists, and data analysts.

[Video introduction]

Read less

The Data-Process Connection

Alec Sharp illustrates the many ways concept models (conceptual data models) support business process change and business analysis. [Video introduction]
Read more

Whether you call it a conceptual data model, a domain map, a business object model, or even a “thing model,” a concept model is invaluable to process and architecture initiatives. Why? Because processes, capabilities, and solutions act on “things” – Settle Claim, Register Unit, Resolve Service Issue, and so on. Those things are usually “entities” or “objects” in the concept model, and clarity on “what is one of these things?” contributes immensely to clarity on what the corresponding processes are.
After introducing methods to get people, even C-level executives, engaged in concept modelling, we’ll introduce and get practice with guidelines to ensure proper naming and definition of entities/concepts/business objects. We’ll also see that success depends on recognising that a concept model is a description of a business, not a description of a database. Another key – don’t call it a data model!
Drawing on almost forty years of successful modelling, on projects of every size and type, this session introduces proven techniques backed up with current, real-life examples.

Topics include:

  • Concept modelling essentials – things, facts about things, and the policies and rules governing things
  • “Guerrilla modelling” – how to get started on concept modelling without anyone realising it
  • Naming conventions and graphic guidelines – ensuring correctness, consistency, and readability
  • Concept models as a starting point for process discovery
  • Practical examples of concept modelling supporting process work, architecture work, and commercial software selection.

[Video introduction]

Read less

DataOps for Better and Faster Analytics

In this half day virtual seminar John O'Brien will explain the benefits of the DataOps methodology to improve on your Analytics. [Video introduction]
Read more

Adopting the DataOps Methodology is helping agile teams deliver data and analytics faster and more manageable in modern data infrastructure and ecosystems. DataOps is critical for companies to become resilient with data and analytics delivery in a volatile and uncertain global business environment. Going beyond DevOps for continuous deployments, DataOps leverages principles from other disciplines to evolve data engineering and management.

Companies need data and analytics more than ever to be agile and competitive in today’s fast-changing environment. DataOps can be an enterprise-wide initiative or an independent agile delivery team working to improve how they deliver data analytics for their customer. Gaining traction takes time and ongoing support.

This seminar will cover:

  • The challenges in current data environments and IT
  • What DataOps is and how it differs from other approaches
  • Which principles and technologies to focus on initially
  • How to adopt DataOps to speed analytics development and delivery
  • How to continuously engineer, deploy, and operationalize data pipelines with automation and monitoring
  • Setting expectations and planning for DataOps maturity.

 

Course Description

1. Understanding why we need to change

  • How business Analytics has changed from diagnostic to predictive
  • How data sources are increasing
  • The impact of data integration on Data Management
  • Changes in IT development methodologies and organizations
  • Supporting new data products
  • How DataOps is emerging as the next era
  • Reviewing the Agile Manifesto
  • Important aspect of DevOps
  • Review statistical process control for DataOps
  • How DataOps can embed Data Quality and Data Governance
  • Defining DataOps and the DataOps Manifesto
  • Comparing DevOps to DataOps

2. Making DataOps Work

The 7 key concepts to focus on for DataOps

  • How Connectors can make a difference
  • How engineered data pipelines will work
  • How “data drift” will impact data work
  • Set up repositories for Data Governance and Data Quality
  • The role of data hubs and MDM
  • How to set up measurements correctly
  • Leveraging DataOps Platform instrumentation

The 2 key processes to focus on for DataOps

  • Components needed to deliver on business ideation
  • Building data and Analytics deliverables with DataOps

3. Managing DataOps: defining Metrics and Maturity Models

  • Defining Metrics for Data and Analytics delivery
  • Key DataOps metrics
  • How to leverage reusability metrics
  • Reviewing metrics for process improvement
  • Maturity stage of DataOps adoption
  • CMMI-based Maturity Model
  • IBM Maturity Model.

[Video introduction]

Read less

 

Limited time? Join one day!
Can you only attend one day? It is possible to attend only the first or only the second conference day and of course the full conference. Delegates also gain four months access to the conference recordings of the elected day so there’s no need to miss out on any session.
Payment by credit card is also available. Please mention this in the Comment-field upon registration and find further instructions for credit card payment on our customer service page.

29 march

09:00 - 09:15 | Opening
Plenary, Room 1    Werner Schoots
| Chairman
Plenary, Room 1, Room 2    Dennis van Gelder, Tanja Ubert
09:15 - 10:05 | Connecting the Dots, Creating Data Centric Business Value (Dutch spoken)
Plenary    Jan Henderyckx
10:10 - 11:00 | The transition to the new pension system – Data Management approach and tooling (Dutch spoken)
Plenary    Arjen Bouman
11:20 - 12:20 | Enterprise Semantic Data Management (Dutch spoken)
Plenary    Marco Brattinga
12:20 - 13:20 | Lunch break
Plenary 
13:20 - 14:10 | Responsible Data Science (Dutch spoken)
Plenary    Tanja Ubert, Gabriella Obispa
14:15 - 14:45 | Simplified Data Architecture: Data Warehouse Automation with Datavault Builder
Plenary    Erik Fransen, Petr Beles
15:00 - 15:50 | DataOps – best practices and lessons learned (Dutch spoken)
Plenary    Niels Naglé, Vincent Goris
15:55 - 16:45 | Data Minimization: A New Design principle for Data Architectures (Dutch spoken)
Plenary    Rick van der Lans

30 march

09:30 - 09:45 | Opening – virtual conference day
Plenary    Werner Schoots
09:45 - 10:35 | Cutting Data Fabric and Mesh to Measure
Plenary    Barry Devlin
10:45 - 11:35 | Remote Data Modelstorming with BEAM: Lessons learnt from 2 years of data modeling training and consulting online
Plenary    Lawrence Corr
11:45 - 12:35 | Driving measurable value in Established Industries with Traditional Machine Learning
Plenary    Keith McCormick
12:30 - 13:30 | Lunch break
Plenary, Room 1 
13:30 – 14:20 | Building a Business-Driven Roadmap for Modern Cloud Data Architecture
Plenary    John O’Brien
14:30 - 15:00 | Creating a Predictable and Mature BI Value Stream with Data Automation
Plenary    Erik van der Hoeven
15:10 - 16:00 | Openness and ownership – the balancing act of enterprise data
Plenary    Donald Farmer
16:10 - 17:00 | Concept Modelling – An Angst-Free Framework for Engaging your Executives
Plenary    Alec Sharp

Workshop

13:30 – 17:00 | Profiting with Practical Supervised Machine Learning
11 May    Keith McCormick
13:30 - 17:00 | Data and Analytics as a Line of Business
31 March    Donald Farmer
13:30 - 17:00 | The Data-Process Connection
6 October    Alec Sharp
13:30 - 17:00 | DataOps for Better and Faster Analytics
7 April    John O’Brien