Adept Events
  • nl
  • Home
  • Conference Outline
  • Speakers
  • Location
  • Sponsors
  • About
    • About Adept Events
    • Photos
    • Contact
    • Other editions
  • nl

Data Warehousing & Business Intelligence Summit 2017

Date Price Contact
March 28 and 29, 2017 € 1.340,- seminars@adeptevents.nl
+31 (0)172 742680
Time Location
9:30 - 17:00 Mercure Hotel Amsterdam City
Next EditionTYPE
March 2018
Date Price
March 28 and 29, 2017 € 1.340,-
Time
9:30 - 17:00
Location Contact
Mercure Hotel Amsterdam City seminars@adeptevents.nl
+31 (0)172 742680
Next Edition
March 2018
TYPE
Date
March 28 and 29, 2017
Price
€ 1.340,-
Time
9:30 - 17:00
Location
Mercure Hotel Amsterdam City
Contact
seminars@adeptevents.nl
+31 (0)172 742680
Next Edition
March 2018
TYPE

Schedule

YOU CAN ATTEND DAY-1 OR DAY-2 OR BOTH! AND SELECTED WORKSHOPS. CREDIT CARD PAYMENT AVAILABLE.

  • Day 1 29 march
  • Day 2 30 march
  • Workshops in March-April
    Werner Schoots
    09:00 - 09:15 | Plenary, Room 1

    Opening

    Live Stream
    Read more
    Read less
    Dennis van Gelder
    | Plenary, Room 1, Room 2

    Chairman

    Read more
    Read less
      Dennis van Gelder | Consultant | Deltiq, Tanja Ubert | Docent and researcher | Hogeschool Rotterdam, Huub Meertens | IT and data expert | Eviden
    Jan Henderyckx
    09:15 - 10:05 | Plenary

    Connecting the Dots, Creating Data Centric Business Value (Dutch spoken)

    Live Stream
    How does one move from a technology centric to an integrated data strategy? And how do we improve the data literacy of the stakeholders and ensure that the data products can be used effectively?
    Read more

    It’s no longer about convincing management of the value of using data to create business value. The real question has now shifted to ensuring that the value is delivered in a sustainable way. Too many organizations still fail to actually get value from their data initiatives. What are the key elements that need to be put in place to ensure success? How do you move from a technology-centric to an integrated data strategy? How do we improve the data literacy of the stakeholders and ensure that the data products can be used effectively? With the regularity of the clock, we introduce new concepts such as data fabric and data mesh, where the question remains to what extent they solve problems or introduce new problems.

     You will learn:

    • Which aspects are really  relevant to get value from your data
    • How to increase  the “data literacy” of  your  employees
    • How more accurate data can contribute to better algorithms
    • Why it is important to look not only at internal but also at external data
    • Which technological solutions are needed to set up good data management
    • Which data governance model can lead to the best results.
    Read less
      Jan Henderyckx | Managing Director | BearingPoint
    Arjen Bouman
    10:10 - 11:00 | Plenary

    The transition to the new pension system - Data Management approach and tooling (Dutch spoken)

    Live Stream
    With the arrival of te new pension system in The Netherlands, APG is facing a serious challenge: How to convert the pension rights of millions of participants to the new pension scheme?
    Read more

    With the arrival of te new pension system in The Netherlands, APG is facing a serious challenge: How to convert the pension rights of millions of participants to the new pension scheme? Datamanagement plays an important role in this transition. Arjen Bouman offers a look behind the scenes at this huge operation and shares his experiences and learnings that he has experienced during this process.

    Read less
      Arjen Bouman | Data Manager | APG
    Marco Brattinga
    11:20 - 12:20 | Plenary

    Enterprise Semantic Data Management (Dutch spoken)

    Live Stream
    Data lineage - the traceability of data to its meaning and the reason for which the data is used - is becoming a critical success factor. Marco Brattinga takes you into the world of enterprise semantic data management as a means to help tackle this important issue.
    Read more

    The interest in the meaning of data is increasing. Data lineage – the traceability of data to its meaning and the reason for which the data is used – is becoming a critical success factor. Additionally, the increasing variety of data calls for a grip on the individual data sources. The lack of available data specialists makes it necessary to make available knowledge explicit. The introduction of a distributed data architecture provides the final push to “clean up the attic of data”.

    The processing of data is therefore not only a logistical challenge, but also requires a reliable approach to map the meaning of data. This approach goes beyond the traditional description of the structure of the data warehouse: a semantic approach is required.

    This semantic approach takes the problem space as the starting point for the description: the domain for which data is acquired. An accurate analysis and model of the domain is the basis for a translation to a model of the data itself as it manifests in the solution space. The result can be seen as a knowledge graph: a network of connected (linked) data, including the definition of this data and the lineage to the basis for this data in legislation, compliance guidelines and company definitions.

    Such an approach is not only relevant for the data warehouse: the result is an explicit, unambiguous recording of the knowledge about the relevant data in an organization. Marco Brattinga takes you into the world of enterprise semantic data management through the following topics:

    • The relevance of semantics for the data warehouse
    • The knowledge graph: linking data by and with metadata
    • The problem space versus the solution space
    • Semantic modeling and data lineage
    • The importance of an augmented data catalog
    • Best-practices to implement data lineage.
    Read less
      Marco Brattinga | Principal Consultant | Ordina
    12:20 - 13:20 | Plenary

    Lunch break

    Read more
    Read less
    Tanja Ubert
    13:20 - 14:10 | Plenary

    Responsible Data Science (Dutch spoken)

    Live Stream
    How do you embed 'responsible' data science and ethics in your projects and applications? Lecturers at Hogeschool Rotterdam, Tanja Ubert en Gabriella Obispa, will share which questions you should ask yourself and offer an approach to responsible data science.
    Read more

    As data scientists, our impact on the world is growing more significant every day. But what are the concrete steps we can make to become more responsible data scientists? In this session, we introduce you to the realm of responsible data science and how to embed ethics into your technology. Tanja Ubert and Gabriella Obispa will share their vision on how we need to include responsibility into our work with data. What questions do we need to ask, what responsibility do we, as specialists, have to take on collecting, using and implementing data solutions in our world?

    • Why responsible data science?
    • A practical entry point to embedding ethics into your data team and organization
    • Value sensitive design – values in your technology innovations
    • Workshop self-assessment current projects, the data ethics framework: how to start implementing transparency, accountability and fairness: design the first version of a (redefined) guideline to start towards responsible data science
    • Ethical dialogue: share insights in groups – the start of a community.
    Read less
      Tanja Ubert | Docent and researcher | Hogeschool Rotterdam, Gabriella Obispa | Lecturer and researcher | Hogeschool Rotterdam
    Erik Fransen
    14:15 - 14:45 | Plenary

    Simplified Data Architecture: Data Warehouse Automation with Datavault Builder

    Live Stream
    In this session Erik Franssen and Petr Beles will show how to design the simplified architecture by focusing on data modelling with Data Vault and automating the data engineering process with Datavault Builder.
    Read more

    Data architectures are becoming increasingly complex due to the need to serve many purposes: multiple personas, ranging from operational data users to data scientists need to have access to a variety of managed, governed data and demand real-time, self-service reporting and analytics. Applying principles while designing data architectures will help simplify the development and usage of those architectures by developers and end users. We apply the following principles:

    1. Data and the meaning of data are managed separately;
    2. Storage and Compute are implemented and managed separately;
    3. Requirements determine the needed capabilities and technological solutions, in that order;
    4. Automate data engineering as much as possible.

    In this session we will show how Connected Data Group and 2150 Datavault Builder work together in designing the simplified architecture by focusing on data modelling with Data Vault and automating the data engineering process with Datavault Builder.

    During this session you will learn:

    • Principles of the simplified architecture
      • Simplify complexity
      • Separating data from metadata
      • Eliminating data replication
      • Automating data engineering
    • Data Vault Builder
      • Relevance
      • Overview of the development process
      • Layers
      • USP’s
      • 5 minute demo
    • Q&A

    How to use the full benefits of Data Vault? Data Vault is the modeling approach to become agile in Data Warehousing. The Data Vault approach is unbeatable, especially when the technical implementation is abstracted through automation. Datavault Builder has combined its Data Vault driven Data Warehouse approach with a standardized development process that allows to scale and allocate development resources flexible.

    Quickly develop your own Data Warehouse. Rely on the visual element of Datavault Builder to facilitate the collaboration between business users and IT for fully accepted and sustainable project outcomes. Immediately lay the foundation for new reports or integrate new sources of data in an agile way. Deliver new requirements and features with fully automated deployment. Agile Data Warehouse development and CI/CD become a reality.

    Read less
      Erik Fransen | Founder | Connected Data Group, Petr Beles | Senior Consultant | 2150 GmbH
    Niels Naglé
    15:00 - 15:50 | Plenary

    DataOps - best practices and lessons learned (Dutch spoken)

    Live Stream
    In practice, DataOps is not as common for data & analytics as DevOps is for software engineering. For the latter, Development and Operations are jointly responsible for developing a system, deploying it and maintaining the system. With the aim of delivering faster, being more agile and creating maximum business value. This is where DataOps is the same as DevOps: the objective is similar. But ‘How’ we do this, differs considerably.
    Read more

    Having the right data in the right place at the right time with the right quality, is becoming increasingly important for supporting business decisions, optimizing, automating and powering AI models. Just like with software development, you want to deliver new functionalities with premium quality much faster. You don’t want to make new data, new insights, new AI models available to the user every month, but when it is ready for deployment. That is what DataOps can achieve in theory. But in practice one faces serious challenges that make it a lot more difficult to effectuate the DataOps process in an organization. For example, how to deal with development sandboxes and representative test data across systems.

    In this session Niels Naglé en Vincent Goris will show what DataOps is and that it is not just DevOps for data. They will discuss the unique challenges, solutions for these challenges and their lessons learned.

    • How does DataOps relate to DevOps and what are the differences?
    • A roadmap to implement DataOps in your organization
    • The effect on your teams and organization
    • The importance of metadata, the data catalog and automation
    • The challenges and practical solutions.
    Read less
      Niels Naglé | Data Solutions Architect | Info Support, Vincent Goris | IT Consultant | Info Support
    Rick van der Lans
    15:55 - 16:45 | Plenary

    Data Minimization: A New Design principle for Data Architectures (Dutch spoken)

    Live Stream
    The amount of data that is being copied each day is enormous and growing. Data minimization means looking for ways to limit or stop this. Less copying offers many advantages for any organisation, but what does it mean for your data architecture and how do you go about it? [Video introduction (Dutch)]
    Read more

    We’ve all seen studies that showed the enormous amounts of data that are created on this planet every day. However, a large part of this data is not new but copied data. In existing data architectures, such as data warehouses, a lot of copying is taking place. But modern architectures, such as data lakes and data hubs, also rely heavily on copying data. This rampant copying must be reduced. We don’t always think about it, but copying data has many disadvantages, including higher data latency, complex forms of data synchronization, more complex data security and data privacy, higher development and maintenance costs, and degraded data quality. It is time to apply the data minimization principle when designing new data architectures. This means that the aim is to minimize copied data. In other words, users gain more access to original data and move from data-by-delivery to data-on-demand. The latter corresponds to what has happened in the movie industry: from collecting videos at a store to video-on-demand. In short, data minimization means that we are going to ‘Netflix’ our data.

    • The effect of data minimization on data warehouses, data lakes, and data hubs
    • The network becomes the database
    • Use of translytical databases, analytical databases, and data virtualization to apply data minimization
    • Focus on business rules and not on data storage
    • Examples of applying data minimization to existing data architectures.

    [Video introduction]

    Read less
      Rick van der Lans | Managing Director | R20/Consultancy
      Werner Schoots
      09:30 - 09:45 | Plenary

      Opening - virtual conference day

      Live Stream
      Read more
      Read less
      Barry Devlin
      09:45 - 10:35 | Plenary

      Cutting Data Fabric and Mesh to Measure

      Live Stream
      What can the two new kids on the block, data architecture frameworks Data Mesh and Data Fabric offer? Dr. Barry Devlin explains how they relate to existing patterns and when should you consider these frameworks. [Video introduction]
      Read more

      The data warehouse is over thirty years old. The data lake just turned ten. So, is it time for something new? In fact, two new patterns have recently emerged—data fabric and data mesh—promising to revolutionise the delivery of BI and analytics.

      Data fabric focuses on the automation of data delivery and discovery using artificial intelligence and active metadata. Data mesh has a very novel take on today’s problems, suggesting we must take a domain driven approach to development to eliminate centralised bottlenecks. Each approach has its supporters and detractors, but who is right? More importantly, should you be planning to replace your existing systems with one or the other?

      In this session, Dr. Barry Devlin will explore what data fabric and mesh are, what they offer, and how they differ. We will compare them to existing patterns, such as data warehouse and data lake, data hub and even data lakehouse, using the Digital Information Systems Architecture (DISA) as a base. This will allow us to clearly see their strengths and weaknesses and understand when and how you might choose to move to one or the other.

      What You Will Learn:

      • Why we are seeing new patterns emerge
      • What are data fabric and data mesh and how they differ
      • Why you would want to use them
      • What the roadblocks are to each
      • Under what circumstances would you use them and where would you start.

      [Video introduction]

      Read less
        Barry Devlin | Founder | 9sight Consulting
      Lawrence Corr
      10:45 - 11:35 | Plenary

      Remote Data Modelstorming with BEAM: Lessons learnt from 2 years of data modeling training and consulting online

      Live Stream
      Learn about the use of collaboration platforms and how to benefit from these in remote data modellling and requirements gathering activities.
      Read more

      In this interactive session Lawrence Corr shares his thoughts and experiences on using visual collaboration platforms such as Miro and MURAL for gathering BI data requirements remotely with BEAM (Business Event Analysis and Modeling) for designing star schemas. Learn how visual thinking, narrative, a simple script with 7Ws and lots of real and digital Post-it ™ notes can get your stakeholders thinking dimensionally and capturing their own data requirements with agility in-person and at a distance.

      Attendees will have the opportunity to vote visually on a virtual whiteboard and should have their smartphones ready to send Lawrence some digital notes to play the ‘7W game’ using the Post-it app.

      This session will cover:

      • Using BEAM (Business Event Analysis and Modeling) remotely to discover key business activity and define rich dimensional data sets
      • Playing the 7W game as an icebreaker and introduction to BEAM
      • Comparisons of key virtual whiteboard features in Miro, MURAL, InVision Freehand and LucidSpark
      • Hybrid modelstorming – starting in-person with real Post-its, capturing work digitally and completing in the cloud
      • BEAM modelstorming templates available in Miro and MURAL which you can using straight away
      Read less
        Lawrence Corr | Chief Data Modelstormer | DecisionOne Consulting
      Keith McCormick
      11:45 - 12:35 | Plenary

      Driving measurable value in Established Industries with Traditional Machine Learning

      Live Stream
      There is no one size fits all when it comes to machine learing, supervised or unsupervised and deep learning technologies. But what are the use cases for these, so when do you use traditional machine learning and when do you apply the latest techniques? [Video introduction]
      Read more

      Developing a machine learning strategy designed to maximize business value in the age of Deep Learning

      Deep Learning is so dominant in some discussions of AI and machine learning that many organizations feel that they need to try to keep up with the latest trends. But does it offer the best path for your organization? What is this technology all about and why should both executives and practitioners understand its history?

      All business leaders know that they have to embrace analytics or be left behind. However, technology changes so rapidly that it is difficult to know who to hire, which technologies to embrace, and how to proceed. The truth is that traditional machine learning techniques are a better fit for more organizations than chasing after the latest trends. The hyped techniques are popular for a reason so leaders with a responsibility for analytics need to have a high-level understanding of them.

      Learning objectives

      • Learn what makes Deep Learning so powerful and what are its limitations
      • Understand why for many use cases traditional machine learning continues to be a much better option
      • Learn the use cases in established industries where machine learning is driving measurable value
      • Learn the industries and use cases where Deep Learning has made recent revolutionary progress and why
      • Discuss the implications of these approaches for hiring and managing your analytics teams
      • Learn how to maximize the value of your analytics portfolio by choosing the right projects and assigning the ideal resources.

       [Video introduction]

      Read less
        Keith McCormick | Data Science Consultant and Instructor |
      12:30 - 13:30 | Plenary

      Lunch break

      Read more
      Read less
      John O’Brien
      13:30 – 14:20 | Plenary

      Building a Business-Driven Roadmap for Modern Cloud Data Architecture

      Live Stream
      What are the steps you need to take to align the data strategy and roadmap with the business? John O'Brien will take you through a four step approach covering data architecture components, patterns, integration components and also modern data infrastructure on AWS, Azure, and GCP. [Video introduction]
      Read more

      Companies rely on modern cloud data architectures to transform their organizations into the agile analytics-driven cultures needed to be competitive and resilient. The modern cloud reference architecture applies data architecture principles into cloud platforms with current database and analytics technologies. However, many organizations quickly get in over their head without a carefully prioritized and actionable roadmap aligned with business initiatives and priorities. Building such a roadmap follows a step-by-step process that produces a valuable communication tool for everyone to deliver together.

      This session will cover the four significant steps to align the data strategy and roadmap with the business. We’ll start with translating business strategy into data and analytics strategies with the Enterprise Analytics Capabilities Framework. This is followed with a logical modern cloud reference data architecture that can leverage agile architecture techniques for implementation as a modern data infrastructure on any cloud, hybrid or multi-cloud environment. This will provide the basis for drilling deeper into architecture patterns and developing proficiency with DataOps and MLOps.

      This session will cover:

      • How to identify and translate business priorities into analytic capabilities
      • How the Enterprise Analytics Capabilities Framework guides architecture roadmaps
      • Modern data architecture components: data lake, DW, data hubs, and sandboxes
      • Modern architecture patterns: polyglot persistence, data lakehouse, data fabric, data mesh
      • Modern integration architecture components: ingestion, data pipelines, event streaming
      • Modern data infrastructure on AWS, Azure, and GCP.

      [Video introduction]

      Read less
        John O’Brien | CEO and Principal Advisor | Radiant Advisors
      Erik van der Hoeven
      14:30 - 15:00 | Plenary

      Creating a Predictable and Mature BI Value Stream with Data Automation

      Live Stream
      How to benefit from Data (warehouse) Automation using WhereScape in a Data Vault 2.0 environment.
      Read more

      Do you want to generate more value out of your data with less effort and cost?

      This presentation will help you to reduce your time to market and increase your development efficiency. Erik discusses projects he has been involved in and explains how he was able to accelerate and streamline them using WhereScape. His main focus will be on a Data Vault 2.0 implementation he was involved in at a large bank.

      WhereScape Data Automation software accelerates the design, build, documentation and management of complex data ecosystems. It automates repetitive manual tasks such as hand coding and enables developers can produce architectures in a fraction of the time, without human error.

      Read less
        Erik van der Hoeven | BI Specialist |
      Donald Farmer
      15:10 - 16:00 | Plenary

      Openness and ownership - the balancing act of enterprise data

      Live Stream
      Governance, compliance and data privacy increasingly seem to conflict with matters like time-to-market, agility and openness. In this session Donald Farmer offers some approaches to this growing challenge. [Video-introduction]
      Read more

      The role of data in business processes has never been more critical. But as we develop new technologies and new skills it feels like we meet new dilemmas at every turn. Concerns about governance and compliance seem to conflict with demands for agility and collaboration. The expanding scope of the data we work with brings new ethical concerns to light. 

      So, are we doomed to a constant struggle for control of our data assets? I don’t think so. In this session, I’ll sketch out a provocative, but hopefully useful idea – that we have confused ownership and accountability, governance and compliance, openness and collaboration. We’ll look at some potentially new approaches, which aim to resolve some of the complex puzzles of enterprise data.

      • Getting to know your enterprise data – do you really know what you have?
      • Why would anyone share enterprise data?
      • Security, privacy, governance, compliance – the essential differences
      • The process of data sharing
      • Catalogs vs Warehouses vs Lakes
      • Roles and responsibilities in data ownership.

      [Video-introduction]

      Read less
        Donald Farmer | Director | TreeHive Strategy
      Alec Sharp
      16:10 - 17:00 | Plenary

      Concept Modelling - An Angst-Free Framework for Engaging your Executives

      Live Stream
      We all realise the importance of Concept Modelling, but how do you involve executive management in it, and how do you keep them engaged? And do you even need to involve executive management? [Video introduction]
      Read more

      We have all heard “This is the golden age of data” and “Data is the new oil” but that does not necessarily mean your senior executives are anxious to participate in Conceptual Data Modelling / Concept Modelling. The speaker recently had an interesting exception to the reluctance of senior executives to participate in data modelling. Led by the Chief Strategy Officer, a group of C-level executives and other senior leaders at a mid-size financial institution asked Alec to facilitate three days of Concept Modelling sessions.

      Fundamentally, a Concept Model is all about improving communication among various stakeholders, but the communication often gets lost – in the clouds, in the weeds, or somewhere off to the side. This is bad enough in any modelling session, but is completely unacceptable when working at the C-level. Drawing on forty years of successful consulting and modelling experience, this presentation will illustrate core techniques and necessary behaviors to keep even your senior executives involved and engaged,

      Key points in the presentation include:

      • What got the executives interested in the first place
      • How we prepared for and structured the sessions
      • How we communicated with the executives before, during, and after the sessions
      • An angst-free framework for developing definitions
      • The evolution of the Concept Model evolved, and the crucial findings
      • The executives’ reaction during the retrospective.

      [Video introduction]

      Read less
        Alec Sharp | Founder | Clariteq Systems Consulting
        Keith McCormick
        13:30 – 17:00 | 11 May

        Profiting with Practical Supervised Machine Learning

        Practical half day seminar on getting started with Supervised machine learning by Keith McCormick. How do you decide when to use Supervised versus Unsupervised machine learning? How to properly prepare data for different kinds of supervised models and how can data preparation be automated in parallel with the model if deployment is to succeed? [Video introduction]
        Read more

        Regression, decision trees, neural networks—along with many other supervised learning techniques—provide powerful predictive insights. Once built, the models can produce key indicators to optimize the allocation of organizational resources.
        New users of these established techniques are often impressed with how easy it all seems to be. Modeling software to build these models is widely available but often results in disappointing results. Many fail to even recognize that proper problem definition was the problem. They likely conclude that the data was not capable of better performance.
        The deployment phase includes proper model interpretation and looking for clues that the model will perform well on unseen data. Although the predictive power of these machine-learning models can be very impressive, there is no benefit unless they inform value-focused actions. Models must be deployed in an automated fashion to continually support decision-making for residual impact. The instructor will show how to interpret supervised models with an eye toward decisioning automation.

        The seminar
        In this half-day seminar, Keith McCormick will overview the two most important and foundational techniques in supervised machine learning, and explain why 70-80% or more of everyday problems faced in established industries can be addressed with one particular machine learning strategy. The focus will be on highly practical techniques for maximizing your results whether you are brand new to predictive analytics or you’ve made some attempts but have been disappointed in the results so far. Veteran users of these techniques will also benefit because a comparison will be made between these traditional techniques and some features of newer techniques. We will explore that while tempting, the newer techniques are rarely the best fit except in a handful of niche application areas that many organizations will not face (at least not in the short term). Participants will leave with specific ideas to apply to their current and future projects.

        Learning Objectives

        • When to apply supervised or unsupervised modeling methods
        • Options for inserting machine learning into the decision making of your organization
        • How to use multiple models for value estimation and classification
        • How to properly prepare data for different kinds of supervised models
        • Interpret model coefficients and output to translate across platforms and languages, including the widely used Predictive Modeling Markup Language (PMML)
        • Explore the pros and cons of “black box” models including ensembles
        • How data preparation must be automated in parallel with the model if deployment is to succeed
        • Compare model accuracy scores to model propensity scores that drive decisions at deployment.

         

        Who is it for?

        • Analytic Practitioners
        • Data Scientists
        • IT Professionals
        • Technology Planners
        • Consultants; Business Analysts
        • Analytic Project Leaders.

         

        Course Description

        1. How to choose the best machine learning strategy

        • How supervised learning compares to other options
        • The reality and the hype regarding machine learning
        • What are the classic traditional machine learning techniques?
        • The two main types of supervised machine learning

        2. Decision Trees: Still the best choice for many everyday challenges

        • Exploring and interpreting insights with a completed decision tree model
        • A brief primer on the various types of decision tree algorithms
        • Strategic considerations and advantages of decision trees
        • Deployment and bringing you ML models into production

        3. Introducing the CART decision tree

        • CART under the hood
        • Processing various variable types with CART
        • Understanding pruning
        • How CART handles missing data with “surrogates”
        • The “Roshoman effect” in machine learning

        4. Additional Supervised Techniques

        • Comparing linear regression to neural networks
        • How to embrace the benefits of neural networks without actually using them
        • Regression trees: how to use decision trees to address regression problems
        • What are “ensemble” methods and why are they so popular?
        • Keeping your solutions practical and transparent

        [Video introduction]

        Read less
          Keith McCormick | Data Science Consultant and Instructor |
        Donald Farmer
        13:30 - 17:00 | 31 March

        Data and Analytics as a Line of Business

        Practical workshop by Donald Farmer on how to leverage on your analytics efforts and developing this into a line of business. [Video introduction]
        Read more

        By the end of this workshop your team will have a sound understanding of how data and analytics can expand, enhance and strengthen your business and your relationships with clients. You’ll have some practical guidelines for strategy, messaging and design which can get you started on your own analytics journey.

         

        Learning objectives

        • The nature of data as asset
        • The importance of analytics to realising that value
        • Ways in which data and analytics can be developed as a line of business
        • Simple models for data and analytics as a line of business
        • The value of benchmarking
        • How to encourage and support internal and external communities
        • How to be strategically more agile when creating data and analytics lines of business
        • Addressing ethics and governance concerns.

         

        Course Description

        1. Introduction: Data as a resource, analytics as a differentiator

        We believe that data without analytics is a wasted resource; analytics without action is a wasted effort. We review the value of data to software companies and the potential for analytics as a new line of business.

         2. Case studies

        Real-world examples of software companies who have developed analytic products and services using a gameplan methodology.
         
        3. Three simple models to get you started

        Although there are many ways in which you can leverage data as a resource and analytics as an offering, we have found three to be relatively easy and effective to start with. We’ll review the components and technologies of each, with some guidelines for success and pitfalls to avoid.

        • User dashboards
        • Predictive analytics and alerting
        • Benchmarking and associative analytics

        4. Communities of practice and tools of choice

        When you introduce analytics as a line of business, users and their social interactions, whether in the office or online, will be critical to your success. We show how communities of practice develop around the tools we choose – and we describe how to ensure your tool is chosen.

         5. Governance and privacy

        In any discussion of data and analytics today, concerns about privacy and compliance always come to the surface. We’ll introduce the subject with enough detail for you take the first, important, practical steps to being well governed for today’s regulatory environment.

         6. Narratives and gameplans

        These are simple tools for mapping and aligning strategy. However, although simple, they offer subtle and effective capabilities for planning features and releases and for aligning teams such as marketing and management around a vision.

         

        Who’s it for? 

        • CIOs, CTOs, analytics leaders and data management leaders
        • Data scientists, and data analysts.

        [Video introduction]

        Read less
          Donald Farmer | Director | TreeHive Strategy
        Alec Sharp
        13:30 - 17:00 | 6 October

        The Data-Process Connection

        Alec Sharp illustrates the many ways concept models (conceptual data models) support business process change and business analysis. [Video introduction]
        Read more

        Whether you call it a conceptual data model, a domain map, a business object model, or even a “thing model,” a concept model is invaluable to process and architecture initiatives. Why? Because processes, capabilities, and solutions act on “things” – Settle Claim, Register Unit, Resolve Service Issue, and so on. Those things are usually “entities” or “objects” in the concept model, and clarity on “what is one of these things?” contributes immensely to clarity on what the corresponding processes are.
        After introducing methods to get people, even C-level executives, engaged in concept modelling, we’ll introduce and get practice with guidelines to ensure proper naming and definition of entities/concepts/business objects. We’ll also see that success depends on recognising that a concept model is a description of a business, not a description of a database. Another key – don’t call it a data model!
        Drawing on almost forty years of successful modelling, on projects of every size and type, this session introduces proven techniques backed up with current, real-life examples.

        Topics include:

        • Concept modelling essentials – things, facts about things, and the policies and rules governing things
        • “Guerrilla modelling” – how to get started on concept modelling without anyone realising it
        • Naming conventions and graphic guidelines – ensuring correctness, consistency, and readability
        • Concept models as a starting point for process discovery
        • Practical examples of concept modelling supporting process work, architecture work, and commercial software selection.

        [Video introduction]

        Read less
          Alec Sharp | Founder | Clariteq Systems Consulting
        John O’Brien
        13:30 - 17:00 | 7 April

        DataOps for Better and Faster Analytics

        In this half day virtual seminar John O'Brien will explain the benefits of the DataOps methodology to improve on your Analytics. [Video introduction]
        Read more

        Adopting the DataOps Methodology is helping agile teams deliver data and analytics faster and more manageable in modern data infrastructure and ecosystems. DataOps is critical for companies to become resilient with data and analytics delivery in a volatile and uncertain global business environment. Going beyond DevOps for continuous deployments, DataOps leverages principles from other disciplines to evolve data engineering and management.

        Companies need data and analytics more than ever to be agile and competitive in today’s fast-changing environment. DataOps can be an enterprise-wide initiative or an independent agile delivery team working to improve how they deliver data analytics for their customer. Gaining traction takes time and ongoing support.

        This seminar will cover:

        • The challenges in current data environments and IT
        • What DataOps is and how it differs from other approaches
        • Which principles and technologies to focus on initially
        • How to adopt DataOps to speed analytics development and delivery
        • How to continuously engineer, deploy, and operationalize data pipelines with automation and monitoring
        • Setting expectations and planning for DataOps maturity.

         

        Course Description

        1. Understanding why we need to change

        • How business Analytics has changed from diagnostic to predictive
        • How data sources are increasing
        • The impact of data integration on Data Management
        • Changes in IT development methodologies and organizations
        • Supporting new data products
        • How DataOps is emerging as the next era
        • Reviewing the Agile Manifesto
        • Important aspect of DevOps
        • Review statistical process control for DataOps
        • How DataOps can embed Data Quality and Data Governance
        • Defining DataOps and the DataOps Manifesto
        • Comparing DevOps to DataOps

        2. Making DataOps Work

        The 7 key concepts to focus on for DataOps

        • How Connectors can make a difference
        • How engineered data pipelines will work
        • How “data drift” will impact data work
        • Set up repositories for Data Governance and Data Quality
        • The role of data hubs and MDM
        • How to set up measurements correctly
        • Leveraging DataOps Platform instrumentation

        The 2 key processes to focus on for DataOps

        • Components needed to deliver on business ideation
        • Building data and Analytics deliverables with DataOps

        3. Managing DataOps: defining Metrics and Maturity Models

        • Defining Metrics for Data and Analytics delivery
        • Key DataOps metrics
        • How to leverage reusability metrics
        • Reviewing metrics for process improvement
        • Maturity stage of DataOps adoption
        • CMMI-based Maturity Model
        • IBM Maturity Model.

        [Video introduction]

        Read less
          John O’Brien | CEO and Principal Advisor | Radiant Advisors
        Keith McCormick
        13:30 – 17:00 | 11 May

        Profiting with Practical Supervised Machine Learning

        Practical half day seminar on getting started with Supervised machine learning by Keith McCormick. How do you decide when to use Supervised versus Unsupervised machine learning? How to properly prepare data for different kinds of supervised models and how can data preparation be automated in parallel with the model if deployment is to succeed? [Video introduction]
        Read more

        Regression, decision trees, neural networks—along with many other supervised learning techniques—provide powerful predictive insights. Once built, the models can produce key indicators to optimize the allocation of organizational resources.
        New users of these established techniques are often impressed with how easy it all seems to be. Modeling software to build these models is widely available but often results in disappointing results. Many fail to even recognize that proper problem definition was the problem. They likely conclude that the data was not capable of better performance.
        The deployment phase includes proper model interpretation and looking for clues that the model will perform well on unseen data. Although the predictive power of these machine-learning models can be very impressive, there is no benefit unless they inform value-focused actions. Models must be deployed in an automated fashion to continually support decision-making for residual impact. The instructor will show how to interpret supervised models with an eye toward decisioning automation.

        The seminar
        In this half-day seminar, Keith McCormick will overview the two most important and foundational techniques in supervised machine learning, and explain why 70-80% or more of everyday problems faced in established industries can be addressed with one particular machine learning strategy. The focus will be on highly practical techniques for maximizing your results whether you are brand new to predictive analytics or you’ve made some attempts but have been disappointed in the results so far. Veteran users of these techniques will also benefit because a comparison will be made between these traditional techniques and some features of newer techniques. We will explore that while tempting, the newer techniques are rarely the best fit except in a handful of niche application areas that many organizations will not face (at least not in the short term). Participants will leave with specific ideas to apply to their current and future projects.

        Learning Objectives

        • When to apply supervised or unsupervised modeling methods
        • Options for inserting machine learning into the decision making of your organization
        • How to use multiple models for value estimation and classification
        • How to properly prepare data for different kinds of supervised models
        • Interpret model coefficients and output to translate across platforms and languages, including the widely used Predictive Modeling Markup Language (PMML)
        • Explore the pros and cons of “black box” models including ensembles
        • How data preparation must be automated in parallel with the model if deployment is to succeed
        • Compare model accuracy scores to model propensity scores that drive decisions at deployment.

         

        Who is it for?

        • Analytic Practitioners
        • Data Scientists
        • IT Professionals
        • Technology Planners
        • Consultants; Business Analysts
        • Analytic Project Leaders.

         

        Course Description

        1. How to choose the best machine learning strategy

        • How supervised learning compares to other options
        • The reality and the hype regarding machine learning
        • What are the classic traditional machine learning techniques?
        • The two main types of supervised machine learning

        2. Decision Trees: Still the best choice for many everyday challenges

        • Exploring and interpreting insights with a completed decision tree model
        • A brief primer on the various types of decision tree algorithms
        • Strategic considerations and advantages of decision trees
        • Deployment and bringing you ML models into production

        3. Introducing the CART decision tree

        • CART under the hood
        • Processing various variable types with CART
        • Understanding pruning
        • How CART handles missing data with “surrogates”
        • The “Roshoman effect” in machine learning

        4. Additional Supervised Techniques

        • Comparing linear regression to neural networks
        • How to embrace the benefits of neural networks without actually using them
        • Regression trees: how to use decision trees to address regression problems
        • What are “ensemble” methods and why are they so popular?
        • Keeping your solutions practical and transparent

        [Video introduction]

        Read less
          Keith McCormick | Data Science Consultant and Instructor |
        Donald Farmer
        13:30 - 17:00 | 31 March

        Data and Analytics as a Line of Business

        Practical workshop by Donald Farmer on how to leverage on your analytics efforts and developing this into a line of business. [Video introduction]
        Read more

        By the end of this workshop your team will have a sound understanding of how data and analytics can expand, enhance and strengthen your business and your relationships with clients. You’ll have some practical guidelines for strategy, messaging and design which can get you started on your own analytics journey.

         

        Learning objectives

        • The nature of data as asset
        • The importance of analytics to realising that value
        • Ways in which data and analytics can be developed as a line of business
        • Simple models for data and analytics as a line of business
        • The value of benchmarking
        • How to encourage and support internal and external communities
        • How to be strategically more agile when creating data and analytics lines of business
        • Addressing ethics and governance concerns.

         

        Course Description

        1. Introduction: Data as a resource, analytics as a differentiator

        We believe that data without analytics is a wasted resource; analytics without action is a wasted effort. We review the value of data to software companies and the potential for analytics as a new line of business.

         2. Case studies

        Real-world examples of software companies who have developed analytic products and services using a gameplan methodology.
         
        3. Three simple models to get you started

        Although there are many ways in which you can leverage data as a resource and analytics as an offering, we have found three to be relatively easy and effective to start with. We’ll review the components and technologies of each, with some guidelines for success and pitfalls to avoid.

        • User dashboards
        • Predictive analytics and alerting
        • Benchmarking and associative analytics

        4. Communities of practice and tools of choice

        When you introduce analytics as a line of business, users and their social interactions, whether in the office or online, will be critical to your success. We show how communities of practice develop around the tools we choose – and we describe how to ensure your tool is chosen.

         5. Governance and privacy

        In any discussion of data and analytics today, concerns about privacy and compliance always come to the surface. We’ll introduce the subject with enough detail for you take the first, important, practical steps to being well governed for today’s regulatory environment.

         6. Narratives and gameplans

        These are simple tools for mapping and aligning strategy. However, although simple, they offer subtle and effective capabilities for planning features and releases and for aligning teams such as marketing and management around a vision.

         

        Who’s it for? 

        • CIOs, CTOs, analytics leaders and data management leaders
        • Data scientists, and data analysts.

        [Video introduction]

        Read less
          Donald Farmer | Director | TreeHive Strategy
        Alec Sharp
        13:30 - 17:00 | 6 October

        The Data-Process Connection

        Alec Sharp illustrates the many ways concept models (conceptual data models) support business process change and business analysis. [Video introduction]
        Read more

        Whether you call it a conceptual data model, a domain map, a business object model, or even a “thing model,” a concept model is invaluable to process and architecture initiatives. Why? Because processes, capabilities, and solutions act on “things” – Settle Claim, Register Unit, Resolve Service Issue, and so on. Those things are usually “entities” or “objects” in the concept model, and clarity on “what is one of these things?” contributes immensely to clarity on what the corresponding processes are.
        After introducing methods to get people, even C-level executives, engaged in concept modelling, we’ll introduce and get practice with guidelines to ensure proper naming and definition of entities/concepts/business objects. We’ll also see that success depends on recognising that a concept model is a description of a business, not a description of a database. Another key – don’t call it a data model!
        Drawing on almost forty years of successful modelling, on projects of every size and type, this session introduces proven techniques backed up with current, real-life examples.

        Topics include:

        • Concept modelling essentials – things, facts about things, and the policies and rules governing things
        • “Guerrilla modelling” – how to get started on concept modelling without anyone realising it
        • Naming conventions and graphic guidelines – ensuring correctness, consistency, and readability
        • Concept models as a starting point for process discovery
        • Practical examples of concept modelling supporting process work, architecture work, and commercial software selection.

        [Video introduction]

        Read less
          Alec Sharp | Founder | Clariteq Systems Consulting
        John O’Brien
        13:30 - 17:00 | 7 April

        DataOps for Better and Faster Analytics

        In this half day virtual seminar John O'Brien will explain the benefits of the DataOps methodology to improve on your Analytics. [Video introduction]
        Read more

        Adopting the DataOps Methodology is helping agile teams deliver data and analytics faster and more manageable in modern data infrastructure and ecosystems. DataOps is critical for companies to become resilient with data and analytics delivery in a volatile and uncertain global business environment. Going beyond DevOps for continuous deployments, DataOps leverages principles from other disciplines to evolve data engineering and management.

        Companies need data and analytics more than ever to be agile and competitive in today’s fast-changing environment. DataOps can be an enterprise-wide initiative or an independent agile delivery team working to improve how they deliver data analytics for their customer. Gaining traction takes time and ongoing support.

        This seminar will cover:

        • The challenges in current data environments and IT
        • What DataOps is and how it differs from other approaches
        • Which principles and technologies to focus on initially
        • How to adopt DataOps to speed analytics development and delivery
        • How to continuously engineer, deploy, and operationalize data pipelines with automation and monitoring
        • Setting expectations and planning for DataOps maturity.

         

        Course Description

        1. Understanding why we need to change

        • How business Analytics has changed from diagnostic to predictive
        • How data sources are increasing
        • The impact of data integration on Data Management
        • Changes in IT development methodologies and organizations
        • Supporting new data products
        • How DataOps is emerging as the next era
        • Reviewing the Agile Manifesto
        • Important aspect of DevOps
        • Review statistical process control for DataOps
        • How DataOps can embed Data Quality and Data Governance
        • Defining DataOps and the DataOps Manifesto
        • Comparing DevOps to DataOps

        2. Making DataOps Work

        The 7 key concepts to focus on for DataOps

        • How Connectors can make a difference
        • How engineered data pipelines will work
        • How “data drift” will impact data work
        • Set up repositories for Data Governance and Data Quality
        • The role of data hubs and MDM
        • How to set up measurements correctly
        • Leveraging DataOps Platform instrumentation

        The 2 key processes to focus on for DataOps

        • Components needed to deliver on business ideation
        • Building data and Analytics deliverables with DataOps

        3. Managing DataOps: defining Metrics and Maturity Models

        • Defining Metrics for Data and Analytics delivery
        • Key DataOps metrics
        • How to leverage reusability metrics
        • Reviewing metrics for process improvement
        • Maturity stage of DataOps adoption
        • CMMI-based Maturity Model
        • IBM Maturity Model.

        [Video introduction]

        Read less
          John O’Brien | CEO and Principal Advisor | Radiant Advisors
        Keith McCormick
        13:30 – 17:00 | 11 May

        Profiting with Practical Supervised Machine Learning

        Practical half day seminar on getting started with Supervised machine learning by Keith McCormick. How do you decide when to use Supervised versus Unsupervised machine learning? How to properly prepare data for different kinds of supervised models and how can data preparation be automated in parallel with the model if deployment is to succeed? [Video introduction]
        Read more

        Regression, decision trees, neural networks—along with many other supervised learning techniques—provide powerful predictive insights. Once built, the models can produce key indicators to optimize the allocation of organizational resources.
        New users of these established techniques are often impressed with how easy it all seems to be. Modeling software to build these models is widely available but often results in disappointing results. Many fail to even recognize that proper problem definition was the problem. They likely conclude that the data was not capable of better performance.
        The deployment phase includes proper model interpretation and looking for clues that the model will perform well on unseen data. Although the predictive power of these machine-learning models can be very impressive, there is no benefit unless they inform value-focused actions. Models must be deployed in an automated fashion to continually support decision-making for residual impact. The instructor will show how to interpret supervised models with an eye toward decisioning automation.

        The seminar
        In this half-day seminar, Keith McCormick will overview the two most important and foundational techniques in supervised machine learning, and explain why 70-80% or more of everyday problems faced in established industries can be addressed with one particular machine learning strategy. The focus will be on highly practical techniques for maximizing your results whether you are brand new to predictive analytics or you’ve made some attempts but have been disappointed in the results so far. Veteran users of these techniques will also benefit because a comparison will be made between these traditional techniques and some features of newer techniques. We will explore that while tempting, the newer techniques are rarely the best fit except in a handful of niche application areas that many organizations will not face (at least not in the short term). Participants will leave with specific ideas to apply to their current and future projects.

        Learning Objectives

        • When to apply supervised or unsupervised modeling methods
        • Options for inserting machine learning into the decision making of your organization
        • How to use multiple models for value estimation and classification
        • How to properly prepare data for different kinds of supervised models
        • Interpret model coefficients and output to translate across platforms and languages, including the widely used Predictive Modeling Markup Language (PMML)
        • Explore the pros and cons of “black box” models including ensembles
        • How data preparation must be automated in parallel with the model if deployment is to succeed
        • Compare model accuracy scores to model propensity scores that drive decisions at deployment.

         

        Who is it for?

        • Analytic Practitioners
        • Data Scientists
        • IT Professionals
        • Technology Planners
        • Consultants; Business Analysts
        • Analytic Project Leaders.

         

        Course Description

        1. How to choose the best machine learning strategy

        • How supervised learning compares to other options
        • The reality and the hype regarding machine learning
        • What are the classic traditional machine learning techniques?
        • The two main types of supervised machine learning

        2. Decision Trees: Still the best choice for many everyday challenges

        • Exploring and interpreting insights with a completed decision tree model
        • A brief primer on the various types of decision tree algorithms
        • Strategic considerations and advantages of decision trees
        • Deployment and bringing you ML models into production

        3. Introducing the CART decision tree

        • CART under the hood
        • Processing various variable types with CART
        • Understanding pruning
        • How CART handles missing data with “surrogates”
        • The “Roshoman effect” in machine learning

        4. Additional Supervised Techniques

        • Comparing linear regression to neural networks
        • How to embrace the benefits of neural networks without actually using them
        • Regression trees: how to use decision trees to address regression problems
        • What are “ensemble” methods and why are they so popular?
        • Keeping your solutions practical and transparent

        [Video introduction]

        Read less
          Keith McCormick | Data Science Consultant and Instructor |
        Donald Farmer
        13:30 - 17:00 | 31 March

        Data and Analytics as a Line of Business

        Practical workshop by Donald Farmer on how to leverage on your analytics efforts and developing this into a line of business. [Video introduction]
        Read more

        By the end of this workshop your team will have a sound understanding of how data and analytics can expand, enhance and strengthen your business and your relationships with clients. You’ll have some practical guidelines for strategy, messaging and design which can get you started on your own analytics journey.

         

        Learning objectives

        • The nature of data as asset
        • The importance of analytics to realising that value
        • Ways in which data and analytics can be developed as a line of business
        • Simple models for data and analytics as a line of business
        • The value of benchmarking
        • How to encourage and support internal and external communities
        • How to be strategically more agile when creating data and analytics lines of business
        • Addressing ethics and governance concerns.

         

        Course Description

        1. Introduction: Data as a resource, analytics as a differentiator

        We believe that data without analytics is a wasted resource; analytics without action is a wasted effort. We review the value of data to software companies and the potential for analytics as a new line of business.

         2. Case studies

        Real-world examples of software companies who have developed analytic products and services using a gameplan methodology.
         
        3. Three simple models to get you started

        Although there are many ways in which you can leverage data as a resource and analytics as an offering, we have found three to be relatively easy and effective to start with. We’ll review the components and technologies of each, with some guidelines for success and pitfalls to avoid.

        • User dashboards
        • Predictive analytics and alerting
        • Benchmarking and associative analytics

        4. Communities of practice and tools of choice

        When you introduce analytics as a line of business, users and their social interactions, whether in the office or online, will be critical to your success. We show how communities of practice develop around the tools we choose – and we describe how to ensure your tool is chosen.

         5. Governance and privacy

        In any discussion of data and analytics today, concerns about privacy and compliance always come to the surface. We’ll introduce the subject with enough detail for you take the first, important, practical steps to being well governed for today’s regulatory environment.

         6. Narratives and gameplans

        These are simple tools for mapping and aligning strategy. However, although simple, they offer subtle and effective capabilities for planning features and releases and for aligning teams such as marketing and management around a vision.

         

        Who’s it for? 

        • CIOs, CTOs, analytics leaders and data management leaders
        • Data scientists, and data analysts.

        [Video introduction]

        Read less
          Donald Farmer | Director | TreeHive Strategy
        Alec Sharp
        13:30 - 17:00 | 6 October

        The Data-Process Connection

        Alec Sharp illustrates the many ways concept models (conceptual data models) support business process change and business analysis. [Video introduction]
        Read more

        Whether you call it a conceptual data model, a domain map, a business object model, or even a “thing model,” a concept model is invaluable to process and architecture initiatives. Why? Because processes, capabilities, and solutions act on “things” – Settle Claim, Register Unit, Resolve Service Issue, and so on. Those things are usually “entities” or “objects” in the concept model, and clarity on “what is one of these things?” contributes immensely to clarity on what the corresponding processes are.
        After introducing methods to get people, even C-level executives, engaged in concept modelling, we’ll introduce and get practice with guidelines to ensure proper naming and definition of entities/concepts/business objects. We’ll also see that success depends on recognising that a concept model is a description of a business, not a description of a database. Another key – don’t call it a data model!
        Drawing on almost forty years of successful modelling, on projects of every size and type, this session introduces proven techniques backed up with current, real-life examples.

        Topics include:

        • Concept modelling essentials – things, facts about things, and the policies and rules governing things
        • “Guerrilla modelling” – how to get started on concept modelling without anyone realising it
        • Naming conventions and graphic guidelines – ensuring correctness, consistency, and readability
        • Concept models as a starting point for process discovery
        • Practical examples of concept modelling supporting process work, architecture work, and commercial software selection.

        [Video introduction]

        Read less
          Alec Sharp | Founder | Clariteq Systems Consulting
        John O’Brien
        13:30 - 17:00 | 7 April

        DataOps for Better and Faster Analytics

        In this half day virtual seminar John O'Brien will explain the benefits of the DataOps methodology to improve on your Analytics. [Video introduction]
        Read more

        Adopting the DataOps Methodology is helping agile teams deliver data and analytics faster and more manageable in modern data infrastructure and ecosystems. DataOps is critical for companies to become resilient with data and analytics delivery in a volatile and uncertain global business environment. Going beyond DevOps for continuous deployments, DataOps leverages principles from other disciplines to evolve data engineering and management.

        Companies need data and analytics more than ever to be agile and competitive in today’s fast-changing environment. DataOps can be an enterprise-wide initiative or an independent agile delivery team working to improve how they deliver data analytics for their customer. Gaining traction takes time and ongoing support.

        This seminar will cover:

        • The challenges in current data environments and IT
        • What DataOps is and how it differs from other approaches
        • Which principles and technologies to focus on initially
        • How to adopt DataOps to speed analytics development and delivery
        • How to continuously engineer, deploy, and operationalize data pipelines with automation and monitoring
        • Setting expectations and planning for DataOps maturity.

         

        Course Description

        1. Understanding why we need to change

        • How business Analytics has changed from diagnostic to predictive
        • How data sources are increasing
        • The impact of data integration on Data Management
        • Changes in IT development methodologies and organizations
        • Supporting new data products
        • How DataOps is emerging as the next era
        • Reviewing the Agile Manifesto
        • Important aspect of DevOps
        • Review statistical process control for DataOps
        • How DataOps can embed Data Quality and Data Governance
        • Defining DataOps and the DataOps Manifesto
        • Comparing DevOps to DataOps

        2. Making DataOps Work

        The 7 key concepts to focus on for DataOps

        • How Connectors can make a difference
        • How engineered data pipelines will work
        • How “data drift” will impact data work
        • Set up repositories for Data Governance and Data Quality
        • The role of data hubs and MDM
        • How to set up measurements correctly
        • Leveraging DataOps Platform instrumentation

        The 2 key processes to focus on for DataOps

        • Components needed to deliver on business ideation
        • Building data and Analytics deliverables with DataOps

        3. Managing DataOps: defining Metrics and Maturity Models

        • Defining Metrics for Data and Analytics delivery
        • Key DataOps metrics
        • How to leverage reusability metrics
        • Reviewing metrics for process improvement
        • Maturity stage of DataOps adoption
        • CMMI-based Maturity Model
        • IBM Maturity Model.

        [Video introduction]

        Read less
          John O’Brien | CEO and Principal Advisor | Radiant Advisors
        Keith McCormick
        13:30 – 17:00 | 11 May

        Profiting with Practical Supervised Machine Learning

        Practical half day seminar on getting started with Supervised machine learning by Keith McCormick. How do you decide when to use Supervised versus Unsupervised machine learning? How to properly prepare data for different kinds of supervised models and how can data preparation be automated in parallel with the model if deployment is to succeed? [Video introduction]
        Read more

        Regression, decision trees, neural networks—along with many other supervised learning techniques—provide powerful predictive insights. Once built, the models can produce key indicators to optimize the allocation of organizational resources.
        New users of these established techniques are often impressed with how easy it all seems to be. Modeling software to build these models is widely available but often results in disappointing results. Many fail to even recognize that proper problem definition was the problem. They likely conclude that the data was not capable of better performance.
        The deployment phase includes proper model interpretation and looking for clues that the model will perform well on unseen data. Although the predictive power of these machine-learning models can be very impressive, there is no benefit unless they inform value-focused actions. Models must be deployed in an automated fashion to continually support decision-making for residual impact. The instructor will show how to interpret supervised models with an eye toward decisioning automation.

        The seminar
        In this half-day seminar, Keith McCormick will overview the two most important and foundational techniques in supervised machine learning, and explain why 70-80% or more of everyday problems faced in established industries can be addressed with one particular machine learning strategy. The focus will be on highly practical techniques for maximizing your results whether you are brand new to predictive analytics or you’ve made some attempts but have been disappointed in the results so far. Veteran users of these techniques will also benefit because a comparison will be made between these traditional techniques and some features of newer techniques. We will explore that while tempting, the newer techniques are rarely the best fit except in a handful of niche application areas that many organizations will not face (at least not in the short term). Participants will leave with specific ideas to apply to their current and future projects.

        Learning Objectives

        • When to apply supervised or unsupervised modeling methods
        • Options for inserting machine learning into the decision making of your organization
        • How to use multiple models for value estimation and classification
        • How to properly prepare data for different kinds of supervised models
        • Interpret model coefficients and output to translate across platforms and languages, including the widely used Predictive Modeling Markup Language (PMML)
        • Explore the pros and cons of “black box” models including ensembles
        • How data preparation must be automated in parallel with the model if deployment is to succeed
        • Compare model accuracy scores to model propensity scores that drive decisions at deployment.

         

        Who is it for?

        • Analytic Practitioners
        • Data Scientists
        • IT Professionals
        • Technology Planners
        • Consultants; Business Analysts
        • Analytic Project Leaders.

         

        Course Description

        1. How to choose the best machine learning strategy

        • How supervised learning compares to other options
        • The reality and the hype regarding machine learning
        • What are the classic traditional machine learning techniques?
        • The two main types of supervised machine learning

        2. Decision Trees: Still the best choice for many everyday challenges

        • Exploring and interpreting insights with a completed decision tree model
        • A brief primer on the various types of decision tree algorithms
        • Strategic considerations and advantages of decision trees
        • Deployment and bringing you ML models into production

        3. Introducing the CART decision tree

        • CART under the hood
        • Processing various variable types with CART
        • Understanding pruning
        • How CART handles missing data with “surrogates”
        • The “Roshoman effect” in machine learning

        4. Additional Supervised Techniques

        • Comparing linear regression to neural networks
        • How to embrace the benefits of neural networks without actually using them
        • Regression trees: how to use decision trees to address regression problems
        • What are “ensemble” methods and why are they so popular?
        • Keeping your solutions practical and transparent

        [Video introduction]

        Read less
          Keith McCormick | Data Science Consultant and Instructor |
        Donald Farmer
        13:30 - 17:00 | 31 March

        Data and Analytics as a Line of Business

        Practical workshop by Donald Farmer on how to leverage on your analytics efforts and developing this into a line of business. [Video introduction]
        Read more

        By the end of this workshop your team will have a sound understanding of how data and analytics can expand, enhance and strengthen your business and your relationships with clients. You’ll have some practical guidelines for strategy, messaging and design which can get you started on your own analytics journey.

         

        Learning objectives

        • The nature of data as asset
        • The importance of analytics to realising that value
        • Ways in which data and analytics can be developed as a line of business
        • Simple models for data and analytics as a line of business
        • The value of benchmarking
        • How to encourage and support internal and external communities
        • How to be strategically more agile when creating data and analytics lines of business
        • Addressing ethics and governance concerns.

         

        Course Description

        1. Introduction: Data as a resource, analytics as a differentiator

        We believe that data without analytics is a wasted resource; analytics without action is a wasted effort. We review the value of data to software companies and the potential for analytics as a new line of business.

         2. Case studies

        Real-world examples of software companies who have developed analytic products and services using a gameplan methodology.
         
        3. Three simple models to get you started

        Although there are many ways in which you can leverage data as a resource and analytics as an offering, we have found three to be relatively easy and effective to start with. We’ll review the components and technologies of each, with some guidelines for success and pitfalls to avoid.

        • User dashboards
        • Predictive analytics and alerting
        • Benchmarking and associative analytics

        4. Communities of practice and tools of choice

        When you introduce analytics as a line of business, users and their social interactions, whether in the office or online, will be critical to your success. We show how communities of practice develop around the tools we choose – and we describe how to ensure your tool is chosen.

         5. Governance and privacy

        In any discussion of data and analytics today, concerns about privacy and compliance always come to the surface. We’ll introduce the subject with enough detail for you take the first, important, practical steps to being well governed for today’s regulatory environment.

         6. Narratives and gameplans

        These are simple tools for mapping and aligning strategy. However, although simple, they offer subtle and effective capabilities for planning features and releases and for aligning teams such as marketing and management around a vision.

         

        Who’s it for? 

        • CIOs, CTOs, analytics leaders and data management leaders
        • Data scientists, and data analysts.

        [Video introduction]

        Read less
          Donald Farmer | Director | TreeHive Strategy
        Alec Sharp
        13:30 - 17:00 | 6 October

        The Data-Process Connection

        Alec Sharp illustrates the many ways concept models (conceptual data models) support business process change and business analysis. [Video introduction]
        Read more

        Whether you call it a conceptual data model, a domain map, a business object model, or even a “thing model,” a concept model is invaluable to process and architecture initiatives. Why? Because processes, capabilities, and solutions act on “things” – Settle Claim, Register Unit, Resolve Service Issue, and so on. Those things are usually “entities” or “objects” in the concept model, and clarity on “what is one of these things?” contributes immensely to clarity on what the corresponding processes are.
        After introducing methods to get people, even C-level executives, engaged in concept modelling, we’ll introduce and get practice with guidelines to ensure proper naming and definition of entities/concepts/business objects. We’ll also see that success depends on recognising that a concept model is a description of a business, not a description of a database. Another key – don’t call it a data model!
        Drawing on almost forty years of successful modelling, on projects of every size and type, this session introduces proven techniques backed up with current, real-life examples.

        Topics include:

        • Concept modelling essentials – things, facts about things, and the policies and rules governing things
        • “Guerrilla modelling” – how to get started on concept modelling without anyone realising it
        • Naming conventions and graphic guidelines – ensuring correctness, consistency, and readability
        • Concept models as a starting point for process discovery
        • Practical examples of concept modelling supporting process work, architecture work, and commercial software selection.

        [Video introduction]

        Read less
          Alec Sharp | Founder | Clariteq Systems Consulting
        John O’Brien
        13:30 - 17:00 | 7 April

        DataOps for Better and Faster Analytics

        In this half day virtual seminar John O'Brien will explain the benefits of the DataOps methodology to improve on your Analytics. [Video introduction]
        Read more

        Adopting the DataOps Methodology is helping agile teams deliver data and analytics faster and more manageable in modern data infrastructure and ecosystems. DataOps is critical for companies to become resilient with data and analytics delivery in a volatile and uncertain global business environment. Going beyond DevOps for continuous deployments, DataOps leverages principles from other disciplines to evolve data engineering and management.

        Companies need data and analytics more than ever to be agile and competitive in today’s fast-changing environment. DataOps can be an enterprise-wide initiative or an independent agile delivery team working to improve how they deliver data analytics for their customer. Gaining traction takes time and ongoing support.

        This seminar will cover:

        • The challenges in current data environments and IT
        • What DataOps is and how it differs from other approaches
        • Which principles and technologies to focus on initially
        • How to adopt DataOps to speed analytics development and delivery
        • How to continuously engineer, deploy, and operationalize data pipelines with automation and monitoring
        • Setting expectations and planning for DataOps maturity.

         

        Course Description

        1. Understanding why we need to change

        • How business Analytics has changed from diagnostic to predictive
        • How data sources are increasing
        • The impact of data integration on Data Management
        • Changes in IT development methodologies and organizations
        • Supporting new data products
        • How DataOps is emerging as the next era
        • Reviewing the Agile Manifesto
        • Important aspect of DevOps
        • Review statistical process control for DataOps
        • How DataOps can embed Data Quality and Data Governance
        • Defining DataOps and the DataOps Manifesto
        • Comparing DevOps to DataOps

        2. Making DataOps Work

        The 7 key concepts to focus on for DataOps

        • How Connectors can make a difference
        • How engineered data pipelines will work
        • How “data drift” will impact data work
        • Set up repositories for Data Governance and Data Quality
        • The role of data hubs and MDM
        • How to set up measurements correctly
        • Leveraging DataOps Platform instrumentation

        The 2 key processes to focus on for DataOps

        • Components needed to deliver on business ideation
        • Building data and Analytics deliverables with DataOps

        3. Managing DataOps: defining Metrics and Maturity Models

        • Defining Metrics for Data and Analytics delivery
        • Key DataOps metrics
        • How to leverage reusability metrics
        • Reviewing metrics for process improvement
        • Maturity stage of DataOps adoption
        • CMMI-based Maturity Model
        • IBM Maturity Model.

        [Video introduction]

        Read less
          John O’Brien | CEO and Principal Advisor | Radiant Advisors

         

        Limited time? Join one day!
        Can you only attend one day? It is possible to attend only the first or only the second conference day and of course the full conference. Delegates also gain four months access to the conference recordings of the elected day so there’s no need to miss out on any session.
        Payment by credit card is also available. Please mention this in the Comment-field upon registration and find further instructions for credit card payment on our customer service page.

        29 march

        09:00 - 09:15 | Opening
        Plenary, Room 1    Werner Schoots
        | Chairman
        Plenary, Room 1, Room 2    Dennis van Gelder, Tanja Ubert, Huub Meertens
        09:15 - 10:05 | Connecting the Dots, Creating Data Centric Business Value (Dutch spoken)
        Plenary    Jan Henderyckx
        10:10 - 11:00 | The transition to the new pension system – Data Management approach and tooling (Dutch spoken)
        Plenary    Arjen Bouman
        11:20 - 12:20 | Enterprise Semantic Data Management (Dutch spoken)
        Plenary    Marco Brattinga
        12:20 - 13:20 | Lunch break
        Plenary 
        13:20 - 14:10 | Responsible Data Science (Dutch spoken)
        Plenary    Tanja Ubert, Gabriella Obispa
        14:15 - 14:45 | Simplified Data Architecture: Data Warehouse Automation with Datavault Builder
        Plenary    Erik Fransen, Petr Beles
        15:00 - 15:50 | DataOps – best practices and lessons learned (Dutch spoken)
        Plenary    Niels Naglé, Vincent Goris
        15:55 - 16:45 | Data Minimization: A New Design principle for Data Architectures (Dutch spoken)
        Plenary    Rick van der Lans

        30 march

        09:30 - 09:45 | Opening – virtual conference day
        Plenary    Werner Schoots
        09:45 - 10:35 | Cutting Data Fabric and Mesh to Measure
        Plenary    Barry Devlin
        10:45 - 11:35 | Remote Data Modelstorming with BEAM: Lessons learnt from 2 years of data modeling training and consulting online
        Plenary    Lawrence Corr
        11:45 - 12:35 | Driving measurable value in Established Industries with Traditional Machine Learning
        Plenary    Keith McCormick
        12:30 - 13:30 | Lunch break
        Plenary 
        13:30 – 14:20 | Building a Business-Driven Roadmap for Modern Cloud Data Architecture
        Plenary    John O’Brien
        14:30 - 15:00 | Creating a Predictable and Mature BI Value Stream with Data Automation
        Plenary    Erik van der Hoeven
        15:10 - 16:00 | Openness and ownership – the balancing act of enterprise data
        Plenary    Donald Farmer
        16:10 - 17:00 | Concept Modelling – An Angst-Free Framework for Engaging your Executives
        Plenary    Alec Sharp

        Workshop

        13:30 – 17:00 | Profiting with Practical Supervised Machine Learning
        11 May    Keith McCormick
        13:30 - 17:00 | Data and Analytics as a Line of Business
        31 March    Donald Farmer
        13:30 - 17:00 | The Data-Process Connection
        6 October    Alec Sharp
        13:30 - 17:00 | DataOps for Better and Faster Analytics
        7 April    John O’Brien

        Speakers

        Rick van der Lans

        Mark Madsen

        Bart Baesens

        William McKnight

        Erik Fransen

        Pieter den Hamer

        Jos Kuiper

        Kishan Shri

        Rutger Rienks

        Gold and Platinum Partners

        Exhibitors & Media partners

        Related events

        View the Adept Events calendar

        • Logisch Datawarehouse – Architectuur, Ontwerp en Technologie
        • Schedule
        • Big Databasetechnologie
        • Designing, Operating and Managing an Enterprise Data Lake
        • Machine Learning & Advanced Analytics
        • Business Intelligence & Datawarehousing Fundamentals
        • A Business-oriented Approach to Data Modelling

        News

        12-02-2025

        Nicola Askham presents keynote and workshop at DW & BI Summit 2025

        View

        28-01-2025

        Winfried Etzel presents keynote and workshop at DW & BI Summit 2025

        View

        08-01-2025

        Alec Sharp presents keynote and workshop at DW & BI Summit 2025

        View

        20-12-2024

        Linda Terlouw presents keynote at DW & BI Summit 2025

        View

        05-02-2024

        Ron Tolido presents keynote at DW & BI Summit 2024

        View

        05-02-2024

        Peter Boncz presents keynote on MotherDuck at DW & BI Summit 2024

        View

        05-02-2024

        MicroStrategy sponsor of DW & BI Summit 2024

        View

        24-01-2023

        Thomas Frisendal presents keynote and workshop about Graph Technology on DW & BI Summit jubilee

        View

        16-01-2023

        Nigel Turner presents keynote and workshop about Data Quality on DW & BI Summit jubilee

        View

        06-01-2023

        Alec Sharp presents keynote and workshop about Conceptual Data Modelling on DW & BI Summit jubilee

        View

        More news
        Adept Events
        KvK Den Haag: 56059825
        E: seminars@adeptevents.nl
        T: +31 (0)172 742680
        M: +31 (0)6 113 118 60
        W: www.adeptevents.nl

        Release
        www.release.nl
        @Release_nl
        Download the Release App

        BI-Platform
        www.biplatform.nl
        @BIPlatform
        Download the BI-Platform App

        © Adept Events is a registered trademark of Array Media B.V.
        Share to Twitter Share to Facebook Share to LinkedIn
        © 2025 Array Media b.v. - All rights reserved | Privacy | Disclaimer