Accelerated Machine Learning and Deep Learning with Intel

Europe/Berlin
Online/Zoom

Online/Zoom

Andreas Marek (MPCDF), Timoteo Colnaghi (MPCDF)
Description

During this workshop, we will showcase how to accelerate your classical Machine Learning and Deep Learning workloads on Intel architecture. We will present the Intel optimizations of commonly used data science libraries such as Pandas, classical Machine Learning libraries such as Scikit-learn and XGBoost and of course, the optimizations of Deep Learning libraries such TF and PyTorch. We will also discuss various topics regarding hyperparameter optimization, AI on HPC system, inference optimization with OpenVINO library and compression techniques using Intel Neural Compressor.
 

Registration
Registration for Accelerated Machine Learning and Deep Learning with Intel
  • Wednesday, 8 March
    • 1
      Welcome and Introduction

      Agenda and speakers' presentation

    • 2
      Hardware acceleration for AI and Intel® oneAPI AI Analytics Toolkit

      In this session, we will first introduce the hardware features that are powering AI on Intel, we will then get a first glance at the software stack harnessing them, namely the Intel® oneAPI AI Analytics Toolkit.

      Speaker: Dr Séverine Habert (Intel)
    • 3
      How to accelerate Classical Machine Learning on Intel Architecture

      In this session, we will cover the Intel-optimized libraries for Machine Learning. Python is currently ranked as the most popular programming language and is widely used in Data Science and Machine Learning. We will begin by covering the Intel® Distribution for Python and its optimizations. We will then cover the optimizations for ML Python packages such as Modin, Intel® Extension for Scikit-learn and XGBoost. The presentations will be accompanied with demos to showcase the performance speedup.

      Speaker: Vladimir Kilyazov (Intel)
    • 10:45
      Break
    • 4
      Hands-on
    • 5
      Closure day 1
  • Thursday, 9 March
    • 6
      Optimize Deep Learning on Intel

      In this session, we present to you what is behind the scenes of Deep Learning with the highly-optimized Intel® oneDNN library in order to get the best-in-class performance on Intel hardware. We then show you Intel® oneDNN in action in DL frameworks such as the Intel-optimized TensorFlow, Intel-optimized PyTorch and the Intel® Extension for PyTorch (IPEX) and Tensorflow (ITEX).

      Speaker: Akash Dhamasia (Intel)
    • 7
      Hands-on
    • 10:50
      Break
    • 8
      AI-driven multiphysics HPC applications on Intel architecture: Bridging the gap between HPC and ML

      A major challenge in HPC is to make use of and understand the massive amounts of data that are being produced when running numerical simulations. For ML on the other hand, the challenge is to have access to enough data so that we have the confidence that our models truly understand the world. Therefore, researchers are looking to replace components of HPC applications with ML models to (a) reduce the need for data storage, (b) accelerate the simulations by ML models to capture longer timescales, and (c) achieve accurate simulations in some problems that the classical solvers are not applicable to. In this session we present this interdisciplinary field and highlight recent achievements on Intel® architectures.

      Speaker: Dr Massoud Rezavand (Intel)
    • 12:00
      Lunch
    • 9
      Introduction to Neural Network Compression Techniques

      In this session, we will explain various network compression techniques in Deep Learning—such as quantization, pruning, and knowledge distillation—, their benefits in terms of performance speed-up, and finally we will showcase you the Intel tools that help you compress your model, like the Intel® Neural Compressor.

      Speaker: Dr Nikolai Solmsdorf (Intel)
    • 10
      Hands-on
    • 14:20
      Break
    • 11
      Uncertainty estimation

      In this session, we will talk about the limitations of conventional deep learning techniques such as being not explainable, overconfident, and being susceptible to adversarial attacks and why in safety critical applications, it is important to incorporate reliable uncertainty estimation to DNNs for trustworthy and informed decision making.
      Demo with IPEX.

      Speaker: Akash Dhamasia (Intel)
    • 12
      Easily speed up Deep Learning inference – Write once deploy anywhere

      In this session, we will showcase the Intel® Distribution of OpenVINO™ Toolkit that allows you to optimize for high-performance inference models that you trained with TensorFlow or with PyTorch. We will demonstrate how to use it to write once and deploy on multiple Intel hardware.

      Speaker: Anas Ahouzi (Intel)
    • 13
      Closure