Accelerated Machine Learning and Deep Learning with Intel

Europe/Berlin
Online/Zoom

Online/Zoom

Andreas Marek (MPCDF), Timoteo Colnaghi (MPCDF)
Description

During this workshop, we will showcase how to accelerate your classical Machine Learning and Deep Learning workloads on Intel architecture. We will present the Intel optimizations of commonly used data science libraries such as Pandas, classical Machine Learning libraries such as Scikit-learn and XGBoost and of course, the optimizations of Deep Learning libraries such TF and PyTorch. We will also discuss various topics regarding hyperparameter optimization, AI on HPC system, inference optimization with OpenVINO library and compression techniques using Intel Neural Compressor.
 

Registration
Registration for Accelerated Machine Learning and Deep Learning with Intel
  • Wednesday, 8 March
    • 09:30 09:35
      Welcome and Introduction 5m

      Agenda and speakers' presentation

    • 09:35 10:00
      Hardware acceleration for AI and Intel® oneAPI AI Analytics Toolkit 25m

      In this session, we will first introduce the hardware features that are powering AI on Intel, we will then get a first glance at the software stack harnessing them, namely the Intel® oneAPI AI Analytics Toolkit.

      Speaker: Dr Séverine Habert (Intel)
    • 10:00 10:45
      How to accelerate Classical Machine Learning on Intel Architecture 45m

      In this session, we will cover the Intel-optimized libraries for Machine Learning. Python is currently ranked as the most popular programming language and is widely used in Data Science and Machine Learning. We will begin by covering the Intel® Distribution for Python and its optimizations. We will then cover the optimizations for ML Python packages such as Modin, Intel® Extension for Scikit-learn and XGBoost. The presentations will be accompanied with demos to showcase the performance speedup.

      Speaker: Vladimir Kilyazov (Intel)
    • 10:45 11:00
      Break 15m
    • 11:00 11:45
      Hands-on 45m
    • 11:45 12:00
      Closure day 1 15m
  • Thursday, 9 March
    • 09:30 10:05
      Optimize Deep Learning on Intel 35m

      In this session, we present to you what is behind the scenes of Deep Learning with the highly-optimized Intel® oneDNN library in order to get the best-in-class performance on Intel hardware. We then show you Intel® oneDNN in action in DL frameworks such as the Intel-optimized TensorFlow, Intel-optimized PyTorch and the Intel® Extension for PyTorch (IPEX) and Tensorflow (ITEX).

      Speaker: Akash Dhamasia (Intel)
    • 10:05 10:50
      Hands-on 45m
    • 10:50 11:00
      Break 10m
    • 11:00 12:00
      AI-driven multiphysics HPC applications on Intel architecture: Bridging the gap between HPC and ML 1h

      A major challenge in HPC is to make use of and understand the massive amounts of data that are being produced when running numerical simulations. For ML on the other hand, the challenge is to have access to enough data so that we have the confidence that our models truly understand the world. Therefore, researchers are looking to replace components of HPC applications with ML models to (a) reduce the need for data storage, (b) accelerate the simulations by ML models to capture longer timescales, and (c) achieve accurate simulations in some problems that the classical solvers are not applicable to. In this session we present this interdisciplinary field and highlight recent achievements on Intel® architectures.

      Speaker: Dr Massoud Rezavand (Intel)
    • 12:00 13:00
      Lunch 1h
    • 13:00 13:35
      Introduction to Neural Network Compression Techniques 35m

      In this session, we will explain various network compression techniques in Deep Learning—such as quantization, pruning, and knowledge distillation—, their benefits in terms of performance speed-up, and finally we will showcase you the Intel tools that help you compress your model, like the Intel® Neural Compressor.

      Speaker: Dr Nikolai Solmsdorf (Intel)
    • 13:35 14:20
      Hands-on 45m
    • 14:20 14:30
      Break 10m
    • 14:30 15:15
      Uncertainty estimation 45m

      In this session, we will talk about the limitations of conventional deep learning techniques such as being not explainable, overconfident, and being susceptible to adversarial attacks and why in safety critical applications, it is important to incorporate reliable uncertainty estimation to DNNs for trustworthy and informed decision making.
      Demo with IPEX.

      Speaker: Akash Dhamasia (Intel)
    • 15:15 16:00
      Easily speed up Deep Learning inference – Write once deploy anywhere 45m

      In this session, we will showcase the Intel® Distribution of OpenVINO™ Toolkit that allows you to optimize for high-performance inference models that you trained with TensorFlow or with PyTorch. We will demonstrate how to use it to write once and deploy on multiple Intel hardware.

      Speaker: Anas Ahouzi (Intel)
    • 15:45 16:00
      Closure 15m