Previous session videos
The center hosted a free online training series that helped prepare people for deep learning projects on NCSA’s HAL cluster. Each session featured an expert instructor covering topics focused on familiarizing novel users with different aspects and uses of HAL and training them to build deep neural network models.
Getting Started with HAL
Overview: This tutorial will introduce new users to HAL system, how to interact with it through Open OnDemand web interface, including Jupiter notebook and VS Code, and through command line interface via SSH. The tutorial will cover the Anaconda environment, batch job submission, data transfer, as well as an overview to the basic machine learning workflow.
- Training Slides
- Instructor: Volodymyr Kindratenko, Director of the Center for Artificial Intelligence Innovation
- Date: January 25, 2023
Introduction to PyTorch
Overview: This tutorial will introduce basics of PyTorch framework necessary to build a neural network, train it and evaluate the accuracy of the model.
- GitHub Repo
- Instructor: Priyam Mazumdar
- Date: February 1, 2023
Introduction to TensorFlow
Overview: This tutorial will introduce basics of TensorFlow necessary to build a neural network, train it and evaluate the accuracy of the model.
- Training Slides
- Instructor: Shirui Luo
- Date: February 8, 2023
Distributed Data Parallel Model Training in PyTorch
Overview: This tutorial walks through distributed data parallel training in PyTorch via DDP. We will start with a simple non-distributed training job, and end with deploying a training job across several GPUs in a single HAL node. Along the way, you will learn about DDP to accelerate your model training. You will also learn how to monitor GPU status to help profile code performance to fully utilize GPU computing power.
- Instructor: Shirui Luo
- Date: February 15, 2023
Robust Physics-Informed Neural Networks
Overview: Physics Informed Neural Networks (PINNs) have recently been found to be effective PDE solvers. This talk will focus on how traditional PINN architectures along with physics-inspired regularizers fail to retrieve the intended solution when training data is noisy and how this problem can be solved using Gaussian Process based smoothing techniques.
- Training Slides
- Instructor: Avik Roy
- Date: February 22, 2023
How to use Pretrained Models
Overview: There are several popular AI model repositories that provide access to pre-trained models via easy-to-use APIs. Hugging Face is one of the latest such repositories that hosts a number of very recent models, such as Facebook’s OPT and OpenAI’s GPT models, as well as many datasets. This tutorial will show how to use basic pre-trained PyTorch models and then will introduce how to get started with using models from the Hugging Face ecosystem.
- Training Slides
- Instructor: Priyam Mazumdar
- Date: March 8, 2023
DRYML an open source meta-library for machine learning and more
Overview: DRYML aims to empower the Machine Learning practitioner to spend less time writing boilerplate code, and more time implementing new techniques. DRYML provides a model serialization framework along with serialization implementation for many common ML frameworks and model types, a framework for defining and training models on a specific problem, and a system to compare models from different ML frameworks on the same footing. Students will leverage their knowledge across multiple ML frameworks such as tensorflow, pytorch, and sklearn to build new models, optimize them, and compare them across ML frameworks.
- Github Repo
- Training Slides
- Instructor: Matthew Krafczyk
- Date: March 22, 2023
Explainability of Deep Neural Networks
Overview: Deep Neural Networks are often treated as black boxes. This talk will focus on some of the modern methods of explainability for DNNs and discuss their implementation, usage, and limitations.
- Training Slides
- Instructor: Avik Roy
- Date: March 29, 2023
Introduction to Transformer models
Overview: In this tutorial, you’ll gain an understanding of how to developing and training Transformer models. We’ll compare with Sequence models and explore how to construct and utilize a Vision Transformer. You’ll learn about essential components such as PatchEmbeddings, Attention Layers, Class tokens, and Positional Embeddings, which play crucial roles in the architecture of the Transformer. By the end of this tutorial, you’ll have a foundation to build and train your Transformer models effectively.
- Training Slides
- Instructor: Priyam Mazumdar – MS Student
- Date: April 12, 2023
Weights & Biases workshop: Track, Visualize, and Improve Your Experiments
Overview: Andrea will take you through a W&B introduction and product walkthrough, including Experiment tracking, W&B Tables, Sweeps, Artifacts, Dashboards/Reports, and Integrations! Followed by a Colab classification competition in Kaggle with swag for top submissions.
University of Illinois Only. Link to Recording
- Presentation
- Weights & Biases student and research-focused resources on their Academic Page!
- Instructor: Andrea Parker
- Date: April 19, 2023
Introduction to Machine Learning and Neural Networks
Overview: This tutorial will provide an easy to follow introduction to machine learning with neural networks. We will go over various examples from simple dense networks to convolutional network architectures using TensorFlow framework on HAL system.
- GitHub Notebook
- Instructor: Aaron Saxton, Data Engineer, NCSA
- Date: September 14, 2022
Introduction to PyTorch
Overview: This tutorial will introduce basics of PyTorch framework necessary to build a neural network, train it and evaluate the accuracy of the model.
- Training Notebook
- Instructor: Priyam Mazumdar, Graduate Student, NCSA
- Date: September 28, 2022
Data Loading and Tools in PyTorch and TensorFlow
Overview: This tutorial will demonstrate how to use data loaders provided with PyTorch and TensorFlow and how to develop application-specific data loaders. Other topics will include how to make use of different tools, such as Weight&Biases, to help with model development and training.
- Instructor: William Eustis, Undergraduate Student, NCSA
- Date: October 5, 2022
Machine Learning First Deep Neural Network
Overview: Everyone talks about Deep Learning and using Neural Networks. The algorithms have been found in certain cases to produce better estimates than humans! They are great for data prediction and understanding more about your data. This workshop is an introduction to Deep Learning using Tensorflow and Keras. A Jupyter notebook is used for hands-on student exercises along with a set of lecture slides.
- Training Github Notebook
- Instructor: Kristopher Keipert, NVIDIA
- Date: March 2, 2022
Convolution Neural Network Models
Overview: Convolutional Neural Networks (CNNs) are used primarily for anything involving images – image classification, to image segmentation, to putting bounding boxes around an object. This interactive workshop includes a Jupyter notebook using TensorFlow and Keras, along with a set of lecture slides. In addition, this Lecture introduces the concept of Inference. This is processing data through a trained model without any back propagation.
- Instructor: Jeff Layton, NVIDIA
- Date: March 9, 2022
Machine Learning with H2O on HAL
Overview: This tutorial will cover basic machine learning techniques, such as Linear Regression, Decision Tree, Support Vector Machine, Naive Bayes, K-Nearest Neighbors, K-Means, and Random Forest, using H2O machine learning platform on HAL system.
Instructor: Dawei Mu
Date: March 23, 2022
Data Loaders
Overview: This tutorial will demonstrate how to use data loaders provided with PyTorch and TensorFlow and how to develop application-specific data loaders.
- Training items:
- Instructor: William Eustis
- Date: March 30, 2022
Graph Neural Networks (GNNs)
Overview: This tutorial will introduce basic concepts in Graph Neural Networks (GNNs) and will show how to use PyTorch Geometric Library to start implementing GNNs.
- Instructors: Minyang Tian & Michael Volk
- Date: April 6, 2022
Inference Optimization with NVIDIA TensorRT
Overview: In many applications of deep learning models, we would benefit from reduced latency (time taken for inference). This tutorial will introduce NVIDIA TensorRT, an SDK for high-performance deep learning inference. We will go through all the steps necessary to convert a trained deep learning model to an inference-optimized model on HAL.
- Instructors: Nikil Ravi & Pranshu Chaturvedi
- Date: April 13, 2022
Hands-On Deep Learning for Computer Vision
Overview: This tutorial will introduce how machine learning can be accomplished with neural networks and will go over various examples from simple dense networks to convolutional network architectures using TensorFlow on HAL system.
- Training Instructions
- Github Repository Notebook
- Instructor: Asad Khan
- Date: September 15, 2021
Distributed Deep Learning on HAL
Overview: This tutorial will explain how to use multiple GPUs for accelerating deep learning on HAL using frameworks such as TensorFlow or PyTorch.
- Training Slides
- Instructor: Dawei Mu
- Date; October 20, 2021
INTRO TO NATURAL LANGUAGE PROCESSING (NLP) WITH PYTORCH
Overview: This tutorial will introduce neural network architectures for processing natural language texts. We will cover Recurrent Neural Networks (RNNs) and Generative Neural Networks (GNNs), attention mechanisms, and how to build text classification models.
- Instructor: Volodymyr Kindratenko
- Date: October 27, 2021
OPENVINO™ TOOLKIT INTEGRATION WITH TENSORFLOW: HANDS-ON PRACTICE INTEL DEVCLOUD
Overview: This tutorial will introduce Tensorflow integration available with the OpenVINO™ toolkit and hands-on practice in the Intel® DevCloud explaining the minimal code changes needed with Tensorflow to gain greater inference performance with OpenVINO. Suggested to pre-register for access to the Intel® DevCloud here. Access is free.
- DevCloud Example Instructions
- GitHub Repository Notebook: OpenVINO TensorFlow Classification Example
- GitHub Repository Notebook: OpenVINO TensorFlow Object Detection Example
- Instructor: Kumar Vishwesh and Yamini Mimmagadda
- Date: November 3, 2021
ROBUST PHYSICS INFORMED NEURAL NETWORKS
Overview: Physics Informed Neural Networks (PINNs) have recently been found to be effective PDE solvers. This talk will focus on how traditional PINN architectures along with physics-inspired regularizers fail to retrieve the intended solution when training data is noisy and how this problem can be solved using Gaussian Process based smoothing techniques.
- Training Slides
- arxiv 2110.13330
- GitHub Repository Notebook
- Instructor: Avik Roy
- Date: November 10, 2021
PHYSICS INFORMED DEEP LEARNING
Overview; This tutorial will explore how to incorporate physics into deep learning models with various examples ranging from using physics informed neural networks (PINNs) for forward and inverse problems to employing physics informed DeepONets for a hybrid data and physics approach to problems.
- Training Slides
- GitHub Repository Notebook
- Instructor: Shawn Rosofsky
- Date: November 17, 2021