Rui's Blog
  • Rui's Blog/Paper Reading Notes - Introduction
  • Personal Blog
    • Personal Blog - Index
      • How to Create Picture-in-Picture Effect / Video Overlay for a Presentation Video
      • How to Do Your Part to Protect the Environment in Wisconsin
      • How to Get a Driver's License in Wisconsin
      • How to Travel from the U.S. to China onboard AA127 in June 2021
      • How to Transfer Credits Back to UW-Madison
      • Resources on Learning Academic Writing (for Computer Science)
    • Towards applying to CS Ph.D. programs
  • Machine Learning Systems
    • Machine Learning Systems - Index
      • MLSys Papers - Short Notes
      • [2011 NSDI] Dominant Resource Fairness: Fair Allocation of Multiple Resource Types
      • [2014 OSDI] Scaling Distributed Machine Learning with the Parameter Server
      • [2018 OSDI] Gandiva: Introspective Cluster Scheduling for Deep Learning
      • [2018 SIGCOMM] Chameleon: Scalable Adaptation of Video Analytics via Temporal and Cross-camera ...
      • [2018 NIPS] Dynamic Space-Time Scheduling for GPU Inference
      • [2019 ATC] Analysis of Large-Scale Multi-Tenant GPU Clusters for DNN Training Workloads
      • [2019 NSDI] Tiresias: A GPU Cluster Manager for Distributed Deep Learning
      • [2019 SOSP] ByteScheduler: A Generic Communication Scheduler for Distributed DNN Training ...
      • [2019 SOSP] PipeDream: Generalized Pipeline Parallelism for DNN Training
      • [2019 SOSP] Parity Models: Erasure-Coded Resilience for Prediction Serving Systems
      • [2019 NIPS] GPipe: Efficient Training of Giant Neural Networks using Pipeline Parallelism
      • [2019 SC] ZeRO: memory optimizations toward training trillion parameter models
      • [2020 OSDI] Gavel: Heterogeneity-Aware Cluster Scheduling Policies for Deep Learning Workloads
      • [2020 OSDI] AntMan: Dynamic Scaling on GPU Clusters for Deep Learning
      • [2020 OSDI] BytePS: A High Performance and Generic Framework for Distributed DNN Training
      • [2020 SIGCOMM] Reducto: On-Camera Filtering for Resource-Efficient Real-Time Video Analytics
        • [2020 MLSys] Salus: Fine-Grained GPU Sharing Primitives for Deep Learning Applications
      • [2020 EuroSys] AlloX: Compute Allocation in Hybrid Clusters
      • [2020 VLDB] PyTorch Distributed: Experiences on Accelerating Data Parallel Training
      • [2020 NetAI] Is Network the Bottleneck of Distributed Training?
      • [2020 NSDI] Themis: Fair and Efficient GPU Cluster Scheduling
      • [2021 MLSys] Accordion: Adaptive Gradient Communication via Critical Learning Regime Identification
      • [2021 VLDB] Analyzing and Mitigating Data Stalls in DNN Training
      • [2021 FAST] CheckFreq: Frequent, Fine-Grained DNN Checkpointing
      • [2021 EuroMLSys] Interference-Aware Scheduling for Inference Serving
      • [2021 OSDI] Pollux: Co-adaptive Cluster Scheduling for Goodput-Optimized Deep Learning
      • [2021 MLSys] Wavelet: Efficient DNN Training with Tick-Tock Scheduling
      • [2021 NSDI] SwitchML: Scaling Distributed Machine Learning with In-Network Aggregation
    • Big Data Systems - Index
      • Big Data Systems Papers - Short Notes
      • [2003 SOSP] The Google File System
      • [2004 OSDI] MapReduce: Simplified Data Processing on Large Clusters
      • [2010 SIGMOD] Pregel: A System for Large-Scale Graph Processing
      • [2011 NSDI] Mesos: A Platform for Fine-Grained Resource Sharing in the Data Center
      • [2012 NSDI] Resilient Distributed Datasets: A Fault-Tolerant Abstraction for In-Memory Cluster ...
      • [2012 OSDI] PowerGraph: Distributed Graph-Parallel Computation on Natural Graphs
      • [2019 FAST] DistCache: Provable Load Balancing for Large-Scale Storage Systems with Distributed...
      • [2021 HotOS] From Cloud Computing to Sky Computing
      • [2021 EuroSys] NextDoor: Accelerating graph sampling for graph machine learning using GPUs
  • Earlier Readings & Notes
    • High Performance Computing Course Notes
      • Lecture 1: Course Overview
      • Lecture 2: From Code to Instructions. The FDX Cycle. Instruction Level Parallelism.
      • Lecture 3: Superscalar architectures. Measuring Computer Performance. Memory Aspects.
      • Lecture 4: The memory hierarchy. Caches.
      • Lecture 5: Caches, wrap up. Virtual Memory.
      • Lecture 6: The Walls to Sequential Computing. Moore’s Law.
      • Lecture 7: Parallel Computing. Flynn's Taxonomy. Amdahl's Law.
      • Lecture 8: GPU Computing Intro. The CUDA Programming Model. CUDA Execution Configuration.
      • Lecture 9: GPU Memory Spaces
      • Lecture 10: GPU Scheduling Issues.
      • Lecture 11: Execution Divergence. Control Flow in CUDA. CUDA Shared Memory Issues.
      • Lecture 12: Global Memory Access Patterns and Implications.
      • Lecture 13: Atomic operations in CUDA. GPU ode optimization rules of thumb.
      • Lecture 14: CUDA Case Studies. (1) 1D Stencil Operation. (2) Vector Reduction in CUDA.
      • Lecture 15: CUDA Case Studies. (3) Parallel Prefix Scan on the GPU. Using Multiple Streams in CUDA.
      • Lecture 16: Streams, and overlapping data copy with execution.
      • Lecture 17: GPU Computing: Advanced Features.
      • Lecture 18: GPU Computing with thrust and cub.
      • Lecture 19: Hardware aspects relevant in multi-core, shared memory parallel computing.
      • Lecture 20: Multi-core Parallel Computing with OpenMP. Parallel Regions.
      • Lecture 21: OpenMP Work Sharing.
      • Lecture 22: OpenMP Work Sharing
      • Lecture 23: OpenMP NUMA Aspects. Caching and OpenMP.
      • Lecture 24: Critical Thinking. Code Optimization Aspects.
      • Lecture 25: Computing with Supercomputers.
      • Lecture 26: MPI Parallel Programming General Introduction. Point-to-Point Communication.
      • Lecture 27: MPI Parallel Programming Point-to-Point communication: Blocking vs. Non-blocking sends.
      • Lecture 28: MPI Parallel Programming: MPI Collectives. Overview of topics covered in the class.
    • Cloud Computing Course Notes
      • 1.1 Introduction to Clouds, MapReduce
      • 1.2 Gossip, Membership, and Grids
      • 1.3 P2P Systems
      • 1.4 Key-Value Stores, Time, and Ordering
      • 1.5 Classical Distributed Algorithms
      • 4.1 Spark, Hortonworks, HDFS, CAP
      • 4.2 Large Scale Data Storage
    • Operating Systems Papers - Index
      • CS 736 @ UW-Madison Fall 2020 Reading List
      • All File Systems Are Not Created Equal: On the Complexity of Crafting Crash-Consistent Applications
      • ARC: A Self-Tuning, Low Overhead Replacement Cache
      • A File is Not a File: Understanding the I/O Behavior of Apple Desktop Applications
      • Biscuit: The benefits and costs of writing a POSIX kernel in a high-level language
      • Data Domain: Avoiding the Disk Bottleneck in the Data Domain Deduplication File System
      • Disco: Running Commodity Operating Systems on Scalable Multiprocessors
      • FFS: A Fast File System for UNIX
      • From WiscKey to Bourbon: A Learned Index for Log-Structured Merge Trees
      • LegoOS: A Disseminated, Distributed OS for Hardware Resource Disaggregation
      • LFS: The Design and Implementation of a Log-Structured File System
      • Lottery Scheduling: Flexible Proportional-Share Resource Management
      • Memory Resource Management in VMware ESX Server
      • Monotasks: Architecting for Performance Clarity in Data Analytics Frameworks
      • NFS: Sun's Network File System
      • OptFS: Optimistic Crash Consistency
      • RAID: A Case for Redundant Arrays of Inexpensive Disks
      • RDP: Row-Diagonal Parity for Double Disk Failure Correction
      • Resource Containers: A New Facility for Resource Management in Server Systems
      • ReVirt: Enabling Intrusion Analysis through Virtual-Machine Logging and Replay
      • Scheduler Activations: Effective Kernel Support for the User-Level Management of Parallelism
      • SnapMirror: File-System-Based Asynchronous Mirroring for Disaster Recovery
      • The Linux Scheduler: a Decade of Wasted Cores
      • The Unwritten Contract of Solid State Drives
      • Venti: A New Approach to Archival Storage
    • Earlier Notes
      • How to read a paper
  • FIXME
    • Template for Paper Reading Notes
Powered by GitBook
On this page
  • Acknowledgments
  • Table of Contents

Was this helpful?

  1. Earlier Readings & Notes

High Performance Computing Course Notes

ECE/ME/EMA/CS 759: High Performance Computing for Engineering Applications, Spring 2021 by Prof. Dan Negrut

Previous[2021 EuroSys] NextDoor: Accelerating graph sampling for graph machine learning using GPUsNextLecture 1: Course Overview

Last updated 4 years ago

Was this helpful?

Acknowledgments

  • All slides/files linked are accessible on Box using a UW-Madison account

  • Almost every figure and piece of code in these notes is excerpted from Prof. Dan Negrut's course slides. Some of the slides are taken from other places by Prof. Negrut -- he cited those in his slides.

Table of Contents

Date

Title

Recommended Readings

1/25

1/27

1/29

2/1

Build mgmt & cmake (ME459 p354-)

2/3

2/5

2/8

2/10

2/12

2/15

2/17

2/19

2/22

2/24

2/26

3/1

3/3

3/5

3/8

3/10

Cache Coherence on Power 9 - Volta systems w/ NVLINK2

3/12

3/15

3/17

3/19

3/22

3/24

3/26

3/29

; Slurm usage (ME459 p95-97)

C recap (ME459 p114-);

gdb recap (ME459 p649-); Ch.5 of the

Git (ME459 p449-);

;

Slides for ME759 (of the whole semester)
Slides from ME459 (Computing Concepts for Applications in Engineering)
Lecture 1: Course Overview
Basic Linux Command Line Usage
Lecture 2: From Code to Instructions. The FDX Cycle. Instruction Level Parallelism.
Euler usage
Lecture 3: Superscalar architectures. Measuring Computer Performance. Memory Aspects.
C book
Lecture 4: The memory hierarchy. Caches.
Lecture 5: Caches, wrap up. Virtual Memory.
How to Write a Git Commit
Lecture 6: The Walls to Sequential Computing. Moore’s Law.
Validity of the single processor approach to achieving large scale computing capabilities (Amdahl, '67)
Lecture 7: Parallel Computing. Flynn’s Taxonomy. Amdahl’s Law.
Structured Programming w/ go to Statements (Knuth, '74)
Lecture 8: GPU Computing Intro. The CUDA Programming Model. CUDA Execution Configuration
Modern Microprocessors: A 90-Minute Guide (Patterson, '01)
Lecture 9: GPU Memory Spaces.
Optimizations in C++ Compilers (Godbolt, 2019)
Lecture 10: GPU Scheduling Issues.
NVIDIA Tesla Architecture
Lecture 11: Execution Divergence. Control Flow in CUDA. CUDA Shared Memory Issues.
CUDA C++ Programming Guide
Lecture 12: Global Memory Access Patterns and Implications.
The GPU Computing Era (Nickolls & Dally, '10)
Lecture 13: Atomic operations in CUDA. GPU ode optimization rules of thumb.
Unified Memory in CUDA 6: A Brief Overview
Lecture 14: CUDA Case Studies. (1) 1D Stencil Operation. (2) Vector Reduction in CUDA
Maximizing Unified Memory Performance in CUDA (Sakharnykh, '17)
Lecture 15: CUDA Case Studies. (3) Parallel Prefix Scans on the GPU. Using Multiple Streams in CUDA.
Titles of GTC '21 Talks
Lecture 16: Streams, and overlapping data copy with execution.
Dissecting the NVIDIA Volta GPU Architecture via Microbenchmarking (Citadel, '18)
CUDA C++ Best Practices Guide
Lecture 17: GPU Computing: Advanced Features.
GTC '18 Talk on Unified Memory
Lecture 18: GPU Computing with thrust and cub.
Thrust: A Productivity-Oriented Library for CUDA (Bell & Hoberock, '11)
Lecture 19: Hardware aspects relevant in multi-core, shared memory parallel computing.
Unified Memory in CUDA 6: A Brief Overview and Related Data Access/Transfer Issues (by Dan and some other guys! '14)
Lecture 20: Multi-core Parallel Computing with OpenMP. Parallel Regions.
Lecture 21: OpenMP Work Sharing.
Node-Level Performance Engineering (SC '19)
Lecture 22: OpenMP Work Sharing.
Advanced OpenMP: Performance and 5.0 Features (SC '19)
Lecture 23: OpenMP NUMA Aspects. Caching and OpenMP.
Mastering Tasking with OpenMP (SC '19)
Lecture 24: Critical Thinking. Code Optimization Aspects.
Ch. 12 of Optimizing Software in C++
Lecture 25: Computing with Supercomputers.
Lecture 26: MPI Parallel Programming General Introduction. Point-to-Point Communication.
HPC Perspectives (Dongarra et. al., '05)
Lecture 27: MPI Parallel Programming Point-to-Point communication: Blocking vs. Non-blocking sends.
Advanced MPI Programming (SC '19)
Lecture 28: MPI Parallel Programming: MPI Collectives. Overview of topics covered in the class.