Non quia difficilia sunt non audemus, sed quia non audemus difficilia sunt
Home -> Teaching -> MPI Tutorials -> PPoPP'13
Home
  Publications
  Awards
  Research
  Teaching
    
MPI Tutorials
      
CSCS'12
      ISC'12
      Speedup'15
      PPoPP'13
      ICS'13
      EuroMPI'14
      ISC'16
    CS498

  Miscellaneous
  Full CV [pdf]
  BLOG






  Events








  Past Events





Advanced MPI and MPI for Dummies at PPoPP'13

MPI for Dummies

Presenters

  Pavan Balaji, Argonne National Laboratory
  Torsten Hoefler, ETH Zurich

Host

PPoPP'13

Slides

All slides are available for download.

Abstract

The Message Passing Interface (MPI) has been the de facto standard for parallel programming for nearly two decades now, and knowledge of MPI is considered a pre-requisite for most people aiming for a career in parallel programming. This is a beginner-level tutorial aimed at introducing parallel programming with MPI. This tutorial will provide an overview of MPI, its offered features, current implementations of MPI, and its suitability for parallel computing environments. Together with a brief overview of MPI and its features the tutorial will also discuss good programming practices and issues to watch out for in MPI programming. Finally, several application case studies, including examples in nuclear physics, combustion, and quantum chemistry applications, and how they use MPI, will be shown.

Detailed Description

Tutorial Goals

MPI is widely recognized as the de facto standard for parallel programming. Even though knowledge of MPI is increasingly becoming prerequisite for researchers and developers involved in parallel programming institutes, including universities, research labs, and the industry, very few institutes offer formal training in MPI. The goal of this tutorial is to educate users with basic programming knowledge in MPI and equip them to the capability to get started with MPI programming. Based on these emerging trends and the associated challenges, the goals of this tutorial are:

  • Making the attendees familiar with MPI programming and its associated benefits
  • Providing an overview of available MPI implementations and the status of their capabilities with respect to the MPI standard
  • Illustrating MPI usage models from various example application domains including nuclear physics, computational chemistry, and combustion.

Targeted audience

This tutorial is targeted for various categories of people working in the areas of high performance communication and I/O, storage, networking, middleware, programming models, and applications related to high-end systems. Specific audience this tutorial is aimed at include:

  • Newcomers to the field of distributed memory programming models who are interested in familiarizing themselves with MPI
  • Managers and administrators responsible for setting-up next generation high-end systems, and facilities in their organizations/laboratories
  • Scientists, engineers, and researchers working on the design and development of next generation high-end systems including clusters, data centers, and storage centers
  • System administrators of large-scale clusters
  • Developers of next generation parallel middleware and applications.

Advanced Parallel Programming with MPI-1, MPI-2, and MPI-3

Presenters

  Pavan Balaji, Argonne National Laboratory
  Torsten Hoefler, ETH Zurich

Host

PPoPP'13

Slides

All slides are available for download.

Abstract

The Message Passing Interface (MPI) has been the de facto standard for parallel programming for nearly two decades now. However, a vast majority of applications only rely on basic MPI-1 features without taking advantage of the rich set of functionality the rest of the standard provides. Further, with the advent of MPI-3 (to be released September 2012), a vast number of new features are being introduced in MPI, including efficient one-sided communication, support for external tools, non-blocking collective operations, and improved support for topology-aware data movement. This is an advanced-level tutorial that will provide an overview of various powerful features in MPI, especially with MPI-2 and MPI-3.

Detailed Description

Tutorial Goals

MPI is widely recognized as the de facto standard for parallel programming. Even though knowledge of MPI is increasingly becoming a prerequisite for researchers and developers involved in parallel programming institutes, including universities, research labs, and the industry, very few institutes offer formal training in MPI. The goal of this tutorial is to educate users with advanced programming knowledge in MPI and equip them with the knowledge of powerful techniques present in various MPI versions including the upcoming MPI-3 standard. Based on these emerging trends and the associated challenges, the goals of this tutorial are:

  • Providing an overview of current large-scale applications and data movement efficiency issues they are facing
  • Providing an overview of the advanced powerful features available in MPI-2 and upcoming in MPI-3
  • Illustrating how scientists, researchers and developers can use these features to design new applications

Targeted audience

This tutorial is targeted for various categories of people working in the areas of high performance communication and I/O, storage, networking, middleware, programming models, and applications related to high-end systems. Specific audience this tutorial is aimed at include:

  • Scientists, engineers, and researchers working on the design and development of next generation high-end systems including clusters, data centers, storage centers
  • System administrators of large-scale clusters
  • Developers of next generation middleware and applications


serving: 3.128.94.112:1660© Torsten Hoefler