This workshop series offers an in-depth exploration of massively parallel programming with Graphics Processing Units (GPUs). Participants will gain a comprehensive understanding of parallel computing concepts, the distinctions between CPU…
Read More
MPI
Parallel Parallel Computing 4: Solving Partial Differential Equations on Parallel Computers
MPI (Message Passing Interface) is a standardized interface for writing portable distributed-memory parallel scientific code. The portability of MPI ensures the same MPI program works the same way on different…
Read More
Parallel Parallel Computing 3: Introduction to PETSc
MPI (Message Passing Interface) is a standardized interface for writing portable distributed-memory parallel scientific code. The portability of MPI ensures the same MPI program works the same way on different…
Read More
Parallel Parallel Computing 1: Running MPI programs
MPI (Message Passing Interface) is a standardized interface for writing portable distributed-memory parallel scientific code. The portability of MPI ensures the same MPI program works the same way on different…
Read More
Practical Parallel Computing 3: Introduction to PETSc
MPI (Message Passing Interface) is a standardized interface for portable distributed-memory scientific parallel computing. The portability ensures that a properly-written, standard-conforming MPI program can work the same way on different…
Read More
Practical Parallel Computing 2: MPI Programming
MPI (Message Passing Interface) is a standardized interface for portable distributed-memory scientific parallel computing. The portability ensures that a properly-written, standard-conforming MPI program can work the same way on different…
Read More
Practical Parallel Computing 1: Running MPI Programs
MPI (Message Passing Interface) is a standardized interface for portable distributed-memory scientific parallel computing. The portability ensures that a properly-written, standard-conforming MPI program can work the same way on different…
Read More
Parallel Computing using MPI and Julia
This workshop will discuss the basics of MPI (message passing interface), and several distributed-memory parallel numerical computing examples using MPI. The Julia language will be used in the examples. It would be straightforward to translate the Julia examples into other languages, such as C and Fortran, if desired.