Skip to Main Content
UCLA Logo Institute for Digital Research and Education

HPC

Parallel Parallel Computing 3: Introduction to PETSc

MPI (Message Passing Interface) is a standardized interface for writing portable distributed-memory parallel scientific code. The portability of MPI ensures the same MPI program works the same way on different…

Parallel Parallel Computing 1: Running MPI programs

MPI (Message Passing Interface) is a standardized interface for writing portable distributed-memory parallel scientific code. The portability of MPI ensures the same MPI program works the same way on different…

Big Data on HPC

This workshop will go over using Big Data techniques on HPC resources. Big Data methods are used when data size because so large, it becomes challenging to compute. Also, when…

Hoffman2 Happy Hour: RStudio on Hoffman2

The Hoffman2 Happy Hours Series: Using Rstudio on Hoffman2 is best with a Rstudio Server app. This Happy Hour shows an example of using a Rstudio Server with Singularity/Apptainer. Any…

Hoffman2 Happy Hour: Anaconda on Hoffman2

Anaconda (https://www.anaconda.com/) is a distribution of R and Python that can be used to easily install many popular data science, biostats, and other packages. This Hoffman2 Happy Hour will go…

Building Containers for HPC

This workshop is a follow-up to a previous workshop on “Using Containers on HPC Resources” (https://github.com/ucla/hpc_workshops). This workshop will focus on creating containers to be ran on HPC resources. We…