Skip to content

Optimization of various matrix operations for different hardware- 1) Hierarchical memory (registers, cache, virtual memory), 2) Instruction level parallelism, 3) Multicore processors, 4) Shared memory parallelism, 5) GPU (CUDA), 6) Distributed memory parallelism (MPI). Code developed as a part of amath 583 course at University of Washington.

Notifications You must be signed in to change notification settings

abs711/High-Performance-Scientific-Computing

About

Optimization of various matrix operations for different hardware- 1) Hierarchical memory (registers, cache, virtual memory), 2) Instruction level parallelism, 3) Multicore processors, 4) Shared memory parallelism, 5) GPU (CUDA), 6) Distributed memory parallelism (MPI). Code developed as a part of amath 583 course at University of Washington.

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages