Skip to content

Commit

Permalink
Add Test
Browse files Browse the repository at this point in the history
  • Loading branch information
lemoinep committed Sep 20, 2023
1 parent df97fe1 commit 0841ef2
Show file tree
Hide file tree
Showing 4 changed files with 41 additions and 3,674 deletions.
20 changes: 7 additions & 13 deletions docs/modules/ROOT/pages/PPChapter1.adoc
Original file line number Diff line number Diff line change
Expand Up @@ -51,14 +51,8 @@ They are few in number but very quick to access. They
are used to store variables, the intermediate results of operations
(arithmetic or logical) or processor control information.



image::image1.png[xref=#fragment,width=322,height=220]



image::Session1_ParallelProgramming_Introduction.pdf[xref=#fragment,page=1]

image::image1.png[xref=#fragment1,width=322,height=220]
//image::../assets/images/image1.png[xref=#img1,width=322,height=220]


The register structure varies from processor to processor. This is why
Expand Down Expand Up @@ -158,7 +152,7 @@ application.

Explicit code targeting GPUs: CUDA, HIP, SYCL, Kokkos, RAJA,...

image:image2.png[xref=#fragment,width=488,height=342]
image:image2.png[xref=#fragment2,width=488,height=342]

_Fig: illustrates the main hardware architecture differences between
CPUs and GPUs. The transistor counts associated with various functions
Expand Down Expand Up @@ -235,7 +229,7 @@ GPUs can be 10+X faster than GPUs for parallel code.

*1.3 GPGPU ( General-Purpose Graphics Processing Unit)*

image:image4.png[xref=#fragment,width=642,height=331]
image:image4.png[xref=#fragment4,width=642,height=331]

A *General-Purpose Graphics Processing Unit* (GPGPU) is a graphics
processing unit (GPU) that is programmed for purposes beyond graphics
Expand Down Expand Up @@ -310,7 +304,7 @@ mapping
Such an architecture is said to be "throughput-oriented". The latest
from the Santa-Clara firm, codenamed “Fermi” has 512 cores.

image:image5.png[xref=#fragment,width=530,height=241]
image:image5.png[xref=#fragment5,width=530,height=241]

_CPU architecture vs. GPUs_

Expand Down Expand Up @@ -663,7 +657,7 @@ New containers are continually being added to the hub.
=== AMD Fusion System Architecture
=== Moves to Unify CPUs and GPUs

image:image6.png[xref=#fragment,width=511,height=287]
image:image6.png[xref=#fragment6,width=511,height=287]


*1.7 TPU (Tensor Processing Unit) form Google*
Expand Down Expand Up @@ -702,7 +696,7 @@ at ultra-high speeds but must be paired with a CPU to give and execute
instructions.


image:image22.png[xref=#fragment,width=544,height=419]
image:image22.png[xref=#fragment22,width=544,height=419]


*Applications for TPUs*
Expand Down
10 changes: 6 additions & 4 deletions docs/modules/ROOT/pages/PPChapter2.adoc
Original file line number Diff line number Diff line change
Expand Up @@ -20,8 +20,7 @@ Note: on a cluster of independent shared-memory multiprocessor machines
in the same program can be a major advantage for the parallel
performance of the code.

image:../assets/images/image7.png[image,width=581,height=336]

image:image7.png[xref=#fragment7,width=581,height=336]


[width="100%",cols="50%,50%",]
Expand Down Expand Up @@ -2011,7 +2010,9 @@ pure MPI scalability runs out.

A common hybrid approach

image:../assets/images/image9.png[image,width=307,height=155]
image:image9.png[xref=#fragment9,width=307,height=155]



* From dequential code, alongside MPI first, then try adding OpenMP
* From MPI code, add OpenMP
Expand All @@ -2021,4 +2022,5 @@ parallel region and allow only the master thread to communicate between
MPI tasks.
* Could use MPI in parallel region with thread-safe MPI.

image:../assets/images/image10.png[image,width=264,height=166]
image:image10.png[xref=#fragment10,width=264,height=166]

Loading

0 comments on commit 0841ef2

Please sign in to comment.