Parallel
Computing
Outline
-
What is parallel programming
-
Parallel computer
-
Flynn's classic taxonomy
-
Heterogeneous computing
-
-
Parallel programming model
-
Shared memory model
-
Distributed memory model
-
Hybrid model
-
Outline
-
What is parallel programming
-
Parallel computer
-
Flynn's classic taxonomy
-
Heterogeneous computing
-
-
Parallel programming model
-
Shared memory model
-
Distributed memory model
-
Hybrid model
-
-
What is parallel programming
Sequential Program
-
The single program is computed by one processor
-
Deal the instruction one after another
-
Only one instructin may execute at any moment
Parallel Program
-
Break a program into several parts
-
Instructions from each part execute simultanously
Why parallel programming
-
Advantages
Higher performance: save larger problem
Better resource utilization: taking advantage of multi-core processors
-
Disadvantages
Harder to program
Harder to debug
Not all problem can be parallelized efficiency ( dependency )
Parallel programs & application
-
Scientific applications
-
Computer animations
-
Computer games
-
Image processing
-
Data mining
Outline
-
What is parallel programming
-
Parallel computer
-
Flynn's classic taxonomy
-
Heterogeneous computing
-
-
Parallel programming model
-
Shared memory model
-
Distributed memory model
-
Hybrid model
-
-
Parallel computer
-
Flynn's classic taxonomy
-
Heterogeneous computing
-
Paralle computer classification
-
Flynn's Classical Taxonomy
-
Processing unit, instruction, data
-
SISD
-
MISD
-
SIMD
-
MIMD
SISD
-
Single Instruction, Single Data (SISD)
-
A serial (non-parallel) computer
-
Executes a single instruction stream, to operate on data stored in a single memory
-
Example: old mainframes,
single-core processor
SIMD
-
Single Instruction, Multiple Data (SIMD)
-
Multiple processing elements that perform the same operation on multiple data points concurrency
-
Example: GPU
vector pipelines computer
MISD
-
Multiple Instruction, Single Data (SIMD)
-
Many functional units perform different operations on the same data
-
Fault-tolerant computers execute
the same instructions to detect
and mask errors -
Example:
space shuttle
MIMD
-
Multiple Instruction, Multiple Data (MIMD)
-
At any time, different processors may be executing different instructions on different data
-
Example:
Most modern computers,
multi-core PCs
supercomputer
cluster
Heterogenous Computing
-
Heterogeneous computing is an integrated system that consists of different types of (programmable) computing units
-
DSP (digital signal processor)
-
FPGA (field-programmable gate array)
-
ASIC (application-specific integrated circuit)
-
GPU (graphics processing unit)
-
Co-processor (Intel Xeon Phi)
-
CPU v.s GPU
-
CPU is latency oriented design, can do lots of sophisticated control
-
GPU is throughput oriented design, long latency but heavily pipeline for high throughput
Latency v.s Thoughtput
Performance
Trend
Outline
-
What is parallel programming
-
Parallel computer
-
Flynn's classic taxonomy
-
Heterogeneous computing
-
-
Parallel programming model
-
Shared memory model
-
Distributed memory model
-
Hybrid model
-
-
Parallel programming model
Shared memory model
Distributed memory model
Hybrid model
Shared Memory Model
-
Memory can be simultaneously access by multiple process with an intent to provide communication among them or avoid redundant data copies
Shared Memory/Thread Model
-
A single process can have multiple, concurrent execution paths
-
Threads have local data, but also, shares resources
-
Threads commucnication through global memory
-
Thread can come and go, but the main program remains
-
to provide the necessary shared resources until the application has complete
-
Shared Memory/Thread Model
Shared Memory/Thread Model
-
Implementation methodology
-
A library of subroutines called from parallel source code
e.g: POSIX Thread (Pthread)
-
-
A set of compiler directives embedded in either serial or parallel source code
-
e.g: OpenMP
-
Shared Memory/Thread Model
-
Important issues
-
Race Condition: A situation where the computing output depending on the sequence order of process executions
-
Deadlock: Two or more competing action are waiting for the other to finish
-
Distributed Memory/MPI Model
-
A set of tasks that use their own local memory during computaion
-
Tasks exchange data through communications by sending and receive messages
-
Memory copy
-
-
Implementation: MPI
-
An API specification that allows computers to communicate by means send, receive, broadcast ... etc
-
Distributed Memory/MPI Model
Distributed Memory/MPI Model
-
Important issues
-
Synchronization
programmer should make sure the correctness of timing dependency between processes -
Communication time
Network speed is much slower than CPU speed
Network latency causes a constant delay time
-
Hybrid Parallel computing Model
-
Combine both shared & distributed memory
-
MPI + pthread/OpenMP
-
Implement parallelism with MPI libraries among nodes
-
Implement parallelism with pthread/OpenMP libraries within each node
-
Q & A
Parallel
By zlsh80826
Parallel
- 784