Parallel Architectures Flashcards

1
Q

What are three classifications of parallel computing?

A

Low-level, single instruction multiple data (SIMD) and multiple instructions multiple data

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Define low-level parallelism

A

It has an abacus serial adding algorithm and parallels adding algorithm, it’s essentially a calculation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Define single instruction multiple data

A

Represents multiple data items at once and operate on these arrays with single instructions such as ‘‘VECTOR-ADD’’

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What did complex CISC SIMD change compared to SIMD?

A

Added specialised word length vector registers in CPU which loads them up with data

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

How does SIMD operate inside the GPU?

A

DMA transfers to main RAM, smaller caches, small pipelines of two stages and thousands of ALUs, ultimately it’s optimised for data-parallel and high throughput computations

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

How does SIMD operate inside the CPU?

A

Optimised for low-latency, large caches which gives quick access to data, complex pipelines of total 30 stages but fewer ALUs

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What’s a work-group?

A

Group of Pes within a CU all executing the same instruction, single instruction

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What’s a work item?

A

Processing element executing it on one datum, multiple data

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What’s a kernel?

A

A program of SIMD instructions given to a work-group

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

How does the GPU implement itself?

A

Through OpenCL C, kernel code is compiled into GPU’s own assembly language and machine code

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What exactly is MIMD?

A

Related to vectors idea of architecture a.k.a. ‘‘Very Long Instructions Words’’ = VLIW

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Do all CPU’s have the same address space?

A

Yes

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What’s a cache coherency?

A

Some levels are shared, others not, tricky to get cache write right e.g. L1 cache needs to notify each other on changes and refreshes

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What does NUMA (non-uniform memory access) do?

A
  1. Single address space shared by processors
  2. But access time differ
  3. Used in supercomputers
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What’s a ‘blue gene’ supercomputer?

A

It’s based on custom chips, it’s task is to protein folding problems and it has a very hierarchical architecture

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What’s a Message Passing Interface?

A

A standard library for processes to send and receive discrete messages to one another

17
Q

How does MPI (message passing interface) pass like?

A

Through bus, mesh, shared memory communication

18
Q

What’s a “Beowolf” cluster architecture?

A

It’s a group of commodity PCs linked through regular network and MPI, one master node, other nodes slaves, run one program and master calls slaves to do parallel work

19
Q

What’s grid computing?

A
  1. Hundred or more maintained identical PCs
  2. No shared memory
  3. Ethernet connection
  4. Shared network discs
20
Q

What’s cloud computing?

A

Like grids but the much weaker organisation

21
Q

What’s map reducing?

A

Many cloud computing projects all evolved to use similar structures, splitting tasks into chunks, usually, each running the same code on different data

22
Q

What does map reducing work?

A

Map each chunk to worker computer, send back results, reduce these results to a single result

23
Q

What does ‘SpiNNaker’ do?

A

The task to simulate biological neurons at brain-like scales

24
Q

Does instruction-less parallelism need a CPU?

A

Nope, it doesn’t