L13 - Datacenter CPU Design Flashcards

1
Q

What are some difference between server CPUs and desktop CPUs?

A
  1. higher requests rates
  2. a lot of processing and moving of data
  3. HW is shared

extra: we need high utilisation in server CPUs

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Is optimising the individual server CPU design a good idea?

A

No. We want to optimise the whole performance and cost of the datacenter.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

List CPU design considerations.

A
  1. type of core: brawny vs wimpy
  2. number of cores
  3. number of sockets
  4. caches hierarchy and size
  5. isolation mechanisms (both for security and performance)
  6. integration with HW accelerators
  7. power management
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Why would one use low-end (wimpy) cores?

A

They consume less power. But their performance is worse.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Amdahl’s Law for Speedup

A

slide 15

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

When are brawny CPUs particularly of high interest?

A

Sequential parts of a program can only be made faster by using high-end CPUs.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What is the advantage of using more sockets?

A

more memory capacity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What is the advantage of using less sockets?

A
  • cores are more closely integrated
  • more uniform memory access
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Why don’t we run each application on dedicated CPU cores?

A

inference (in shared caches and memory)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What does Intel’s Cache Allocation Technology achieve?

A

partitions highly-associative last-level cache into subsets with smaller associativity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What would the advantages of allowing some overlaps in the cache?

A

Two different programs may utilise cache in different parts of their execution. By letting them overlap in cache, the overall throughput may be larger.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What is datacenter tax?

A

CPU cycles consumed by moving data

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What are front-end stalls?

A

overheads associated with fetching & decoding instructions

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What are back-end stalls?

A

overheads due to data cache hierarchy and lack of instruction-level parallelism

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Are workloads front-end or back-end bound?

Why?

A

front-end

reason: L1 cache is 100 times smaller than the typical instruction set of an application

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Resource requirements vary over time in a server. What is a possible way of dealing with this?

A

disaggregate resources