---
tags: computer-arch
---
# Quiz7 of Computer Architecture (2021 Fall)
:::info
:information_source: General Information
* You are allowed to read **[lecture materials](http://wiki.csie.ncku.edu.tw/arch/schedule)**.
* That is, an open book exam.
* We are using the honor system during this quiz, and would like you to accept the following:
1. You will not share the quiz with anyone.
2. You will not discuss the material on the quiz with anyone until after solutions are released.
* Each answer has `2` points.
* Of course, you must answer in valid numeric representation and/or English alphabet except your formal name.
* :timer_clock: 10:10 ~ 11:59 AM on Jan 11, 2022
* Fill ==[Google Form](https://docs.google.com/forms/d/e/1FAIpQLSdhyPmWEGOKZX7sveChol4kAZSFeK1yHaWV-yMam0jxRD0JTg/viewform)== to answer
:::
## Problem `A`
Assume that NCKU has been developing its own online grading system, "LearnSomthing", which it hopes will replace how people grade homework and exams both in NCKU and beyond.
"LearnSomthing" prides itself on parallelized grading by starting multiple threads running the `gradeExams` function described below in pseudocode.
```cpp
Shared Memory:
// exams is an array containing the exams
exams = [exam0, exam1, exam2, ... , exam99];
next_exam = 0;
gradeExams:
// Get ungraded exam
exam = exams[next_exam]
next_exam = next_exam + 1
// Grade exam
grade(exam)
goto gradeExams
```
1. Suppose two threads, `A` and `B`, are running the `gradeExams` code above without any synchronization. For each of the following failure scenarios, answer whether it is possible to happen or not: (`Yes` for possible; `No` for impossible)
* `A` and `B` start grading the same exam: __ A01 __
* `A` gets exam $k+1$ before `B` gets exam $k$: __ A02 __
* An exam is skipped and is never graded by either `A` or `B`: __ A03 __
> * A01 = ?, A02 = ?, A03 = ?
Despite LearnSomething's advanced parallel grading technology, it cannot interface with the gradebook module of this course. Instead, teaching assistant (TA) manually enters the grades into the gradebook as "LearnSomthing" completes its grading. When the `gradeExams` function finishes grading an exam, it emails a score report for the student to him.
He will read a score report email, enter the grade into the gradebook, and repeat, as represented by the `enterGrades` function below. He does not like having more than 10 unread score report emails at a time and also does not like checking her email when he does not have any new messages. Assume that there is more than 1 and fewer than 10 threads running the `gradeExams` function.
2. Add semaphores below to enforce these constraints:
* No two LearnSomething threads should ever grade the same exam
* Teaching assistant should never have more than 10 unread LearnScope score report emails
* `readScoreReportEmail()` should not be called until there is an unread score report email from "LearnSomething"
* After all 100 exams are claimed by "LearnSomething" threads, LearnSomething should stop attempting to grade exams
* As long as there are still unclaimed exams, avoid deadlock
* Use as few semaphores as possible, and do not add any additional precedence constraints
```cpp
Shared Memory:
// exams is an array containing the exams
exams = [exam0, exam1, exam2, ... , exam99];
next_exam = 0;
// Specify the semaphores and initial values here
semaphore remainingExams = A04;
semaphore nextExamLock = A05;
semaphore emailSpaces = A06;
semaphore unreadEmails = A07;
```
| `gradeExams:` | `enterGrades:` |
| -------- | -------- |
| __ A08 __ | __ A13 __ |
| __ A09 __ | `readScoreReportEmail()` |
| `// Get ungraded exam` | __ A14 __ |
| `exam = exams[next_exam]` | |
| | `enterGrade()` |
| `next_exam += 1` | |
| __ A10 __ | |
| | `goto enterGrades` |
| `// Grade exam` | |
| ` grade(exam)` | |
| __ A11 __ | |
| | |
| `emailScoreReport(exam)` | |
| __ A12 __ | |
| | |
| `goto gradeExams` | |
> * A04 = ?, A05 = ?, A06 = ?, A07 = ?
> * A08 = ?, A09 = ?, A10 = ? A11 = ? A12 = ? A13 = ? A14 = ?
---
## Problem `B`
The following semaphore-synchronized processes are run on a shared processor. They can coordinate their execution via shared semaphores, using standard signal and wait calls. Assume that execution may switch between any of the three processes at any point in time.
| Process~1~ | Process~2~ | Process~3~ |
| -------- | -------- | -------- |
| `Loop1:` | `Loop2:` | `Loop3:` |
| ` Print("I")` | ` Print("P")` | ` Print("L")` |
| ` Print("E")` | | ` Print("N")` |
| ` goto Loop1` | `goto Loop2` | ` goto Loop3` |
1. Assuming that no semaphores are being used, for each of the following sequences of characters, specify whether or not this system could produce that output. Answer in "Yes" or "No."
* "PIEPIE" : __ B01 __
* "LIPNIP" : __ B02 __
> * B01 = ?, B02 = ?
2. In the code listing below, add calls to signal and wait to ensure the three processes print "PIPELINE" exactly once and then stop printing. You should specify a semaphore in each call to signal or wait. You must use 4 semaphores needed.
```cpp
s1 = B03
s2 = B04
s3 = B05
p2p3stop = B06
```
| Process~1~ | Process~2~ | Process~3~ |
| -------- | -------- | -------- |
| `Loop1:` | `Loop2:` | `Loop3:` |
| __ B07 __ | __ B08 __ | __ B09 __ |
| ` Print("I")` | | ` Print("L")` |
| __ B10 __ | | __ B11 __ |
| __ B12 __ | ` Print("P")` | |
| | __ B13 __ | __ B14 __ |
| ` Print("E")` | __ B15 __ | ` Print("N")` |
| __ B16 __ | | __ B17 __ |
| | | __ B18 __ |
| ` goto Loop1` | `goto Loop2` | ` goto Loop3` |
> * B03 = ?, B04 = ?, B05 = ?, B06 = ?
> * B07 = ?, B08 = ?, B09 = ?, B10 = ?, B11 = ?, B12 = ?
> * B13 = ?, B14 = ?, B15 = ?, B16 = ?, B17 = ?, B18 = ?
---
## Problem `C`
Consider the execution of the following C code, which sums across the columns of a 2D array stored in [row-major order](https://en.wikipedia.org/wiki/Row-_and_column-major_order). Row-major order means that elements in the same row of the array are adjacent in memory, i.e., `A[i][j]` is next to `A[i][j+1]`. The array elements are 32-bit integers.
```cpp
int A[N][M]; // N = 32, M = 256
int sum = 0
for (int j = 0; j < M; j++) {
for (int i = 0; i < N; i++) {
sum += A[i][j];
}
}
```
1. Consider an 8-way set-associative cache with 16-byte cache lines and LRU replacement. What is the minimum number of sets that this cache needs such that this code will only produce compulsory misses? __ C01 __
> * C01 = ?
2. Suppose the cache is [virtually indexed and physically tagged](https://en.wikipedia.org/wiki/CPU_cache#Address_translation). Does the number of sets you derived in previous question introduce a virtual aliasing issue assuming a 4 KiB page size? __ C02 __ (Answer in Yes/No with brief explanation.)
> * C02 = ?
3. We would like to reduce the frequency of compulsory misses in this code by adding [software prefetching](https://en.wikipedia.org/wiki/Cache_prefetching) instructions. We measure the L1 miss penalty to be 40 cycles. When the prefetch instruction is replaced with a NOP, the first 32 iterations of the inner loop each take 50 cycles to complete. What should the `OFFSET` of the prefetch instruction be to maximize timeliness? __ C03 __
```cpp
int A[N][M]; // N = 32, M = 256
int sum = 0;
for (int j = 0; j < M; j++) {
for (int i = 0; i < N; i++) {
prefetch(&A[i][j] + OFFSET); // prefetches from (A + M * i + j + OFFSET)
sum += A[i][j];
}
}
```
> * C03 = ?
---
## Problem `D`
Consider the effectiveness evaluation of multithreading using a simple database benchmark. The benchmark searches for an entry in a linked list built from the following structure, which contains a key, a pointer to the next node in the linked list, and a pointer to the data entry.
```cpp
struct node {
int key;
struct node *next;
struct data *ptr;
}
```
The following RISC-V code shows the core of the benchmark, which traverses the linked list and finds an entry with a particular key.
```cpp
loop: LW x3, 0(x1) # load a key
LW x4, 4(x1) # load the next pointer
SEQ x3, x3, x2 # set x3 if x3 == x2
BNEZ x3, end # found the entry
ADD x1, x0, x4
BNEZ x4, loop # check the next node
end:
```
We run this benchmark on a single-issue in-order processor. The processor can fetch and issue (dispatch) one instruction per cycle. If an instruction cannot be issued due to a data dependency, the processor stalls. Integer instructions take one cycle to execute and the result can be used in the next cycle. For example, if `SEQ` is executed in cycle 1, `BNEZ` can be executed in cycle 2. We also assume that the processor has a perfect branch predictor with no penalty for both taken
and not-taken branches.
1. Assume that our system does not have a cache. Each memory operation directly accesses main memory and takes 100 CPU cycles. The load/store unit is fully pipelined, and non-blocking. After the processor issues a memory operation, it can continue executing instructions until it reaches an instruction that is dependent on an outstanding memory operation. How many cycles does it take to execute one iteration of the loop in steady state? __ D01 __ cycles
> * D01 = ?
2. Now we add zero-overhead multithreading to our pipeline. A processor executes multiple threads, each of which performs an independent search. Hardware mechanisms schedule a thread to execute each cycle. In our first implementation, the processor switches to a different thread every cycle using fixed round robin scheduling. Each of the N threads executes one instruction every N cycles. What is the minimum number of threads that we need to fully utilize the processor, i.e., execute one instruction per cycle? __ D02 __ thread(s)
> * D02 = ?
3. Then, we change the processor to only switch to a different thread when an instruction cannot execute due to data dependency. What is the minimum number of threads to fully utilize the processor now? Note that the processor issues instructions in-order in each thread. __ D03 __ thread(s)
> * D03 = ?
---
## Problem `E`
Assume that we are going to implement SIMD instructions in our pipelined RISC-V datapath. In order to do so, we will take 2 steps:
1. Combine the `IF` and `ID` stages into an `IFD` stage
2. Create 2 paths for the datapath to take after IF: one path for normal RISC-V instructions and one for SIMD RISC-V containing specialized hardware.
The only problem is that the SIMD `EX` stage takes 3 cycles to complete instead of 1, and no other SIMD instruction is allowed to enter the SIMD `EX` stage while another SIMD instruction is there.
![](https://hackmd.io/_uploads/ryGXBzc2Y.png)
1. This "delay" from the SIMD `EX` stage necessitates the use of stalls to ensure proper functionality. Which of the following implementations correctly generates the stall signal?
![](https://hackmd.io/_uploads/HJKKrM52Y.png)
* You may ignore any kinds of stalls caused by hazards; we are only concerned with this special case in our new pipeline. However, we still want to maintain a good instruction throughput. To do this, we should allow normal instructions to continue through the CPU, as they are not blocked from doing so by the SIMD path.
* The signal `SIMD_Inst` is an indicator that the current instruction fetched from `IFD` is a SIMD instruction, while the signal `SIMD_count` refers to the number of the cycle the SIMD instruction is completing in the `EX` stage, i.e. when it is in the first cycle of the EX stage, `SIMD_count` = 1. If there is no instruction in the SIMD `EX` stage, this value is 0. The comparators are unsigned. Select all that apply. __ E01 __ (Answer in A, B, C, D, or None of the other options)
> * E01 = ?
2. Because we wish to actually stall and not flush, how should the PC and PC mux be updated to allow for this?
![](https://hackmd.io/_uploads/ryWKUGc2K.png)
* Assume stall is a signal that is 1 when we should stall, and therefore not fetch a new instruction, or 0 otherwise.
* Select all that apply. __ E02 __ (Answer in A, B, C, D, or None of the other options)
> * E02 = ?
3. How many stalls caused by the SIMD EX stage are needed for the following piece of code? __ E03 __
```cpp=
addi t0, x0, 1
simd_add x1, x2, x3
simd_sub x2, x1, x3
addi t1, t0, 2
simd_mul x2, x2, x2
sub t0, t0, t1
mul t0, t0, t0
simd_div x2, x2, x1
```
> * E03 = ?
---
## Problem `F`
Assume we have a single-level, 1 KiB direct-mapped L1 cache with 16-byte blocks. We have 4 GiB of memory. An
integer is 4 bytes. The array is block-aligned.
```cpp
#define LEN 2048
int ARRAY[LEN];
int main() {
for (int i = 0; i < LEN - 256; i+=256) {
ARRAY[i] = ARRAY[i] + ARRAY[i+1] + ARRAY[i+256];
ARRAY[i] += 10;
}
}
```
1. Calculate the number of tag, index, and offset bits in the L1 cache.
* T: __ F01 __
* I: __ F02 __
* O: __ F03 __
> * F01 = ?, F02 = ?, F03 = ?
2. What is the hit rate for the code above? Assume C processes expressions left-to-right. __ F04 __
> * F04 = ?
3. We decide to add an L2 cache to our system. We shrink our L1 cache, so it now takes 3 cycles to access L1. Since we have a larger L2 cache, it takes 50 cycles to access L2. The L1 cache has a hit rate of 25% while the L2 cache has a hit rate of 90%. It takes 500 cycles to access physical memory. What is the average memory access time (AMAT) in cycles? __ F05 __
> * F05 = ?
---
## Problem `G`
Assume that you have been tasked to assess the performance of two designs:
* Processor~1~
* Processor~2~
You have decided to benchmark both processors on the following assembly sequence.
```cpp
L1: slli x11, x10, 2
lw x12, 0x80(x11)
add x13, x13, x12
addi x10, x10, 1
blt x10, x14, L1
sub x14, x14, x15
xori x15, x14, 0x1
```
Processor~1~ is a 4-stage pipelined processor with full bypassing. It attempts to improve performance by decreasing the number of cycles it takes to execute each instruction. The `EXE` and `MEM` stages are combined. Load requests are sent in the `EXE` stage and received in the `WB` stage one cycle after. The `IF` stage speculatively fetches the instruction at PC+4 and annuls incorrectly fetched instructions after a branch misprediction. Branches are resolved in the `EXE` stage.
Processor~2~ is a 6-stage pipelined processor with full bypassing. It attempts to improve performance by raising the clock speed. The `EXE` stage is split into two stages to break up the critical path through the ALU. The `IF` stage always fetches the instruction at PC+4 and annuls incorrectly fetched instructions after a taken branch. ALU and Branch ALU results are available only in EXE2.
1. How many cycles does it take the Procesor~1~ to execute one iteration of the loop? __ G01 __ cycle(s)
> * G01 = ?
2. How many cycles does it take the Processor~2~ to execute one iteration of the loop? __ G02 __ cycles(s)
> * G02 = ?
3. What is the clock period for Processor~1~? __ G03 __ ns per loop iteration.
> * G03 = ?
---
# Problem `H`
Consider a computer which has 2 processors, each with their own cache. Both have the same design: A 128 B cache size, 2-way set associative, 4 ints per block, write-back, and write-allocate with LRU replacement. Each cache takes in 20-bit addresses. Assume that ints are 4 bytes, and we are using the [MOESI cache-coherence protocol](https://en.wikipedia.org/wiki/MOESI_protocol).
1. Are the 20-bit addresses Virtual Addresses? __ H01 __ (answer in Yes or No)
> * H01 = ?
2. How many Offset bits? __ H02 __
> * H02 = ?
3. How many Index bits? __ H03 __
> * H03 = ?
4. How many Tag bits? __ H04 __
> * H04 = ?
We decide to parallelize a for loop across these 2 processors, but we have each thread do a strided memory access, where processor~0~ handles even indices, while processor~1~ handles odd indices. However, the memory accesses are perfectly interleaved, i.e. the order of array accesses are still `A[0]`, `A[1]`, `A[2]`, `A[3]` ...
```cpp
# define ARR_LEN 32
// A is located at address 0xA0000
int A[ARR_LEN];
// Processor 0’s loop
for (int i = 0; i < ARR_LEN; i += 2) {
A[i] += i
}
// Processor 1’s loop
for (int j = 1; j < ARR_LEN; j += 2) {
A[j] += j
}
```
5. What is the overall hit rate? __ H05 __
> * H05 = ?
6. What fraction of misses are coherency misses? Leave your answer as a fully simplified fraction. __ H06 __
> * H06 = ?
7. In total, how many times did we need to go to main memory to write-back? __ H07 __
> * H07 = ?
8. We want to avoid all the coherency misses, so we look to see if we can rewrite our code to optimize for cache performance. Which of the following methods will lead to a higher hit rate than that from the interleaved accesses? __ H08 __ (select all that apply)
a. Letting processor 0 start and finish, then processor 1 starts and finishes
b. Letting processor 1 start and finish, then processor 0 starts and finishes
c. None of the other options
> * H08 = ?
---
# Problem `I`
Assume that your processor is designed as the traditional five-stage pipeline. You find that your unit tests with single commands work for all instructions, and write some test patterns with multiple instructions. After running the test suite, the following cases fail. You should assume registers are initialized to 0, the error condition is calculated in the fetch stage, and no forwarding is currently implemented.
- [ ] Case 1: Assume the address of an array with all different values is stored in `s0`.
```cpp
addi t0, x0, 1
slli t1, t0, 2
add t1, s0, t1
lw t2, 0(t1)
```
Each time you run this test, there is the same incorrect output for `t2`. All the commands work individually on the single-stage pipeline.
1. What caused the failure? __ I01 __ (select ONE)
a. Control Hazard
b. Structural Hazard
c. Data Hazard
d. None of the above
> * I01 = ?
2. How could you fix it? __ I02 __ (select all that apply)
a. Insert a nop 3 times if you detect this specific error condition
b. Forward execute to write back if you detect this specific error condition
c. Forward execute to memory if you detect this specific error condition
d. Forward execute to execute if you detect this specific error condition
e. Flush the pipeline if you detect this specific error condition
> * I02 = ?
- [ ] Case 2: After fixing that hazard, the following case fails:
```cpp
addi s0, x0, 4
slli t1, s0, 2
bge s0, x0, greater
xori t1, t1, -1
addi t1, t1, 1
greater:
mul t0, t1, s0
```
When this test case is run, `t0` contains `0xFFFFFFC0`, which is not what it should have been.
3. What caused the failure? __ I03 __ (select ONE)
a. Control Hazard
b. Structural Hazard
c. Data Hazard
d. None of the above
> * I03 = ?
4. How could you fix it? __ I04 __ (select all that apply)
a. Insert a nop 3 times if you detect this specific error condition
b. Forward execute to write back if you detect this specific error condition
c. Forward execute to memory if you detect this specific error condition
d. Forward execute to execute if you detect this specific error condition
e. Flush the pipeline if you detect this specific error condition
> * I04 = ?
---
## Problem `J`
Which of the following were discussed in Prof. David Patterson's talk "[A New Golden Age for Computer Architecture History, Challenges, and Opportunities](https://youtu.be/uyc_pDBJotI)"? __ J01 __(select all that apply)
1. The power demands of machine learning are growing 10x per year; Moore’s Law is only 10x in 5 years.
2. The Tensor Processing Unit has a similar performance per watt to a CPU on production applications.
3. The marketplace has decided: Open ISAs (e.g., RISC-V) are better than proprietary ones (e.g., ARM).
4. Domain Specific Architectures achieve incredible performance but just for one application, like ASICs.
> * J01 = ?
---