Common network architectures Flashcards
(10 cards)
Inception: State the multiplication cost for one convolution path and why bottlenecks help.
Mults = H·W·C·K²·F. 1×1 bottlenecks cut C, slashing FLOPs before larger kernels.
ResNet: Write the residual output equation and its purpose.
y = F(x) + x. Lets the block learn a residual; identity mapping is easy so very deep nets avoid degradation.
Precision formula and plain‑language meaning.
Precision = TP / (TP + FP): fraction of predicted boxes that are correct.
Recall formula and plain‑language meaning.
Recall = TP / (TP + FN): proportion of ground‑truth objects the detector finds.
Difference between precision and recall in one sentence.
Precision asks ‘How many predictions are right?’; recall asks ‘How many real objects did I catch?’
Why use 1×1 convs before 3×3/5×5 in Inception modules?
They reduce channel depth, lowering compute while keeping representational power.
How do residual skips help gradient flow?
They create a direct path so gradients bypass deep stacks, making optimisation easier for very deep networks.
Give the IoU formula and common TP threshold.
IoU = overlap area / union area; a detection is TP if IoU ≥ 0.5.
Define AP and mAP briefly.
AP = area under precision‑recall curve for one class; mAP = mean AP over all classes.
Conv output size formula (no padding) and example for 28×28, F=3, S=2.
H_out = ⌊(H_in−F)/S⌋+1 ⇒ ⌊(28−3)/2⌋+1 = 13.