Computer Organisation & Architecture — Complete GATE CS Hub

Computer Organisation & Architecture

Complete GATE CS Notes — ISA, Pipelining, Cache, Memory Hierarchy, I/O & ALU

Last updated: April 2026  |  GATE CS 2025 syllabus aligned

What You Will Learn in This Cluster

  • How a CPU actually executes instructions — the fetch-decode-execute cycle and the role of each hardware unit
  • Instruction Set Architecture (ISA): the contract between hardware and software, RISC vs CISC design philosophy
  • Pipelining: how overlapping execution stages improves throughput, and how hazards disrupt the pipeline
  • Cache memory: why it exists, how direct/set-associative/fully-associative mappings work, and how to calculate hit ratios
  • Memory hierarchy: registers → cache → RAM → disk, and how each level balances speed, size, and cost
  • I/O systems: polling, interrupts, DMA, and how the CPU interacts with peripherals without wasting cycles
  • ALU and arithmetic: two’s complement, fast adder designs, Booth’s algorithm, and IEEE 754 floating-point

1. Why Computer Organisation Matters

Most students treat Computer Organisation (COA) as a subject to memorise before the exam — a bunch of diagrams and formulas about hardware that “software people” do not need. That is a costly mistake.

Every performance decision you make as a programmer — why a cache-friendly loop runs 10× faster, why a context switch is expensive, why branch prediction matters — has a direct explanation in COA. More practically, GATE CS asks 4–6 marks of COA every single year, and these questions are almost always numerical. If you understand the underlying model, the numbers follow naturally.

COA also connects every other CS subject. Compiler Design generates machine instructions described by the ISA. Operating Systems manage the memory hierarchy COA defines. Computer Networks use I/O mechanisms COA specifies. Study COA properly and you strengthen five subjects at once.

2. The Big Picture — How a Computer Works

Before diving into individual topics, it helps to see the full machine in one frame. A modern computer is built from four cooperating subsystems:

SubsystemWhat It DoesKey ComponentsCovered In
CPUFetches, decodes, and executes instructionsALU, Control Unit, Registers, PC, IRISA, Pipelining, ALU
MemoryStores instructions and data at multiple speed/size levelsRegisters, Cache (L1/L2/L3), RAM, Virtual MemoryCache Memory, Memory Hierarchy
I/OTransfers data between CPU/memory and the outside worldDevice controllers, DMA controller, Interrupt handlerI/O Systems
InterconnectCarries data between subsystemsSystem bus (address, data, control), PCI, USBI/O Systems, Memory Hierarchy

The Classical Von Neumann Model

All mainstream computers follow the Von Neumann architecture: a single shared memory holds both program instructions and data, and the CPU reads from it sequentially through the fetch-decode-execute cycle.

Fetch-Decode-Execute Cycle:
1. Fetch: PC → MAR → Memory → MDR → IR (load instruction into Instruction Register)
2. Decode: Control Unit interprets the opcode in IR
3. Execute: ALU performs the operation; result stored in register or memory
4. PC ← PC + instruction_size (advance to next instruction)

The Von Neumann bottleneck — the bus between CPU and memory limits throughput — is precisely why cache memory and pipelining were invented. Understanding this bottleneck makes every subsequent COA topic feel logical rather than arbitrary.

Harvard Architecture vs Von Neumann

FeatureVon NeumannHarvard
Memory for instructions & dataSharedSeparate
BusSingle shared busSeparate instruction & data buses
Fetch + data access simultaneously?No — bottleneckYes — parallel
ExamplesMost general-purpose CPUs (x86, ARM)Most DSPs, microcontrollers (PIC, AVR)
ComplexitySimplerMore complex

3. All Topics in This Cluster

This cluster covers the complete GATE CS Computer Organisation & Architecture syllabus across seven focused pages. Each page contains worked GATE-level examples, formula boxes, and a quick-reference summary.

#TopicWhat You LearnGATE Importance
CO-1Instruction Set ArchitectureRISC vs CISC, addressing modes, instruction formats, machine code★★★☆☆
CO-2PipeliningPipeline stages, hazards (structural/data/control), speedup & efficiency★★★★★
CO-3Cache MemoryDirect/set-associative/fully-associative mapping, replacement policies, AMAT★★★★★
CO-4Memory HierarchyMemory types, speed-cost trade-off, virtual memory & TLB★★★★☆
CO-5I/O SystemsPolling vs interrupts, DMA, I/O buses, device controllers★★★☆☆
CO-6ALU & Arithmetic OperationsTwo’s complement, fast adders, Booth’s algorithm, IEEE 754 floating-point★★★☆☆
CO-7CO Formula SheetEvery formula from this cluster on one page — for last-minute revision★★★★★

4. GATE CS Weightage & Previous Year Trends

COA has appeared in every GATE CS paper since 2010. The table below shows the distribution of marks across subtopics based on past papers.

SubtopicTypical GATE MarksQuestion TypesFrequency
Pipelining2–3 marksSpeedup calculation, hazard identification, stall cyclesEvery year
Cache Memory2–3 marksAddress breakdown, AMAT, miss penalty, replacement policyAlmost every year
Memory Organisation1–2 marksNumber of chips, address lines, memory mapFrequent
ISA / Addressing Modes1 markConceptual MCQ, effective address calculationAlternate years
I/O & DMA1 markDMA transfer time, interrupt handlingOccasional
ALU / Arithmetic1 markBooth’s algorithm steps, IEEE 754 representationOccasional
GATE Strategy for COA:
Master Pipelining + Cache first — they account for 60–70% of COA marks. Then cover Memory Organisation for 1–2 easy marks. ISA, I/O, and ALU are lower-yield but conceptually simple — do them last.

GATE 2024 COA Questions (Representative Examples)

Q1 (2 marks): A 5-stage pipeline has a clock cycle time of 10 ns. Due to a data hazard, 2 stall cycles are inserted. What is the effective CPI for a program with 1000 instructions where 20% have data hazards?

Solution:
Base CPI = 1 (ideal pipeline)
Stall cycles per hazard instruction = 2
Stall fraction = 20% = 0.20
Effective CPI = 1 + (0.20 × 2) = 1.4
Total time = 1000 × 1.4 × 10 ns = 14,000 ns = 14 μs

Q2 (2 marks): A cache has 64 sets, 4-way set-associative, with block size 64 bytes. How many bits are used for the set index and block offset in a 32-bit physical address?

Solution:
Block offset bits = log₂(64) = 6 bits
Set index bits = log₂(64 sets) = 6 bits
Tag bits = 32 − 6 − 6 = 20 bits

5. Recommended Study Order

COA topics build on each other. Following the wrong order leads to confusion — for example, trying to understand cache replacement policies without first understanding why cache exists leads to rote memorisation instead of real understanding.

OrderTopicWhy This Comes FirstTime to Master
1stISA & Addressing ModesEstablishes vocabulary — what an instruction is, how operands are specified3–4 hours
2ndALU & ArithmeticExplains what the CPU actually does inside during execute stage4–5 hours
3rdPipeliningNow you understand what is being pipelined — instructions executing in stages6–8 hours
4thCache MemoryMotivation is clear — pipelining creates a fast CPU that needs fast memory6–8 hours
5thMemory HierarchyCache sits within a broader memory hierarchy — now you can place it in context4–5 hours
6thI/O SystemsFinal piece — how data gets in and out without stalling the CPU3–4 hours
7thFormula Sheet (revision)Consolidate all formulas before the exam1–2 hours

6. Essential Formulas at a Glance

These are the formulas that appear most frequently in GATE COA questions. Each one is derived in detail in the relevant topic page — do not memorise them blind.

Pipeline Speedup:
Speedup = (n × k) / (k + n − 1)
where n = number of instructions, k = number of pipeline stages

Pipeline Efficiency:
Efficiency = Speedup / k = n / (k + n − 1)

Effective CPI with Stalls:
Effective CPI = Ideal CPI + Stall cycles per instruction

Average Memory Access Time (AMAT):
AMAT = Hit time + Miss rate × Miss penalty

Cache Size Calculation:
Number of blocks = Cache size / Block size
Number of sets = Number of blocks / Associativity

Address Breakdown (set-associative):
Offset bits = log₂(block size in bytes)
Index bits = log₂(number of sets)
Tag bits = Address bits − Index bits − Offset bits

Effective Address (Base + Offset):
EA = Base Register + Displacement

Two’s Complement Negation:
−X = Invert all bits + 1

7. Common Mistakes Students Make in COA

  1. Confusing Architecture with Organisation
    Students often use these interchangeably. Architecture = the programmer’s view (ISA, instruction formats). Organisation = how the hardware implements that view (control signals, bus widths, clock logic). GATE questions sometimes test this distinction directly.
  2. Forgetting the fill time in pipeline speedup
    The formula for pipeline speedup is (n × k) / (k + n − 1), not n / k. The (k − 1) in the denominator accounts for the cycles spent filling the pipeline before it reaches steady state. On small-n problems, this makes a significant difference.
  3. Mixing up AMAT for read vs write operations
    AMAT = Hit time + (Miss rate × Miss penalty) applies to reads. Writes depend on the write policy (write-through vs write-back) and whether a write buffer is present. Always check whether the question specifies a write-through or write-back cache.
  4. Miscounting address bits for cache mapping
    A very common GATE mistake: students add index + offset + tag and get more or fewer bits than the physical address width. Always verify: Tag + Index + Offset = Total address bits. If it does not add up, re-examine the cache parameters.
  5. Treating DMA as “CPU-free” — it is not
    DMA does transfer data without CPU involvement cycle-by-cycle, but the CPU still initiates the transfer, handles the completion interrupt, and shares the memory bus with the DMA controller (cycle stealing). Questions on DMA transfer time often catch students who assume zero CPU overhead.

8. Frequently Asked Questions

What is the difference between computer organisation and computer architecture?

Computer Architecture is the programmer-visible specification — instruction set, data types, addressing modes, and the register model. Computer Organisation is how the hardware actually implements that specification — control unit design, bus timing, ALU structure, and memory interface. Think of architecture as the API and organisation as the implementation behind it.

How many marks does Computer Organisation carry in GATE CS?

COA typically accounts for 4–6 marks in GATE CS. Pipelining numericals and cache calculations are the most reliable mark-earners, appearing in almost every paper. The remaining marks come from memory organisation, I/O, and ISA questions, which rotate in frequency year to year.

Which COA topic should I prioritise for GATE?

Pipelining and Cache Memory are the two highest-yield topics — together they usually account for 4–5 of the 4–6 available marks. Master those two first. Then cover Memory Hierarchy for 1–2 additional easy marks. Treat ISA, I/O, and ALU as secondary topics to fill in later.

Is Computer Organisation difficult for GATE CS preparation?

COA is one of the more calculation-heavy subjects in GATE CS, but the calculations are structured and formulaic once you understand the underlying model. Students who struggle with COA usually have a conceptual gap — they memorise formulas without understanding what cache sets, pipeline stages, or DMA controllers actually do. Fix the concept first, and the numericals become straightforward.