Computer Organisation & Architecture
Complete GATE CS Notes — ISA, Pipelining, Cache, Memory Hierarchy, I/O & ALU
Last updated: April 2026 | GATE CS 2025 syllabus aligned
What You Will Learn in This Cluster
- How a CPU actually executes instructions — the fetch-decode-execute cycle and the role of each hardware unit
- Instruction Set Architecture (ISA): the contract between hardware and software, RISC vs CISC design philosophy
- Pipelining: how overlapping execution stages improves throughput, and how hazards disrupt the pipeline
- Cache memory: why it exists, how direct/set-associative/fully-associative mappings work, and how to calculate hit ratios
- Memory hierarchy: registers → cache → RAM → disk, and how each level balances speed, size, and cost
- I/O systems: polling, interrupts, DMA, and how the CPU interacts with peripherals without wasting cycles
- ALU and arithmetic: two’s complement, fast adder designs, Booth’s algorithm, and IEEE 754 floating-point
1. Why Computer Organisation Matters
Most students treat Computer Organisation (COA) as a subject to memorise before the exam — a bunch of diagrams and formulas about hardware that “software people” do not need. That is a costly mistake.
Every performance decision you make as a programmer — why a cache-friendly loop runs 10× faster, why a context switch is expensive, why branch prediction matters — has a direct explanation in COA. More practically, GATE CS asks 4–6 marks of COA every single year, and these questions are almost always numerical. If you understand the underlying model, the numbers follow naturally.
COA also connects every other CS subject. Compiler Design generates machine instructions described by the ISA. Operating Systems manage the memory hierarchy COA defines. Computer Networks use I/O mechanisms COA specifies. Study COA properly and you strengthen five subjects at once.
2. The Big Picture — How a Computer Works
Before diving into individual topics, it helps to see the full machine in one frame. A modern computer is built from four cooperating subsystems:
| Subsystem | What It Does | Key Components | Covered In |
|---|---|---|---|
| CPU | Fetches, decodes, and executes instructions | ALU, Control Unit, Registers, PC, IR | ISA, Pipelining, ALU |
| Memory | Stores instructions and data at multiple speed/size levels | Registers, Cache (L1/L2/L3), RAM, Virtual Memory | Cache Memory, Memory Hierarchy |
| I/O | Transfers data between CPU/memory and the outside world | Device controllers, DMA controller, Interrupt handler | I/O Systems |
| Interconnect | Carries data between subsystems | System bus (address, data, control), PCI, USB | I/O Systems, Memory Hierarchy |
The Classical Von Neumann Model
All mainstream computers follow the Von Neumann architecture: a single shared memory holds both program instructions and data, and the CPU reads from it sequentially through the fetch-decode-execute cycle.
1. Fetch: PC → MAR → Memory → MDR → IR (load instruction into Instruction Register)
2. Decode: Control Unit interprets the opcode in IR
3. Execute: ALU performs the operation; result stored in register or memory
4. PC ← PC + instruction_size (advance to next instruction)
The Von Neumann bottleneck — the bus between CPU and memory limits throughput — is precisely why cache memory and pipelining were invented. Understanding this bottleneck makes every subsequent COA topic feel logical rather than arbitrary.
Harvard Architecture vs Von Neumann
| Feature | Von Neumann | Harvard |
|---|---|---|
| Memory for instructions & data | Shared | Separate |
| Bus | Single shared bus | Separate instruction & data buses |
| Fetch + data access simultaneously? | No — bottleneck | Yes — parallel |
| Examples | Most general-purpose CPUs (x86, ARM) | Most DSPs, microcontrollers (PIC, AVR) |
| Complexity | Simpler | More complex |
3. All Topics in This Cluster
This cluster covers the complete GATE CS Computer Organisation & Architecture syllabus across seven focused pages. Each page contains worked GATE-level examples, formula boxes, and a quick-reference summary.
| # | Topic | What You Learn | GATE Importance |
|---|---|---|---|
| CO-1 | Instruction Set Architecture | RISC vs CISC, addressing modes, instruction formats, machine code | ★★★☆☆ |
| CO-2 | Pipelining | Pipeline stages, hazards (structural/data/control), speedup & efficiency | ★★★★★ |
| CO-3 | Cache Memory | Direct/set-associative/fully-associative mapping, replacement policies, AMAT | ★★★★★ |
| CO-4 | Memory Hierarchy | Memory types, speed-cost trade-off, virtual memory & TLB | ★★★★☆ |
| CO-5 | I/O Systems | Polling vs interrupts, DMA, I/O buses, device controllers | ★★★☆☆ |
| CO-6 | ALU & Arithmetic Operations | Two’s complement, fast adders, Booth’s algorithm, IEEE 754 floating-point | ★★★☆☆ |
| CO-7 | CO Formula Sheet | Every formula from this cluster on one page — for last-minute revision | ★★★★★ |
4. GATE CS Weightage & Previous Year Trends
COA has appeared in every GATE CS paper since 2010. The table below shows the distribution of marks across subtopics based on past papers.
| Subtopic | Typical GATE Marks | Question Types | Frequency |
|---|---|---|---|
| Pipelining | 2–3 marks | Speedup calculation, hazard identification, stall cycles | Every year |
| Cache Memory | 2–3 marks | Address breakdown, AMAT, miss penalty, replacement policy | Almost every year |
| Memory Organisation | 1–2 marks | Number of chips, address lines, memory map | Frequent |
| ISA / Addressing Modes | 1 mark | Conceptual MCQ, effective address calculation | Alternate years |
| I/O & DMA | 1 mark | DMA transfer time, interrupt handling | Occasional |
| ALU / Arithmetic | 1 mark | Booth’s algorithm steps, IEEE 754 representation | Occasional |
Master Pipelining + Cache first — they account for 60–70% of COA marks. Then cover Memory Organisation for 1–2 easy marks. ISA, I/O, and ALU are lower-yield but conceptually simple — do them last.
GATE 2024 COA Questions (Representative Examples)
Q1 (2 marks): A 5-stage pipeline has a clock cycle time of 10 ns. Due to a data hazard, 2 stall cycles are inserted. What is the effective CPI for a program with 1000 instructions where 20% have data hazards?
Base CPI = 1 (ideal pipeline)
Stall cycles per hazard instruction = 2
Stall fraction = 20% = 0.20
Effective CPI = 1 + (0.20 × 2) = 1.4
Total time = 1000 × 1.4 × 10 ns = 14,000 ns = 14 μs
Q2 (2 marks): A cache has 64 sets, 4-way set-associative, with block size 64 bytes. How many bits are used for the set index and block offset in a 32-bit physical address?
Block offset bits = log₂(64) = 6 bits
Set index bits = log₂(64 sets) = 6 bits
Tag bits = 32 − 6 − 6 = 20 bits
5. Recommended Study Order
COA topics build on each other. Following the wrong order leads to confusion — for example, trying to understand cache replacement policies without first understanding why cache exists leads to rote memorisation instead of real understanding.
| Order | Topic | Why This Comes First | Time to Master |
|---|---|---|---|
| 1st | ISA & Addressing Modes | Establishes vocabulary — what an instruction is, how operands are specified | 3–4 hours |
| 2nd | ALU & Arithmetic | Explains what the CPU actually does inside during execute stage | 4–5 hours |
| 3rd | Pipelining | Now you understand what is being pipelined — instructions executing in stages | 6–8 hours |
| 4th | Cache Memory | Motivation is clear — pipelining creates a fast CPU that needs fast memory | 6–8 hours |
| 5th | Memory Hierarchy | Cache sits within a broader memory hierarchy — now you can place it in context | 4–5 hours |
| 6th | I/O Systems | Final piece — how data gets in and out without stalling the CPU | 3–4 hours |
| 7th | Formula Sheet (revision) | Consolidate all formulas before the exam | 1–2 hours |
6. Essential Formulas at a Glance
These are the formulas that appear most frequently in GATE COA questions. Each one is derived in detail in the relevant topic page — do not memorise them blind.
Speedup = (n × k) / (k + n − 1)
where n = number of instructions, k = number of pipeline stages
Pipeline Efficiency:
Efficiency = Speedup / k = n / (k + n − 1)
Effective CPI with Stalls:
Effective CPI = Ideal CPI + Stall cycles per instruction
Average Memory Access Time (AMAT):
AMAT = Hit time + Miss rate × Miss penalty
Cache Size Calculation:
Number of blocks = Cache size / Block size
Number of sets = Number of blocks / Associativity
Address Breakdown (set-associative):
Offset bits = log₂(block size in bytes)
Index bits = log₂(number of sets)
Tag bits = Address bits − Index bits − Offset bits
Effective Address (Base + Offset):
EA = Base Register + Displacement
Two’s Complement Negation:
−X = Invert all bits + 1
7. Common Mistakes Students Make in COA
-
Confusing Architecture with Organisation
Students often use these interchangeably. Architecture = the programmer’s view (ISA, instruction formats). Organisation = how the hardware implements that view (control signals, bus widths, clock logic). GATE questions sometimes test this distinction directly. -
Forgetting the fill time in pipeline speedup
The formula for pipeline speedup is (n × k) / (k + n − 1), not n / k. The (k − 1) in the denominator accounts for the cycles spent filling the pipeline before it reaches steady state. On small-n problems, this makes a significant difference. -
Mixing up AMAT for read vs write operations
AMAT = Hit time + (Miss rate × Miss penalty) applies to reads. Writes depend on the write policy (write-through vs write-back) and whether a write buffer is present. Always check whether the question specifies a write-through or write-back cache. -
Miscounting address bits for cache mapping
A very common GATE mistake: students add index + offset + tag and get more or fewer bits than the physical address width. Always verify: Tag + Index + Offset = Total address bits. If it does not add up, re-examine the cache parameters. -
Treating DMA as “CPU-free” — it is not
DMA does transfer data without CPU involvement cycle-by-cycle, but the CPU still initiates the transfer, handles the completion interrupt, and shares the memory bus with the DMA controller (cycle stealing). Questions on DMA transfer time often catch students who assume zero CPU overhead.
8. Frequently Asked Questions
What is the difference between computer organisation and computer architecture?
Computer Architecture is the programmer-visible specification — instruction set, data types, addressing modes, and the register model. Computer Organisation is how the hardware actually implements that specification — control unit design, bus timing, ALU structure, and memory interface. Think of architecture as the API and organisation as the implementation behind it.
How many marks does Computer Organisation carry in GATE CS?
COA typically accounts for 4–6 marks in GATE CS. Pipelining numericals and cache calculations are the most reliable mark-earners, appearing in almost every paper. The remaining marks come from memory organisation, I/O, and ISA questions, which rotate in frequency year to year.
Which COA topic should I prioritise for GATE?
Pipelining and Cache Memory are the two highest-yield topics — together they usually account for 4–5 of the 4–6 available marks. Master those two first. Then cover Memory Hierarchy for 1–2 additional easy marks. Treat ISA, I/O, and ALU as secondary topics to fill in later.
Is Computer Organisation difficult for GATE CS preparation?
COA is one of the more calculation-heavy subjects in GATE CS, but the calculations are structured and formulaic once you understand the underlying model. Students who struggle with COA usually have a conceptual gap — they memorise formulas without understanding what cache sets, pipeline stages, or DMA controllers actually do. Fix the concept first, and the numericals become straightforward.