Computer Architecture-QBANK
Computer Architecture-QBANK
DEPARTMENT OF
ELECTRONICS AND COMMUNICATION
ENGINEERING
QUESTION BANK
VI SEMESTER
CS6303-COMPUTER ARCHITECTURE
Regulation – 2013
Prepared by
QUESTION BANK
2 Explain and compare the hardwired and microprogrammed (15) BTL 5 Evaluating
control unit designs in terms of their mechanism of generating
control signals with diagram.
3 Consider the following loop: BTL 5 Evaluating
Loop: lw r1,0(r1)
and r1,r1,r2
lw r1,0(r1)
lw r1,0(r1)
beq r1,r0,loop
Assume that perfect branch prediction is used (no stalls) that there
are no delay slots, and that the pipeline has full forwarding
support. Also assume that many iterations of this loop are
executed before the loop exits.
i).Assess a pipeline execution diagram for the third iteration of (8)
this loop.
ii).Show all instructions that are in the pipeline during these (7)
cycles (for all iterations).
4 Plan the pipelining in MIPS architecture and generate the (15) BTL 6 Creating
exceptions handled in MIPS.
UNIT IV-PARALLELISM
Instruction-level-parallelism – Parallel processing challenges – Flynn’s classification –
Hardware multithreading – Multicore processors
PART-A
Q.No Questions BT Competence
Level
1 Describe the main idea of ILP. BTL 2 Understanding
2 Illustrate how to organize a multiple issue processor? BTL 3 Applying
3 List the three important properties of vector instructions. BTL 1 Remembering
4 Analyze the main characteristics of SMT processor. BTL 4 Analyzing
5 Quote the importance of loop unrolling technique. BTL 1 Remembering
6 Define VLIW processor. BTL 1 Remembering
7 Express anti-dependence. How is it removed? BTL 2 Understanding
8 State the overall speedup if a webserver is to be enhanced with a BTL 1 Remembering
new CPU which is 10 times faster on computation than an old
CPU .The original CPU spent 40% of its time processing and
60% of its time waiting for I/O.
9 Differentiate between strong scaling and weak scaling. BTL 2 Understanding
10 Show the performance of cluster organization. BTL 3 Applying
11 Compare SMT and hardware multithreading. BTL 5 Evaluating
12 Identify the Flynn classification and give an example for each BTL 1 Remembering
class in Flynn’s classification.
13 Integrate the ideas of in-order execution and out-of-order execution. BTL 6 Creating
14 Discriminate UMA and NUMA. BTL 5 Evaluating
15 Describe fine grained multithreading. BTL 1 Remembering
16 Express the need for instruction level parallelism. BTL 2 Understanding
17 Formulate the various approaches to hardware multithreading. BTL 6 Creating
18 Categorize the various multithreading options. BTL 4 Analyzing
19 Compare fine grained multithreading and coarse grained BTL 4 Analyzing
multithreading.
20 Classify shared memory multiprocessor based on the memory BTL 3 Applying
access latency.
PART-B
1 i).Define parallelism and its types. (4) BTL 1 Remembering
ii).List the main characteristics and limitations of Instruction level (9)
parallelism.
2 i).Give the software and hardware techniques to achieve Instruction (4) BTL 2 Understanding
level parallelism.
ii).Summarize the facts or challenges faced by parallel processing in (9)
enhancing computer architecture.
3 Express in detail about hardware multithreading. (13) BTL 2 Understanding
4 Solve: suppose you want to achieve a speed up to 90 times faster (13) BTL 3 Applying
with 100 processors. What percentage of the original computation
can be sequential?
5 Describe data level parallelism in BTL 1 Remembering
i).SIMD. (6)
ii).MISD. (7)
6 i).Point out how will you classify shared memory multi-processor BTL 4 Analyzing
based on memory access latency. (7)
ii).Compare and contrast Fine grained, Coarse grained multithreading
and Simultaneous Multithreading. (8)
7 Evaluate the features of Multicore processors. (13) BTL 5 Evaluating
8 i).Classify the types of multithreading. (9) BTL 4 Analyzing
ii).Analyze the advantages in multithreading. (4)
9 Formulate the classes in Flynn’s Taxonomy of computer (13) BTL 6 Creating
Architecture classification.
10 Elaborate in detail about the following
i).SISD. (8) BTL 1 Remembering
ii).MIMD (5)
11 Explain simultaneous Multithreading with example. (13) BTL 4 Analyzing
12 Describe the four principle approaches to multithreading with (13) BTL 1 Remembering
necessary diagrams.
13 Illustrate the following in detail BTL 3 Applying
i).Data Dependence (5)
ii).Name Dependence (4)
iii).Control dependence (4)
14 Discuss the following in detail BTL 2 Understanding
i). Vector processor. (7)
ii).Superscalar processor. (6)
PART C
1 Explain how would this loop be scheduled on a static two issue (15)
pipeline for MIPS?
Loop: lw $t0,0($s1) #$t0=array element
Addu $t0,$t0,$s2 #add scalar in $s2
Sw $t0, 0($s1) # store result
BTL 6 Creating
Addi; %s1,$s1, -4#decrement pointer
Bne $s1,$zero,loop # branch $s1!=0
Decide and reorder the instruction to avoid as many pipeline stalls
as possible. Assume branches are predicted, so that control
hazards are handled by the hardware.
2 A pipelined processor uses delayed branch technique. (15)
Recommend any one of the following possibility for the design of
the processor. In the first possibility, the processor has a 4-satge BTL 5 Evaluating
pipeline and one delay slot. In the second possibility, it has a 6-
stage pipeline and two delay slots. Compare the performance of
these two alternatives, taking only the branch penalty into
account. Assume that 20% of the instructions are branch
instructions and that an optimizing compiler has an 80% success
rate in filling in the single delay slot. For the second alternative,
the compiler is able to fill the second slot 25% of the time.
3 Consider the following portions of two different programs BTL 6 Creating
running at the same time on four processors in a symmetric
multicore processor (SMP). Assume that before this code is run,
both x and y are 0?
Core 1: x=2;
Core 2: y=2;
Core 3: w= x + y +1;
Core 4: z= x + y;
i. What if all the possible resulting values of w,x,y,z ? For each (8)
possible outcomes, explain how we might arrive at those values.
ii. Develop the execution more deterministic so that only one set (7)
of values is possible?
4 Suppose we want to perform 2 sums: one is a sum of 10 scalar (15) BTL 6 Creating
variables and one is a matrix sum of a pair of two dimensional
arrays, with dimensions 10 by 10. For now let’s assume only the
matrix sum is parallelizable. What if the speed up do you get with
10 versus 40 processors and next calculate the speed ups
assuming the matrices grow to 20 by 20.
UNIT V-MEMORY AND I/O SYSTEMS
Memory hierarchy – Memory technologies – Cache basics – Measuring and improving cache performance –
Virtual memory, TLBs – Input/output system, programmed I/O, DMA and interrupts, I/O processors.
PART-A
Q.No Questions BT Competence
Level
1 Distinguish the types of locality of references. BTL 2 Understanding
2 Define the structure of memory hierarchy in a typical computer BTL 1 Remembering
system and draw its diagram.
3 Give the idea of address mapping and memory –mapped I/O. BTL 2 Understanding
4 Compare and contrast SRAM and DRAM. BTL 4 Analyzing
5 What is the need to implement memory as a hierarchy? BTL 1 Remembering
6 Describe Rotational Latency. BTL 1 Remembering
7 State is direct-mapped cache. BTL 1 Remembering
8 Evaluate the following instance wherein the cache size is 64 BTL 5 Evaluating
blocks and block size is 16 bytes. What block number does byte
address 1200 map?
9 Formulate, how many total bits are required for a direct-mapped BTL 6 Creating
cache with 16 KB of data and 4-word blocks, assuming a 32-bit
address?
10 Analyze the writing strategies in cache memory. BTL 4 Analyzing
11 Integrate the functional steps required in an instruction cache BTL 6 Creating
miss.
12 State hit rate and miss rate. BTL 1 Remembering
13 Summarize the various block placement schemes in cache BTL 2 Understanding
memory.
14 Identify the purpose of Dirty/Modified bit in Cache memory. BTL 1 Remembering
15 Point out the use of DMA controller? BTL 4 Analyzing
16 Show the role of TLB in virtual memory. BTL 3 Applying
17 Illustrate the advantages of virtual memory. BTL 3 Applying
18 Assess the relationship between physical address and logical BTL 5 Evaluating
address.
19 Differentiate Programmed I/O and Interrupt I/O. BTL 2 Understanding
20 Demonstrate the sequence of events involved in handling an BTL 3 Applying
interrupt request from a single device.
PART-B
1 i).List the various memory technologies and examine its (8) BTL 1 Remembering
relevance in architecture design.
ii). Identify the characteristics of memory system. (5)
2 Elaborate in detail the memory hierarchy with neat diagram. (13) BTL 1 Remembering
3 i).Give the advantages of cache. (4) BTL 2 Understanding
ii).Identify the basic operations of cache in detail with diagram. (9)
4 Express the need for cache memory. Explain the following three BTL 2 Understanding
mapping methods with examples.
i). Direct. (4)
ii).Associative. (4)
iii).Set associative. (5)
5 i). Analyze the given problem: (8)
A byte addressable computer has a small data cache capable of BTL 4 Analyzing
holding eight 32-bit words. Each cache block contains 132-bit
word. When a given program is executed, the processor reads
data from the following sequence of hex addresses – 200, 204,
208, 20C, 2F4, 2F0, 200,204,218, 21C, 24C, 2F4. The pattern is
repeated four times. Assuming that the cache is initially empty,
show the contents of the cache at the end of each pass, and
compute the hit rate for a direct mapped cache.
ii).What are the methods used to measure and improve the (5)
performance of the cache?
6 i).Define virtual memory and its importance. (5) BTL 1 Remembering
ii).Examine TLB with necessary diagram .What is its use? (8)
7 i).Demonstrate the DMA controller. Discuss how it improves the (4) BTL 3 Applying
overall performance of the system.
ii).Illustrate how DMA controller is used for direct data transfer (9)
between memory and peripherals?
8 i).Evaluate the advantages of interrupts. (5) BTL 5 Evaluating
ii).Summarize the concept of interrupts and interrupts handling
with neat diagrams. (8)
9 Design standard input and output interfaces required to connect (13) BTL 6 Creating
the I/O device to the bus.
10 Classify the bus arbitration techniques of DMA in detail. (13) BTL 4 Analyzing
11 Point out the following in detail BTL 4 Analyzing
i). Programmed I/O associated with computers. (7)
ii).Write the sequence of operations carried out by a processor when(6)
interrupted by peripheral device connected to it.
12 Describe in detail about the methods used to reduce cache misses. (13) BTL 1 Remembering
13 Discuss virtual memory address translation in detail with (13) BTL 2 Understanding
necessary diagram.
14 Calculate the performance the processor : (13) BTL 3 Applying
Assume the miss rate of an instruction cache is 2% and the miss
rate of the data cache is 4%. If a processor has a CPI of 2 without
any memory stalls and the miss penalty is 100 cycles for all
misses, estimate how much faster a processor would run with a
perfect cache that never missed. Assume the frequency of all
loads and stores is 36%.
PART C
1 Mean Time Between Failures (MTBF), Mean Time To
Replacement (MTTR) and Mean Time To Failure (MTTF) are
useful metrics for evaluating the reliability and availability of a
storage resource. Explore these concepts by answering the
questions about devices with the following metrics:
BTL 6 Creating
MTTF : 3 years MTTR: 1 day
i). Develop and calculate the MTBF for each of the devices. (3)
ii).Develop and calculate the availability for each of the devices. (4)
iii).What if happens to availability as the MTTR approaches 0? (4)
iv).What if happens to availability as the MTTR gets very high? (4)
2 Design and explain parallel priority interrupt hardware for a (15)
BTL 6 Evaluating
system with eight interrupt sources.
3 For a direct mapped cache design with a 32 bit address, the BTL 5 Evaluating
following bits of the address are used to access the cache.
Tag : 31-10 Index: 9-5 Offset: 4-0
i). Judge what is the cache block size? (5)
ii).Decide how many entries does the cache have? (5)
iii).Assess what is the ratio between total bits required for such a (5)
cache implementation over the data storage bits?
4 Summarize by considering web application .Assuming both client (15) BTL 5 Evaluating
and servers are involved in the process of web browsing
application, where can caches be placed to speed up the process.
Design a memory hierarchy for the system. Show the typical size
and latency at various levels of the hierarchy. What is the
relationship between the cache size and its access latency? What
are the units of data transfers between hierarchies? What id the
relationship between data location, data size and transfer latency?