sci.tamucc.edusci.tamucc.edu/~dkar/sum2017/2334/ch12.doc · web viewdevices that provide backup...

17
Chapter 12 – Memory Organization Section 12.1 – Memory Hierarchy The memory unit that communicates directly with the CPU is called the main memory where programs and data currently needed by the processor reside Devices that provide backup storage are called auxiliary memory, where all other information is stored and transferred to main memory when needed The goal of using memory hierarchy is to obtain the highest possible access speed while minimizing the total cost of the entire memory system Auxiliary vs. Main Memory Auxiliary Memory Main Memory Provides backup storage Holds programs and data currently needed Low-cost Smaller capacity Large capacity Direct access by CPU Holds data files, system programs, etc not currently needed Higher cost No direct access by CPU Faster access, but still not fast enough…

Upload: lykhuong

Post on 02-Apr-2018

218 views

Category:

Documents


1 download

TRANSCRIPT

Page 1: sci.tamucc.edusci.tamucc.edu/~dkar/sum2017/2334/ch12.doc · Web viewDevices that provide backup storage are called auxiliary memory, where all other information is stored and transferred

Chapter 12 – Memory Organization

Section 12.1 – Memory Hierarchy

The memory unit that communicates directly with the CPU is called the main memory where programs and data currently needed by the processor reside

Devices that provide backup storage are called auxiliary memory, where all other information is stored and transferred to main memory when needed

The goal of using memory hierarchy is to obtain the highest possible access speed while minimizing the total cost of the entire memory system

Auxiliary vs. Main MemoryAuxiliary Memory Main Memory

Provides backup storage Holds programs and data currently neededLow-cost Smaller capacityLarge capacity Direct access by CPUHolds data files, system programs, etc not currently needed

Higher cost

No direct access by CPU Faster access, but still not fast enough…

Cache Memory: Very-high speed memory Smaller capacity, higher cost Access time is close to processor logic clock cycle time Stores program segments currently being executed and data frequently accessed Increases performance of the computer

Page 2: sci.tamucc.edusci.tamucc.edu/~dkar/sum2017/2334/ch12.doc · Web viewDevices that provide backup storage are called auxiliary memory, where all other information is stored and transferred

Section 12.2 – Main Memory

RAM and ROM Chips:Typical RAM chip

Typical ROM chip

Page 3: sci.tamucc.edusci.tamucc.edu/~dkar/sum2017/2334/ch12.doc · Web viewDevices that provide backup storage are called auxiliary memory, where all other information is stored and transferred

Memory Address Map: The addressing of memory can be established by means of a table that specifies the

memory address assigned to each chip Example: 512 bytes RAM and 512 bytes ROM

Memory Connection to CPU: RAM and ROM chips are connected to a CPU through the data and address buses. The low-order lines in the address bus select the byte within the chips and other lines

in the address bus select a particular chip through its chip select inputs.

Page 4: sci.tamucc.edusci.tamucc.edu/~dkar/sum2017/2334/ch12.doc · Web viewDevices that provide backup storage are called auxiliary memory, where all other information is stored and transferred
Page 5: sci.tamucc.edusci.tamucc.edu/~dkar/sum2017/2334/ch12.doc · Web viewDevices that provide backup storage are called auxiliary memory, where all other information is stored and transferred

Section 12.4 – Associative Memory

The time required to find an item in memory can be reduced considerably if stored data can be identified for access by the content of the data itself rather than by an address

A memory unit accessed by content is called an associative memory or content addressable memory (CAM)

When a word is to be read from an associative memory, the content of the word, or part of the word, is specified.

It is more expensive than RAM as each cell must have storage capability as well as logic circuits for matching its content with an external argument

Compare each word in CAM in parallel with the content of A(Argument Register) If CAM Word[i] = A, M(i) = 1 Read sequentially accessing CAM for CAM Word(i) and set M(i) = 1 if there is a

match K(Key Register) provides a mask for choosing a particular field or key in the

argument register A (only those bits in the argument that have 1’s in their corresponding position of K are compared)

Example:

A 101 111100K 111 000000Word 1 100 111100 no matchWord 2 101 000001 match

Page 6: sci.tamucc.edusci.tamucc.edu/~dkar/sum2017/2334/ch12.doc · Web viewDevices that provide backup storage are called auxiliary memory, where all other information is stored and transferred

Section 12.5 – Cache Memory

The references to memory at any given interval of time tend to be confined within a few localized areas in memory. This phenomenon is known as locality of reference.

Temporal Locality: The information which will be used in near future is likely to be in use already( e.g. Reuse of information in loops)

Spatial Locality: If a word is accessed, adjacent(near) words are likely to be accessed soon (e.g. Related data items (arrays) are usually stored together; instructions are executed sequentially)

Cache: The property of locality of reference makes the cache memory systems work Cache is a fast small capacity memory that should hold those information which are

most likely to be accessed

All the memory accesses are directed first to cache If the word is in cache; access cache to provide it to the CPU If the word is not in cache; bring a block (or a line) including that word to replace a

block now in cache

Performance of Cache Memory System:

Hit Ratio - % of memory accesses satisfied by cache memory systemTe: Effective memory access time in cache memory systemTc: Cache access timeTm: Main memory access time

Te = Tc + (1 - h) TmExample: Tc = 0.4ns, Tm = 1.2ns, h = 0.85%Te = 0.4 + (1 - 0.85) × 1.2 = 0.58ns

The transformation of data from main memory to cache memory is referred to as a mapping process

1. Associative mapping2. Direct mapping3. Set-associative mapping

Page 7: sci.tamucc.edusci.tamucc.edu/~dkar/sum2017/2334/ch12.doc · Web viewDevices that provide backup storage are called auxiliary memory, where all other information is stored and transferred

Associative Mapping:

Any block location in cache can store any block in memory Most Flexible, fast, very expensive Mapping table is implemented in an associative memory Mapping table stores both address and the content of the memory word If the address is found, 12-bit data is read and sent to the CPU, if not main memory is

accessed for the word and the address-data pair is transferred to the associative cache memory

If the cache is full, an address-data pair is displaced to make room for the pair needed according to a replacement policy (e.g. round-robin, FIFO)

Direct Mapping:

Each memory block has only one place to load in cacheMapping table is made of RAM instead of CAMn-bit memory address consists of 2 parts; k bits of index field and n - k bits of tag fieldn-bit addresses are used to access main memory and k-bit index is used to access the cache

Page 8: sci.tamucc.edusci.tamucc.edu/~dkar/sum2017/2334/ch12.doc · Web viewDevices that provide backup storage are called auxiliary memory, where all other information is stored and transferred

When CPU generates a memory request, the index field is used for address to access the cache and the tag field is compared with the tag in the word read from the cache

If two tags matches, there is a hit and the desired data word is in cache If there is no match, there is a miss and the required word is read from main memory,

and stored in cache together with the new tag, replacing the old value The disadvantage is that the hit ratio drops if two or more words with same index but

different tags are accessed repeatedly (minimized by the fact that they are far apart- 512 locations)

Page 9: sci.tamucc.edusci.tamucc.edu/~dkar/sum2017/2334/ch12.doc · Web viewDevices that provide backup storage are called auxiliary memory, where all other information is stored and transferred

The index field is now divided into two parts; the block field and the word field. In a 512-word cache we have 64 blocks of 8 words each; the tag field stored within

the cache is common to all eight words of the same block Every time a miss occurs, an entire block of eight words is transferred from main

memory to cache memory; this takes extra time, but the hit ratio will most likely improve with a larger block size because of the sequential nature of programs

Set-Associative Mapping:

Each word of cache can store two or more words of memory under the same index address

The number of tag-data items in one word of cache is said to form a set The word length is 2(6 + 12) = 36 bits, the index address is 9 bits, and the cache can

accommodate 1024 words of main memory The index value of the address is used to access the cache, and the tag field of the

CPU address is compared with both tags; the comparison logic is done by an associative search of the tags in the set similar to an associative memory (thus the name set-associative)

Writing into Cache:

Write-Through method: Update main memory with every memory write operation, with cache memory being updated in parallel if it contains the word at the specified address

Memory is always updated -> Important when CPU and I/O device communicating through DMA are both executing

Slow, due to the memory access time Write-Back (Copy-Back) method: The cache location is updated during a write

operation, and the location is then marked by a flag so that later when the word is removed from cache it is copied into main memory

Page 10: sci.tamucc.edusci.tamucc.edu/~dkar/sum2017/2334/ch12.doc · Web viewDevices that provide backup storage are called auxiliary memory, where all other information is stored and transferred

Memory is not up-to-date, i.e., the same item in cache and memory may have different value

Section 12.6 – Virtual Memory

Give the programmer the illusion that the system has a very large memory (equal to the totality of auxiliary memory), even though the computer actually has a relatively small main memory.

Address Space (Logical) and Memory Space (Physical):

An address used by a programmer is called virtual address, and the set of such addresses the address space (N)

An address in main memory is called a location or physical address, and the set of such locations the memory space (M)

Portions of programs and data need not be in contiguous locations in memory since information is being moved in and out, and empty spaces may be available in scattered locations in memory

The address field of the instruction code has sufficient number of bits to specify all virtual addresses

A table is needed to map a virtual address of 20 bits to a physical address of 15 bits The mapping is a dynamic operation, which means that every address is translated

immediately as a word is referenced by CPU

Page 11: sci.tamucc.edusci.tamucc.edu/~dkar/sum2017/2334/ch12.doc · Web viewDevices that provide backup storage are called auxiliary memory, where all other information is stored and transferred

Address Mapping Using Pages:

Memory space is divided into groups of equal size called blocks and address space is divided into groups of equal size called pages

The mapping from address space to memory space is facilitated if each virtual address is considered to be represented by two numbers: a page number address and a line within a page.

In a computer with 2p words per page, p bits specify the line address and the remaining high-order bits specify the page number

In Fig. 12-18 a virtual address has 13 bits. Since each page consists of 1K = 1024 words, the high-order three bits will specify one of the eight pages and the low-order 10 bits give the line address within the page

The line address in address space and memory space is the same, and the only mapping required is from a page number to a block number

The address in the page table denotes the page number and the content of the word gives the block number where that page is stored in main memory

Page 12: sci.tamucc.edusci.tamucc.edu/~dkar/sum2017/2334/ch12.doc · Web viewDevices that provide backup storage are called auxiliary memory, where all other information is stored and transferred

In Fig. 12-19 the pages 1, 2, 5, and 6 are now available in main memory in blocks 3, 0, 1, and 2, respectively

The presence bit indicates whether the page has been transferred from auxiliary memory to main memory

The CPU references a word in memory with a virtual address of 13 bits If the presence bit in the word read from the page table is 0, it signifies that the

content of the word referenced by the virtual address does not reside in main memory, and a call to the operating system is generated to fetch the required page from auxiliary memory and place it into main memory

Associative Memory Page Table:

A RAM page table is inefficient with respect to storage utilization In Fig. 12-19 we observe that eight words of memory are needed, one for each page,

but only four are available as main memory can accommodate only four blocks For example if we have an address space of 1024K words and memory space of 32K

words, the number of pages is 1024 and the number of blocks is 32. Therefore, at any given time 32 locations may have a presence bit equal to 1, and at least 992 locations will be empty and not in use

Efficient way is to organize the page table with number of words equal to the number of blocks in main memory

Implemented by means of an associative memory with each word in memory containing a page number together with its corresponding block number

Page 13: sci.tamucc.edusci.tamucc.edu/~dkar/sum2017/2334/ch12.doc · Web viewDevices that provide backup storage are called auxiliary memory, where all other information is stored and transferred

Page Replacement:

Page fault is a condition that occurs when a program attempts to reference a page that is still in auxiliary memory

When a page fault occurs, the execution of the present program is suspended until the required page is brought into main memory

If the main memory is full, it would be necessary to remove a page from a memory block to make room for the new page according to a page replacement policy

The FIFO algorithm selects for replacement the page that has been in memory the longest time. Using a queue - every time a page is loaded, its identification number is inserted in the queue. Easy to implement as the identification number of the page to be removed can be accessed first

The OPT (Optimal Replacement) algorithm replaces that page which will not be used for the longest period of time. Lowest page fault rate of all the algorithms, but difficult to implement.

07

1

7

2 0 3 0 4 2 3 0 3 2 1 2 0 1 77 0 1

0 07

1

201

231

230

430

420

423

023

013

012

712

702

701

Page frames

Reference string

07

17

2 0 3 0 4 2 3 0 3 2 1 2 0 1 77 0 1

0 07

1

201

203

243

203

201

701

Page frames

Reference string

Page 14: sci.tamucc.edusci.tamucc.edu/~dkar/sum2017/2334/ch12.doc · Web viewDevices that provide backup storage are called auxiliary memory, where all other information is stored and transferred

The LRU algorithm replaces the page which has not been used for the longest period of time. It is implemented by associating a counter with every page that is in main memory. At fixed intervals of time, the counters associated with all pages presently in main memory are incremented by 1, and when a page is referenced, its associated counter is set to zero. The least recently used page is the page with the highest count.

07

17

2 0 3 0 4 2 3 0 3 2 1 2 0 1 77 0 1

0 07

1

201

203

403

402

432

032

132

102

107

Page frames

Reference string