1 / 81

Chapter Seven Large and Fast: Exploiting Memory Hierarchy

Chapter Seven Large and Fast: Exploiting Memory Hierarchy. Outline. 7.1 Introduction 7.2 The Basic of Caches 7.3 Measuring and Improving Cache Performance 7.4 Virtual Memory 7.5 A Common Framework for Memory Hierarchies 7.6 Real Stuff: The Pentium P4 and the AMD Opteron Memory Hierarchies

silas
Télécharger la présentation

Chapter Seven Large and Fast: Exploiting Memory Hierarchy

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Chapter SevenLarge and Fast: Exploiting Memory Hierarchy

  2. Outline • 7.1 Introduction • 7.2 The Basic of Caches • 7.3 Measuring and Improving Cache Performance • 7.4 Virtual Memory • 7.5 A Common Framework for Memory Hierarchies • 7.6 Real Stuff: The Pentium P4 and the AMD Opteron Memory Hierarchies • 7.7 Fallacies and Pitfalls • 7.8 Concluding Remarks • 7.9 Historical Perspective and Further Reading

  3. 7.1 Introduction The principle of localitystates that programs access a relatively small portion of their address space at any instant of time, just as you accessed a very small portion of the library’s collection.

  4. Keywords • Temporal localityThe principle stating that if a data location is referenced then it will tend to be referenced again soon. • Spatial localityThe locality principle stating that if a data location is referenced, data locations with nearby addresses will tend to be referenced soon. • Memory hierarchyA structure that uses multiple levels of memories; as the distance from the CPU increases, the size of the memories and the access time both increase. • BlockThe minimum unit of information that can be either present or not present in the two-level hierarchy. • Hit rateThe fraction of memory accesses found in a cache. • Miss rateThe fraction of memory accesses not found in a level of the memory hierarchy.

  5. Keywords • Hit timeThe time required to access a level of the memory hierarchy, including the time needed to determine whether the access is a hit or a miss. • Miss penaltyThe time required to fetch a block into a level of the memory hierarchy from the lower level, including the time to access the block, transmit it from one level to the other, and insert it in the level that experienced the miss.

  6. Memories: Review • SRAM: • value is stored on a pair of inverting gates • very fast but takes up more space than DRAM (4 to 6 transistors) • DRAM: • value is stored as a charge on capacitor (must be refreshed) • very small but slower than SRAM (factor of 5 to 10)

  7. Figure 7.1 The basic structure of a memory hierarchy.

  8. Figure 7.2 Every pair of levels in the memory hierarchy can be thought of as having an upper and lower level.

  9. Exploiting Memory Hierarchy • Users want large and fast memories! SRAM access times are .5 – 5ns at cost of $4000 to $10,000 per GB.DRAM access times are 50-70ns at cost of $100 to $200 per GB.Disk access times are 5 to 20 million ns at cost of $.50 to $2 per GB. • Try and give it to them anyway • build a memory hierarchy 2004

  10. Locality • A principle that makes having a memory hierarchy a good idea • If an item is referenced,temporal locality: it will tend to be referenced again soon spatial locality: nearby items will tend to be referenced soon. Why does code have locality? • Our initial focus: two levels (upper, lower) • block: minimum unit of data • hit: data requested is in the upper level • miss: data requested is not in the upper level

  11. 7.2 The Basics of Caches

  12. Keywords • Direct-mapped cacheA cache structure in which each memory location is mapped to exactly one location in the cache. • TagA field in a table used for a memory hierarchy that contains the address information required to identify whether the associated block in the hierarchy corresponds to a requested word. • Valid bitA field in the tables of a memory hierarchy that indicates that the associated block in the hierarchy contains valid data. • Cache missA request for data from the cache that cannot be filled because the data is not present in the cache. • Write-throughA scheme in which writes always update both the cache and the memory, ensuring that data is always consistent between the two. • Write bufferA queue that holds data while the data are waiting to be written to memory.

  13. Keywords • Write-back A scheme that handles writes by updating values only to the block in the cache, then writing the modified block to the lower level of the hierarchy when the block is replaced. • Split cacheA scheme in which a level of the memory hierarchy is composed of two independent caches that operate in parallel with each other with one handling instructions and one handling data.

  14. Cache • Two issues: • How do we know if a data item is in the cache? • If it is, how do we find it? • Our first example: • block size is one word of data • "direct mapped" For each item of data at the lower level, there is exactly one location in the cache where it might be. e.g., lots of items at the lower level share locations in the upper level

  15. Figure 7.4 The cache just before and just after a reference to a word Xn that is not initially in the cache.

  16. Direct Mapped Cache • Mapping: address is modulo the number of blocks in the cache

  17. Figure 7.6 The cache contents are shown after each reference request that misses, with the index and tag fields shown in binary. a. The initial state of the cache after power-on b. After handling a miss of address ( )

  18. c. After handling a miss of address ( ) d. After handling a miss of address ( )

  19. e. After handling a miss of address ( ) f. After handling a miss of address ( )

  20. Direct Mapped Cache • For MIPS: What kind of locality are we taking advantage of?

  21. Bits in Cache • How many total bits are required for a direct-mapped cache with 16 KB of data and 4-word clocks, assuming a 32-bit address? • AnswerWe know that 16 KB is 4 K words, which is words, and, with a block size of 4 words , blocks. Each block has or 128 bits of data plus a tag, which is 32-10-2-2 bits, plus a valid bit. Thus, the total cache size isor 18.4 KB for a 16 KB cache. For this cache, the total number of bits in the cache is about 1.15 times as many as needed just for the storage of the data.

  22. Mapping an Address to a Multiword Cache Block • Consider a cache with 64 blocks and a clock size of 16 bytes. What block number does byte address 1200 map to? • Answer • The address of the block is (formula on page 474): • Notice that this block address is the block containing all addresses betweenand • Thus, with 16 bytes per block, byte address 1200 is block addresswhich maps to cache block number (75 modulo 64) = 11. In fact, this block maps all addresses between 1200 and 1215.

  23. Figure 7.8 Miss rate versus block size.

  24. Direct Mapped Cache • Taking advantage of spatial locality:

  25. Hits vs. Misses • Read hits • this is what we want! • Read misses • stall the CPU, fetch block from memory, deliver to cache, restart • Write hits: • can replace data in cache and memory (write-through) • write the data only into the cache (write-back the cache later) • Write misses: • read the entire block into the cache, then write the word

  26. To understand the impact of different organizations of memory, let’s define a set of hypothetical memory access times, Assume • 1 memory bus clock cycle to send the address • 15 memory bus clock cycles for each DRAM access initiated • 1 memory bus clock cycle to send a word of dataIf we have a cache clock of four words and a one-word-wide bank of DRAMs, the miss penalty would be 1+4*15+4*1=65 memory bus clock cycles. Thus, the number of bytes transferred per bus clock cycle for a single miss would beFigure 7.11 shows three options for designing the memory system. 1. Memory is one word wide, and all access are made sequentially. 2. Increases the bandwidth to memory by widening the memory and the buses between the processor and memory. 3. The memory chips are organized in 4 banks.

  27. Hardware Issues • Make reading multiple words easier by using banks of memory • It can get a lot more complicated...

  28. The miss penalty of each memory system • A. 1+4*15+4*1 = 65 • B. With a main memory width if two words:1+2*15+2*1 = 33 four words:17 • C. 1cycle to transmit the address and read request to the banks, 15 cycles for all four banks to access memory, and 4 cycles to send the four words back to the cache. 1+1*15+4*1 = 20

  29. Figure 7.12 DRAM size increased by multiples of four approximately once every three years until 1996, and thereafter doubling approximately every two years.

  30. 7.3 Measuring and Improving Cache Performance

  31. Keywords • Fully associative cacheA cache structure in which a block can be placed in any location in the cache. • Set-associative cacheA cache that has a fixed number of locations (at least two) where each block can be placed. • Least recently used (LRU)A replacement scheme in which the block replaced is the one that has been unused for the longest time. • Multilevel cacheA memory hierarchy with multiple levels of caches, rather than just a cache and main memory. • Global miss rateThe fraction of references that miss in all levels of a multilevel cache. • Local miss rateThe fraction of references to one level of a cache that miss; used in multilevel hierarchies.

  32. Calculating Cache Performance • Assume an instruction cache miss rate for a program is 2 % and a data cache miss rate is 4%. If a processor has a CPI of 2 without and memory stalls and the miss penalty is 100 cycles for all misses, determine how much faster a processor would run with a perfect cache that never missed. Use the instruction frequencies for SPECint2000 from Chapter 3, Figure 3.26, on page 228. • Answer

  33. Cache Performance with Increased Clock Rate • Suppose we increase the performance of the computer in the previous example by doubling its clock rate. Since the main memory speed is unlikely to change, assume that the absolute time to handle a cache miss does not change. How much faster will the computer be with faster clock, assuming the same miss rate as the previous example? • Answer Thus, the faster computer with cache misses will have a CPI of 2 + 6.88 = 8.88, compared to a CPI with cache misses of 5.44 for the slower computer. Using the formula for APU time from the previous example, we can compute the relative performance as

  34. Performance • Increasing the block size tends to decrease miss rate: • Use split caches because there is more spatial locality in code:

  35. Performance • Simplified model: execution time = (execution cycles + stall cycles) ´ cycle time stall cycles = # of instructions ´ miss ratio ´ miss penalty • Two ways of improving performance: • decreasing the miss ratio • decreasing the miss penalty What happens if we increase block size?

  36. Decreasing miss ratio with associativity Compared to direct mapped, give a series of references that: • results in a lower miss ratio using a 2-way set associative cache • results in a higher miss ratio using a 2-way set associative cache assuming we use the “least recently used” replacement strategy

  37. An implementation

  38. Performance

  39. Decreasing miss penalty with multilevel caches • Add a second level cache: • often primary cache is on the same chip as the processor • use SRAMs to add another cache above primary memory (DRAM) • miss penalty goes down if data is in 2nd level cache • Example: • CPI of 1.0 on a 5 Ghz machine with a 5% miss rate, 100ns DRAM access • Adding 2nd level cache with 5ns access time decreases miss rate to .5% • Using multilevel caches: • try and optimize the hit time on the 1st level cache • try and optimize the miss rate on the 2nd level cache

  40. Cache Complexities • Not always easy to understand implications of caches: Theoretical behavior of Radix sort vs. Quicksort Observed behavior of Radix sort vs. Quicksort

  41. Cache Complexities • Here is why: • Memory system performance is often critical factor • multilevel caches, pipelined processors, make it harder to predict outcomes • Compiler optimizations to increase locality sometimes hurt ILP • Difficult to predict best algorithm: need experimental data

  42. 7.4 Virtual Memory

  43. Keywords • Virtual memoryA technique that uses main memory as a “cache” for secondary storage. • Physical addressAn address in main memory. • Protection A set of mechanisms for ensuring that multiple processes sharing the processor, memory, or I/O devices cannot interfere, intentionally or unintentionally, with one another by reading or writing each other’s data. These mechanisms also isolate the operating system from a user process. • Page faultAn event that occurs when an accessed page is not present in main memory. • Virtual addressAn address that corresponds to a location in virtual space and is translated by address mapping to a physical address when memory is accessed.

  44. Keywords • Address translationAlso called address mapping. The process by which a virtual address is mapped to an address used to access memory. • Segmentation A variable-size address mapping scheme in which an address consists of two parts: a segment number, which is mapped to a physical address, and a segment offset. • Page tableThe table containing the virtual to physical address translations in a virtual memory system. The table, which is stored in memory, is typically indexed by the virtual page number; each entry in the table contains the physical page number for that virtual page if the page is currently in memory. • Swap spaceThe space on the disk reserved for the full virtual memory space of a process. • Reference bit Also called use bit. A field that is set whenever a page is accessed and that is used to implement LRU or other replacement schemes.

  45. Keywords • Translation-lookaside buffer (TLB)A cache that keeps track of recently used address mappings to avoid an access to the page table. • Virtually addressed cacheA cache that is accessed with a virtual address rather than a physical address. • Aliasing A situation in which the same object is accessed by two addresses; can occur in virtual memory when there are two virtual addresses for the same physical page. • Physically addressed cacheA cache that is addressed by a physical address. • Kernel modeAlso called supervisor mode. A mode indicating that a running process is an operating system process. • System callA special instruction that transfers control from user mode to a dedicated location in supervisor code space, invoking the exception mechanism in the process.

  46. Keywords • Context switchA changing of the internal state of the processor to allow a different process to use the processor that includes saving the state needed to return to the currently executing process. • Exception enableAlso called interrupt enable. A signal or action that controls whether the process responds to an exception or not; necessary for preventing the occurrence of exceptions during intervals before the processor has safely saved the state needed to restart. • Restartable instructionAn instruction that can resume execution after an exception is resolved without the exception’s affecting the result of the instruction. • Handler Name of a software routine invoked to “handle” an exception or interrupt. • Unmapped A portion of the address space that cannot have page faults.

More Related