1 / 44

Comparison of Modern CPUs and GPUs And the convergence of both

Comparison of Modern CPUs and GPUs And the convergence of both. Jonathan Palacios Josh Triska. Introduction and Motivation. Graphics Processing Units (GPUs) have been evolving at a rapid rate in recent years In terms of raw processing power gains, they greatly outpace CPUs.

ova
Télécharger la présentation

Comparison of Modern CPUs and GPUs And the convergence of both

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Comparison of Modern CPUs and GPUsAnd the convergence of both Jonathan Palacios Josh Triska

  2. Introduction and Motivation • Graphics Processing Units (GPUs) have been evolving at a rapid rate in recent years • In terms of raw processing power gains, they greatly outpace CPUs

  3. Introduction and Motivation

  4. Introduction and Motivation • Disparity is largely due to the specific nature of problems historically solved by the GPU • Same operations on many primitives (SIMD) • Focus on throughput over Latency • Lots of special purpose hardware • CPUs On the the other hand: • Focus on reducing Latency • Designed to handle a wider range of problems

  5. Introduction and Motivation • Despite differences, we've found that GPUs and CPUs are converging in many ways: • CPUs are adding more cores • GPUs becoming more programmable, general purpose • Examples • NVIDIA Fermi • Intel Larrabee

  6. Overview • Introduction • History of GPU • Chip Layouts • Data-flow • Memory Hierarchy • Instruction Set • Applications • Conclusion

  7. History of the GPU • GPUs have mostly developed in the last 15 years • Before that, graphics handled by Video Graphics Array (VGA) Controller • Memory controller, DRAM, display generator • Takes image data, and arranges it for output device

  8. History of the GPU • Graphics Acceleration hardware components were gradually added to VGA controllers • Triangle rasterization • Texture mapping • Simple shading • Examples of early “graphics accelerators” • 3dfx Voodoo • ATI Rage • NIVDIA RIVA TNT2

  9. History of the GPU • NVIDIA GeForce 256 “first” GPU (1999) • Non-programmable (fixed-function) • Transforming and Lighting • Texture/Environment Mapping

  10. History of the GPU • Fairly early on in the GPU market, there was a severe narrowing of competition • Early companies: • Silicon Graphics International • 3dfx • NVIDIA • ATI • Matrox • Now only AMD and NVIDIA

  11. History of the GPU • Since their inception, GPUs have gradually become more powerful, programmable, and general purpose • Programmable geometry, vertex and pixel processors • Unified Shader Model • Expanding instruction set • CUDA, OpenCL

  12. History of the GPU • The latest NVIDIA Architecture, Fermi offers many more general purpose features • Real floating point quality and performance • Error Correcting Codes • Fast context switching • Unified address space

  13. GPU Chip Layouts • GPU Chip layouts have been moving in the direction of general purpose computing for several years • Some High-level trends • Unification of hardware components • Large increases in functional unit counts

  14. GPU Chip Layouts NVIDIA GeForce 7800

  15. GPU Chip Layouts NVIDIA GeForce 8800

  16. GPU Chip Layouts NVIDIA GeForce 400 (Fermi architecture) 3 billion transisors

  17. GPU Chip Layouts AMD Radeon 6800 (Cayman architecture) 2.64 billion transisors

  18. CPU Chip Layouts • CPUs have also been increasing functional unit counts • However, these units are always added with all of the hardware fanfare that would come with a single core processor • Reorder buffers/reservations stations • Complex branch prediction • This means that CPUs add raw compute power at a much slower rate.

  19. CPU Chip Layouts Intel Core i7 (Nehalem architecture) 125 million transistors

  20. CPU Chip Layouts Intel Core i7 (Nehalem architecture) 731 million transistors

  21. CPU Chip Layouts Nehalem “core” 731 million transistors

  22. CPU Chip Layouts Intel Westmere (Nehalem)

  23. CPU Chip Layouts Intel 8-Core Nehalem EX 2.3 Billion transistors

  24. “Hybrid” Chip Layouts Intel Larrabee project Vaporware

  25. “Hybrid” Chip Layouts NVIDIA Tegra

  26. Chip Layouts Summary • The take-home message is that the real-estate allocation of GPUs and CPUs evolve based on very different fundamental priorities • GPUs • Increase raw compute power • Increase throughput • Still fairly special purpose • CPUs • Reduce Latency • Epitome of general purpose • Backwards compatibility

  27. The (traditional) graphics pipeline Programmable Since 2000 • Programmable elements of the graphics pipeline were historically fixed-function units, until the year 2000

  28. The unified shader • With the introduction of the unified shader model, the GPU becomes essentially a many-core, streaming multiprocessor Nvidia 6800 tech brief

  29. Emphasis on throughput • If your frame rate is 50 Hz, your latency can be approximately 2 ms  • However, you need to do 100 million operations for that one frame  • Result: very deep pipelines and high FLOPS • GeForce 7 had >200 stages for the pixel shader • Fermi: 1.5 TFLOPS, AMD 5870: 2.7 TFLOPS • Unified shader has cut down on the number of stages by allowing breaks from linear execution

  30. Memory hierarchy • Cache size hierarchy caches is backwards from that of CPUs • Caches serve to conserve precious memory bandwidth by intelligently prefetching

  31. Memory prefetching Prefetching algorithm • Graphics pipelines are inherently high-latency • Cache misses simply push another thread into the core • Hit rates of ~90%, as opposed to ~100%

  32. Memory access • GPUs are all about 2D spatial locality, not linear locality • GPU caches read- only (uses registers) • Growing body of research optimizing algorithms for 2D cache model

  33. Instruction set differences • Until very recently, scattered address space • 2009 saw the introduction of modern CPU-style 64-bit addressing • Block operations versus sequential block = 1:4 by 1:4 if y[i][j] = within block y[i][j] = y[i][j] + 1 for i = 1 to 4 for j = 1 to 4 y[i][j] = y[i][j] + 1 Bam! • SIMD: single instruction, multiple data

  34. SIMD vs. SISD Programmable GPU shaders versus Pentium 4

  35. Single Instruction, Multiple Thread (SIMT) Weaving cotton threads • Newer GPUs are using a new kind of scheduling model called SIMT • ~32 threads are bundled together in a “warp” and executed together • Warps are then executed 1 instruction at a time, round robin

  36. Instruction set differences • Branch granularity • If one thread within a processor cluster branches without the rest, you have a branch divergence • Threads become serial until branches converge • Warp scheduling improves, not eliminates, hazards from branch divergence • if/else may stall threads

  37. Instruction set differences • Unified shader • All shaders (since 2006) have the same basic instruction set layered on a (still) specialized core • Cores are very simple: hardware support for things like recursion may not be available • Until very recently, dealing with speed hacks • Floating-point accuracy truncated to save cycles • IEEE FP specs are appearing on some GPUs • Primitives limited to GPU data structures • GPUs operate on textures, etc • Computational variables must be mapped

  38. GPU Limitations • Relatively small amount of memory, < 4GB in current GPUs • I/O directly to GPU memory has complications • Must transfer to host memory, and then back • If 10% of instructions are LD/ST and other instructions are... • 10 times faster 1/(.1 + .9/10) ≈ speedup of 5 • 100 times faster 1/(.1 + .9/100) ≈ speedup of 9

  39. Applications – real-time physics

  40. Applications – protein folding

  41. Applications – fluid dynamics

  42. Applications – bitonic sorting

  43. Applications – n-body problems

  44. Conclusion • GPUs and CPUs fill different niches in the market for high performance architecture. • GPUs: Large throughput; latency hidden; fairly simple, but costly programs; special purpose • CPUs: Low latency; complex programs; general purpose • Both will likely always be needed; combinations of CPUs and GPUs can be much faster than either alone • CPUs are becoming multi-core and parallel • GPUs are adding general-purpose cores

More Related