1 / 31

Architectures and instruction sets

Architectures and instruction sets. Computer architecture taxonomy. Assembly language. von Neumann architecture. Memory holds data, instructions. Central processing unit (CPU) fetches instructions from memory. Separate CPU and memory distinguishes programmable computer.

shonaker
Download Presentation

Architectures and instruction sets

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Architectures and instruction sets • Computer architecture taxonomy. • Assembly language. Overheads for Computers as Components

  2. von Neumann architecture • Memory holds data, instructions. • Central processing unit (CPU) fetches instructions from memory. • Separate CPU and memory distinguishes programmable computer. • CPU registers help out: program counter (PC), instruction register (IR), general-purpose registers, etc. Overheads for Computers as Components

  3. CPU + memory memory address CPU PC 200 data ADD r5,r1,r3 ADD r5,r1,r3 IR 200 Overheads for Computers as Components

  4. Harvard architecture address CPU data memory PC data address program memory data Overheads for Computers as Components

  5. von Neumann vs. Harvard • Harvard can’t use self-modifying code. • Harvard allows two simultaneous memory fetches. • Most DSPs use Harvard architecture for streaming data: • greater memory bandwidth; • more predictable bandwidth. Overheads for Computers as Components

  6. RISC vs. CISC • Complex instruction set computer (CISC): • many addressing modes; • many operations. • Reduced instruction set computer (RISC): • load/store; • pipelinable instructions. Overheads for Computers as Components

  7. Instruction set characteristics • Fixed vs. variable length. • Addressing modes. • Number of operands. • Types of operands. Overheads for Computers as Components

  8. Programming model • Programming model: registers visible to the programmer. • Some registers are not visible (IR). Overheads for Computers as Components

  9. Multiple implementations • Successful architectures have several implementations: • varying clock speeds; • different bus widths; • different cache sizes; • etc. Overheads for Computers as Components

  10. Assembly language • One-to-one with instructions (more or less). • Basic features: • One instruction per line. • Labels provide names for addresses (usually in first column). • Instructions often start in later columns. • Columns run to end of line. Overheads for Computers as Components

  11. ARM assembly language example label1 ADR r4,c LDR r0,[r4] ; a comment ADR r4,d LDR r1,[r4] SUB r0,r0,r1 ; comment destination Overheads for Computers as Components

  12. Pseudo-ops • Some assembler directives don’t correspond directly to instructions: • Define current address. • Reserve storage. • Constants. Overheads for Computers as Components

  13. Pipelining • Execute several instructions simultaneously but at different stages. • Simple three-stage pipe: memory fetch decode execute Overheads for Computers as Components

  14. Pipeline complications • May not always be able to predict the next instruction: • Conditional branch. • Causes bubble in the pipeline: fetch decode Execute JNZ fetch decode execute fetch decode execute Overheads for Computers as Components

  15. Superscalar • RISC pipeline executes one instruction per clock cycle (usually). • Superscalar machines execute multiple instructions per clock cycle. • Faster execution. • More variability in execution times. • More expensive CPU. Overheads for Computers as Components

  16. Simple superscalar • Execute floating point and integer instruction at the same time. • Use different registers. • Floating point operations use their own hardware unit. • Must wait for completion when floating point, integer units communicate. Overheads for Computers as Components

  17. Costs • Good news---can find parallelism at run time. • Bad news---causes variations in execution time. • Requires a lot of hardware. • n2 instruction unit hardware for n-instruction parallelism. Overheads for Computers as Components

  18. Finding parallelism • Independent operations can be performed in parallel: ADD r0, r0, r1 ADD r2, r2, r3 ADD r6, r4, r0 r1 r2 r3 r0 + + r3 r4 r0 + r6 Overheads for Computers as Components

  19. Pipeline hazards • Two operations that require the same resource cannot be executed in parallel: x = a + b; a = d + e; y = a - f; a + x f b - y d a + e Overheads for Computers as Components

  20. Scoreboarding • Scoreboard keeps track of what instructions use what resources: Overheads for Computers as Components

  21. Order of execution • In-order: • Machine stops issuing instructions when the next instruction can’t be dispatched. • Out-of-order: • Machine will change order of instructions to keep dispatching. • Substantially faster but also more complex. Overheads for Computers as Components

  22. VLIW architectures • Very long instruction word (VLIW) processing provides significant parallelism. • Rely on compilers to identify parallelism. Overheads for Computers as Components

  23. What is VLIW? • Parallel function units with shared register file: register file function unit function unit function unit function unit ... instruction decode and memory Overheads for Computers as Components

  24. VLIW cluster • Organized into clusters to accommodate available register bandwidth: cluster cluster cluster ... Overheads for Computers as Components

  25. VLIW and compilers • VLIW requires considerably more sophisticated compiler technology than traditional architectures---must be able to extract parallelism to keep the instructions full. • Many VLIWs have good compiler support. Overheads for Computers as Components

  26. Static scheduling a b e f a b e c g f c nop d d g nop expressions instructions Overheads for Computers as Components

  27. Trace scheduling conditional 1 Rank paths in order of frequency. block 2 block 3 Schedule paths in order of frequency. loop head 4 loop body 5 Overheads for Computers as Components

  28. EPIC • EPIC = Explicitly parallel instruction computing. • Used in Intel/HP Merced (IA-64) machine. • Incorporates several features to allow machine to find, exploit increased parallelism. Overheads for Computers as Components

  29. IA-64 instruction format • Instructions are bundled with tag to indicate which instructions can be executed in parallel: 128 bits tag instruction 1 instruction 2 instruction 3 Overheads for Computers as Components

  30. Memory system • CPU fetches data, instructions from a memory hierarchy: Main memory CPU L2 cache L1 cache Overheads for Computers as Components

  31. Memory hierarchy complications • Program behavior is much more state-dependent. • Depends on how earlier execution left the cache. • Execution time is less predictable. • Memory access times can vary by 100X. Overheads for Computers as Components

More Related