1 / 43

컴파일러

컴파일러. 첫째주 2013/09/02. 권혁철. 과제. It might be the biggest program you ’ ve ever written. It cannot be done the day it ’ s due! Syllabus 에 있는 대로 따라 올 것 개인별 레포트 기본적인 요구조건을 충족해야 함. 팀 구성 / 과제. 과제 컴파일러를 만드는 일반 과정을 따르면 됨 도구 (LEX, YACC) 을 사용해도 됨 어셈블리어는 되도록 Pentium 용일 것. 제한조건.

mitch
Download Presentation

컴파일러

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. 컴파일러 첫째주 2013/09/02 권혁철

  2. 과제 • It might be the biggest program you’ve ever written. • It cannot be done the day it’s due! • Syllabus에 있는 대로 따라 올 것 • 개인별 레포트 • 기본적인 요구조건을 충족해야 함

  3. 팀 구성/과제 • 과제 • 컴파일러를 만드는 일반 과정을 따르면 됨 • 도구(LEX, YACC)을 사용해도 됨 • 어셈블리어는 되도록 Pentium용일 것

  4. 제한조건 • 형선언, 사칙연산, if문, for문은 포함함 • 부프로그램이나 기타 기능 추가 시 이를 평가하여 가점을 줌 • 가상어셈블리어 코드를 만들고, 인터프리터를 만들어도 됨 • Pentium용 어셈블리어나 Byte-code로 결과를 출력하여 수행되면 이를 인정하여 가점을 줌 • 언어 선택이나 문법 선택, 언어정의 따위는 각자 알아서 함 • Lex나 Yacc을 사용해도 됨 • 단, 이를 이용하지 않으면 이를 인정하여 가점을 줌, Parse table은 Yacc의 것을 이용할 수 있음

  5. Making Languages Usable It was our belief that if FORTRAN, during its first months, were to translate any reasonable “scientific” source program into an object program only half as fast as its hand-coded counterpart, then acceptance of our system would be in serious danger... I believe that had we failed to produce efficient programs, the widespread use of languages like FORTRAN would have been seriously delayed. — John Backus 18 person-years to complete!!!

  6. Compiler construction • Compiler writing is perhaps the most pervasive topic in computer science, involving many fields: • Programming languages • Architecture • Theory of computation • Algorithms • Software engineering • In this course, you will put everything you have learned together. Exciting, right??

  7. 간단한 질문 • Consider the grammar shown below(<S> is your start symbol). Circle which of the strings shown on the below are in the language described by the grammar? There may be zero or more correct answers. • Grammar: • <S> ::= <A> a <B> b • <A> ::= b <A> | b • <B> ::= <A> a | a • Strings: A) baab B) bbbabb C) bbaaaa D) baaabb E) bbbabab • Compose the grammar for the language consisting of sentences of an equal number of a’s followed by an equal number of b’s. For example, aaabbb is in the language, aabbb is not, the empty string is not in the language.

  8. What is a compiler? Compiler TargetProgram SourceProgram ErrorMessage • The source language might be • General purpose, e.g. C or Pascal • A “little language” for a specific domain, e.g. SIML • The target language might be • Some other programming language • The machine language of a specific machine

  9. 관련 용어 What is an interpreter? A program that reads an executable program and produces the results of executing that program Target Machine: machine on which compiled program is to be run Cross-Compiler: compiler that runs on a different type of machine than is its target Compiler-Compiler: a tool to simplify the construction of compilers (YACC/JCUP)

  10. Is it hard?? • In the 1950s, compiler writing took an enormous amount of effort. • The first FORTRAN compiler took 18 person-years • Today, though, we have very good software tools • You will write your own compiler in a team of 3 in one semester!

  11. Intrinsic interest Compiler construction involves ideas from many different parts of computer science Greedy algorithms Heuristic search techniques Artificial intelligence Graph algorithms, union-find Dynamic programming Algorithms DFAs & PDAs, pattern matching Fixed-point algorithms Theory Allocation & naming, Synchronization, locality Systems Pipeline & hierarchy management Instruction set use Architecture

  12. Intrinsic merit Compiler construction poses challenging and interesting problems: Compilers must do a lot but also run fast Compilers have primary responsibility for run-time performance Compilers are responsible for making it acceptable to use the full power of the programming language Computer architects perpetually create new challenges for the compiler by building more complex machines Compilers must hide that complexity from the programmer Success requires mastery of complex interactions

  13. Implications Must recognize legal (and illegal) programs Must generate correct code Must manage storage of all variables (and code) Must agree with OS & linker on format for object code High-level View of a Compiler Source code Machine code Compiler Errors

  14. Two Pass Compiler • We break compilation into two phases: • ANALYSIS breaks the program into pieces and creates an intermediate representation of the source program. • SYNTHESIS constructs the target program from the intermediate representation. • Sometimes we call the analysis part the FRONT END and the synthesis part the BACK END of the compiler. They can be written independently.

  15. Traditional Two-pass Compiler Implications Use an intermediate representation (IR) Front end maps legal source code into IR Back end maps IR into target machine code Admits multiple front ends & multiple passes (better code) Typically, front end is O(n) or O(n log n), while back end is NP-Complete Back End Machine code Source code IR Front End Errors

  16. Can we build n x m compilers with n+m components? Must encode all language specific knowledge in each front end Must encode all features in a single IR Must encode all target specific knowledge in each back end Limited success in systems with very low-level IRs A Common Fallacy Front end Front end Front end Back end Back end Back end Fortran SPARC Front end Scheme i86 Java Power PC Smalltalk

  17. Source code analysis • Analysis is important for many applications besides compilers: • STRUCTURE EDITORS try to fill out syntax units as you type • PRETTY PRINTERS highlight comments, indent your code for you, and so on • STATIC CHECKERS try to find programming bugs without actually running the program • INTERPRETERS don’t bother to produce target code, but just perform the requested operations (e.g. Matlab)

  18. Source code analysis • Analysis comes in three phases: • LINEAR ANALYSIS processes characters left-to-right and groups them into TOKENS • HIERARCHICAL ANALYSIS groups tokens hierarchically into nested collections of tokens • SEMANTIC ANALYSIS makes sure the program components fit together, e.g. variables should be declared before they are used

  19. Linear (lexical) analysis The linear analysis stage is called LEXICAL ANALYSIS or SCANNING. Example: position = initial + rate * 60 gets translated as: • he IDENTIFIER “position” • The ASSIGNMENT SYMBOL “=” • The IDENTIFIER “initial” • The PLUS OPERATOR “+” • The IDENTIFIER “rate” • The MULTIPLICATION OPERATOR “*” • The NUMERIC LITERAL 60

  20. Hierarchical (syntax) analysis • The hierarchical stage is called SYNTAX ANALYSIS or PARSING. • The hierarchical structure of the source program can be represented by a PARSE TREE, for example:

  21. assignment statement identifier = expression position expression + expression identifier expression * expression initial identifier identifier rate 60

  22. Syntax analysis • The hierarchical structure of the syntactic units in a programming language is normally represented by a set of recursive rules. Example for expressions: • Any identifier is an expression • Any number is an expression • If expression1 and expression2 are expressions, so are expression1 + expression2 expression1 * expression2 ( expression1 )

  23. Syntax analysis • Example for statements: • If identifier1 is an identifier and expression2 is an expression, then identifier1 = expression2 is a statement. • If expression1 is an expression and statement2 is a statement, then the following are statements: while ( expression1 ) statement2 if ( expression1 ) statement2

  24. Lexical vs. syntactic analysis • Generally if a syntactic unit can be recognized in a linear scan, we convert it into a token during lexical analysis. • More complex syntactic units, especially recursive structures, are normally processed during syntactic analysis (parsing). • Identifiers, for example, can be recognized easily in a linear scan, so identifiers are tokenized during lexical analysis.

  25. Source code analysis • It is common to convert complex parse trees to simpler SYNTAX TREES, with a node for each operator and children for the operands of each operator. position = initial + rate * 60 = position + Analysis initial * rate 60

  26. Semantic analysis • The semantic analysis stage: • Checks for semantic errors, e.g. undeclared variables • Gathers type information • Determines the operators and operands of expressions Example: if rate is a float, the integer literal 60 should be converted to a float before multiplying. = position + initial * rate inttoreal 60

  27. source program The rest of the process lexicalanalyzer syntaxanalyzer semanticanalyzer symbol-tablemanager errorhandler intermediatecode generator codeoptimizer codegenerator target program

  28. Symbol-table management • During analysis, we record the identifiers used in the program. • The symbol table stores each identifier with its ATTRIBUTES. • Example attributes: • How much STORAGE is allocated for the id • The id’s TYPE • The id’s SCOPE • For functions, the PARAMETER PROTOCOL • Some attributes can be determined immediately; some are delayed.

  29. Error detection • Each compilation phase can have errors • Normally, we want to keep processing after an error, in order to find more errors. • Each stage has its own characteristic errors, e.g. • Lexical analysis: a string of characters that do not form a legal token • Syntax analysis: unmatched { } or missing ; • Semantic: trying to add a float and a pointer

  30. position = initial + rate * 60 lexical analyzer Internal Representations Each stage of processing transforms a representation of the source code program into a new representation. id1 = id2 + id3 * 60 syntax analyzer symbol table = id1 + * id2 id3 60 semantic analyzer = id1 + * id2 id3 inttoreal 60

  31. Intermediate code generation • Some compilers explicitly create an intermediate representation of the source code program after semantic analysis. • The representation is as a program for an abstract machine. • Most common representation is “three-address code” in which all memory locations are treated as registers, and most instructions apply an operator to two operand registers, and store the result to a destination register.

  32. Intermediate code generation = position + initial * rate inttoreal 60 semantic analyzer temp1 := inttoreal(60) temp2 := id3 * temp1 temp3 := id2+ temp2 id1 := temp3

  33. The Optimizer (or Middle End) Typical Transformations  performance, code size, power consumptionetc Discover & propagate some constant value Move a computation to a less frequently executed place Specialize some computation based on context Discover a redundant computation & remove it Remove useless or unreachable code Encode an idiom in some particularly efficient form IR IR IR ... IR IR Opt 1 Opt 3 Opt 2 Opt n Errors Modern optimizers are structured as a series of passes

  34. Code optimization • At this stage, we improve the code to make it run faster. temp1 := inttoreal(60) temp2 := id3 * temp1 temp3 := id2+ temp2 id1 := temp3 temp1 := id3 * 60.0 id1 := id2 + temp1 code optimizer

  35. Code generation • In the final stage, we take the three-address code (3AC) or other intermediate representation, and convert to the target language. • We must pick memory locations for variables and allocate registers. MOVF id3, R2 MULF #60.0, R2 MOVF id2, R1 ADDF R2, R1 MOVF R1, id1 temp1 := id3 * 60.0 id1 := id2 + temp1 code generator

  36. The Back End Responsibilities Translate IR into target machine code Choose instructions to implement each IR operation Decide which value to keep in registers Ensure conformance with system interfaces Automation has been less successful in the back end IR IR IR Machine code Instruction Scheduling Instruction Selection Register Allocation Errors

  37. The Back End Instruction Selection Produce fast, compact code Take advantage of target features such as addressing modes Usually viewed as a pattern matching problem ad hoc methods, pattern matching, dynamic programming IR IR IR Instruction Selection Register Allocation Machine code Instruction Scheduling Errors

  38. The Back End Register Allocation Have each value in a register when it is used Manage a limited set of resources Can change instruction choices & insert LOADs & STOREs Optimal allocation is NP-Complete (1 or k registers) Compilers approximate solutions to NP-Complete problems IR IR IR Register Allocation Machine code Instruction Scheduling Instruction Selection Errors

  39. The Back End Instruction Scheduling Avoid hardware stalls and interlocks Use all functional units productively Can increase lifetime of variables (changing the allocation) Optimal scheduling is NP-Complete in nearly all cases Heuristic techniques are well developed IR IR IR Instruction Scheduling Machine code Instruction Selection Register Allocation Errors

  40. Cousins of the compiler • PREPROCESSORS take raw source code and produce the input actually read by the compiler • MACRO PROCESSING: macro calls need to be replaced by the correct text • Macros can be used to define a constant used in many places. E.g. #define BUFSIZE 100 in C • Also useful as shorthand for often-repeated expressions:#define DEG_TO_RADIANS(x) ((x)/180.0*M_PI)#define ARRAY(a,i,j,ncols) ((a)[(i)*(ncols)+(j)]) • FILE INCLUSION: included files (e.g. using #include in C) need to be expanded

  41. Cousins of the compiler • ASSEMBLERS take assembly code and covert to machine code. • Some compilers go directly to machine code; others produce assembly code then call a separate assembler. • Either way, the output machine code is usually RELOCATABLE, with memory addresses starting at location 0.

  42. Cousins of the compiler • LOADERS take relocatable machine code and alter the addresses, putting the instructions and data in a particular location in memory. • The LINK EDITOR (part of the loader) pieces together a complete program from several independently compiled parts.

  43. Compiler writing tools • We’ve come a long way since the 1950s. • SCANNER GENERATORS produce lexical analyzers automatically. • Input: a specification of the tokens of a language (usually written as regular expressions) • Output: C code to break the source language into tokens. • PARSER GENERATORS produce syntactic analyzers automatically. • Input: a specification of the language syntax (usually written as a context-free grammar) • Output: C code to build the syntax tree from the token sequence. • There are also automated systems for code synthesis.

More Related