Dr. J's Compiler and Translator Design Lecture Notes

(C) Copyright 2011-2025 by Clinton Jeffery and/or original authors where appropriate. For use in Dr. J's Compiler classes only. Lots of material in these notes originated with Saumya Debray's Compiler course notes from the University of Arizona, for which I owe him a debt of thanks. Various portions of his notes were in turn inspired by the ASU red dragon book.
  • Intermediate Code Generation
  • Final Code Generation
  • Optimization
  • Final Exam Review
  • lecture #1 began here

    Syllabus

    Yes, go over the syllabus.

    Announcements

    Comments on Implementation Languages

    New This Year

    The target language we are writing a subset for this year is a subset of Kotlin that I will call k0.

    Reading!

    1. Read the Thain text chapters 1-3. Within the Scanning chapter, there are portions on the finite automata that should be CSE 342 review; you may SKIM that material, unless you don't know it or don't remember it, in which case you should READ in detail.
    2. If you have BYOPL, you may want to read the BYOPL text, chapters 1-3. You can sort of skim chapter 1-2. Chapter 3 is important for the next homework(s) #2+.
    3. Read Sections 3-5 of the Flex manual, Lexical Analysis With Flex.
    4. Read the class lecture notes as fast as we manage to cover topics. Please ask questions about whatever is not totally clear. You can Ask Questions in class or via e-mail.

    Although the whole course's lecture notes are ALL available to you up front, I generally revise each lecture's notes, making additions, corrections and adaptations to this year's homeworks, the night before each lecture. The best time to print hard copies of the lecture notes, if you choose to do that, is one day at a time, right before the lecture is given. Or just read online.

    Why study compilers?

    Computer scientists study compiler construction for the following reasons: CSE 423 is labor intensive. This is a good thing. There is no way to learn the skills necessary for writing big programs without this kind of labor-intensive experience.

    Some Tools we will use

    Labs and lectures will discuss all of these, but if you do not know them already, the sooner you go learn them, the better.
    C and "make".
    You should already be intermediate proficiency in these tools. You should expect to practice and gain expertise with these tools as part of passing this class.
    lex and yacc
    These are compiler-writers' tools, but are used for other kinds of applications, almost anything with a complex file format to read in can benefit from them.
    gdb and valgrind
    You will need to know a source-level like gdb debugger well to survive this class. If you have never used valgrind: it can find some bugs that gdb misses!
    e-mail
    Feel free to ask the instructor questions by e-mail, or request a zoom or in-person appointment.

    Compilers - What Are They and What Kinds of Compilers are Out There?

    The purpose of a compiler is: to translate a program in some language (the source language) into a lower-level language (the target language). The compiler itself is written in some language, called the implementation language. To write a compiler you have to be very good at programming in the implementation language, and have to think about and understand the source language and target language.

    There are several major kinds of compilers:

    Native Code Compiler
    Translates source code into hardware (assembly or machine code) instructions. Example: gcc.
    Virtual Machine Compiler
    Translates source code into an abstract machine code, for execution by a virtual machine interpreter. Example: javac.
    JIT Compiler
    Translates virtual machine code to native code. Operates within a virtual machine. Example: Sun's HotSpot java machine.
    Preprocessor
    Translates source code into simpler or slightly lower level source code, for compilation by another compiler. Examples: cpp, m4.
    Pure interpreter
    Executes source code on the fly, without generating machine code. Example: Lisp.
    OK, so a pure interpreter is not really a compiler. Here are some more tools, by way of review, that compiler people might be directly concerned with, even if they are not themselves compilers. You should learn any of these terms that you don't already know.
    assembler
    a translator from human readable (ASCII text) files of machine instructions into the actual binary code (object files) of a machine.
    linker
    a program that combines (multiple) object files to make an executable. Converts names of variables and functions to numbers (machine addresses).
    loader
    Program to load code. On some systems, different executables start at different base addresses, so the loader must patch the executable with the actual base address of the executable.
    preprocessor
    Program that processes the source code before the compiler sees it. Usually, it implements macro expansion, but it can do much more.
    editor
    Editors may operate on plain text, or they may be wired into the rest of the compiler, highlighting syntax errors as you go, or allowing you to insert or delete entire syntax constructs at a time.
    debugger
    Program to help you see what's going on when your program runs. Can print the values of variables, show what procedure called what procedure to get where you are, run up to a particular line, run until a particular variable gets a special value, etc.
    profiler
    Program to help you see where your program is spending its time, so you can tell where you need to speed it up.

    Phases of a Compiler

    Lexical Analysis:
    Converts a sequence of characters into words, or tokens
    Syntax Analysis:
    Converts a sequence of tokens into a parse tree
    Semantic Analysis:
    Manipulates parse tree to verify symbol and type information
    Intermediate Code Generation:
    Converts parse tree into a sequence of intermediate code instructions
    Optimization:
    Manipulates intermediate code to produce a more efficient program
    Final Code Generation:
    Translates intermediate code into final (machine/assembly) code

    lecture #2 began here

    Kotlin of the Day

    Wow. Thank you to those of you who have already turned in HW#1!

    What some of you told me about Kotlin so far:

    variables are declared using a var keyword. Does that mean Kotlin is dynamically typed?
    statically typed and type inferencing are not mutually exclusive. Our "subset" of Kotlin, k0, can be really (and more restrictively, and explicitly) statically typed. "var" keyword was used in Pascal, and Pascal is statically typed...
    Kotlin has a different perspective than C/C++ on "const"-ness
    There is this whole paradigm (functional programming) and this whole family of languages (single assignment languages such as SISAL) that says constness should be in as much of your programs as possible. So maybe Kotlin is const by default and anything non-const is flagged.
    Kotlin has nullable and non-nullable variants of (almost) all types.
    This solves(?) part of Java's null exception hell via the type system. Note that Java famously has a few built-in scalar types and then most stuff is classes/objects. Is Kotlin like that with the non-OO types and then the OO types?
    Kotlin mostly has a lot of class types, and class types will be a major challenge to writing our compiler.
    We will have to have a plan alright. But we will be requiring only a very restructed subset of classes/objects support in k0.
    Some things in kotlin that look hard to implement:

    How big will our Kotlin subset be?

    Example of the Compilation Process

    Consider the example statement; its translation to machine code illustrates some of the issues involved in compiling.
    position = initial + rate * 60
    
    30 or so characters, from a single line of source code, are first transformed by lexical analysis into a sequence of 7 tokens. Those tokens are then used to build a tree of height 4 during syntax analysis.

    Semantic analysis may transform the tree into one of height 5, that includes a type conversion necessary for real addition on an integer operand. Note that names like position, initial and rate in source code are but three instances of the kind of token that allows names, called an identifier or "id". The actual names are stored in the structure, but the terminal symbol is the same for all these names.

    Intermediate code generation uses a simple traversal algorithm to linearize the tree back into a sequence of machine-independent three-address-code instructions. This is typically done bottom-up and the linear sequence might look almost like a reversal of the tree structure, but in general it is not constrained to be just that.

      t1 = inttoreal(60)  
      t2 = id3 * t1
      t3 = id2 + t2
      id1 = t3

    Optimization of the intermediate code allows the four instructions to be reduced to two machine-independent instructions. Final code generation might implement these two instructions using 5 machine instructions in which the actual registers and addressing modes of the CPU are utilized (left column). For comparison the right column shows the same computation using a bytecode instruction set. Many bytecodes are stack-based and have no registers.

    native VM
      MOVF	id3, R2  
      MULF	#60.0, R2
      MOVF	id2, R1
      ADDF	R2, R1
      MOVF	R1, id1
    
    pnull   # push space for a result
    var 0   # load a reference to var slot 0 (position)
    pnull   # push space for a result
    var 1   # load a reference to var slot 1 (initial)
    pnull   # push space for a result
    var 2   # load a reference to var slot 2 (rate)
    int 0   # load a reference to constant slot 0 (60)
    mult    # multiply top two references
    plus    # add top two references
    asgn    # assign value from top ref into next ref

    Revision Controls Yes, Public Repositories No

    It is sensible to use software engineering tools including revision control systems such as git on a large project like writing a compiler. On the other hand it is not OK to share your work with the whole class, intentionally or through stupidity, you must keep things private or within any team where teams are explicitly approved or assigned. If you use a revision control system, figure out how to make it private and share access only with approved team members. Various options:

    Kotlin Reference

    Have you found the Kotlin Language Specification yet?

    Overview of Lexical Analysis

    A lexical analyzer, also called a scanner, typically has the following functionality and characteristics.

    A couple comments on the Lab for this course

    Labs start in earnest next Wednesday
    Generally they will run every Wednesday from 2-3pm
    Main purpose of the lab: auxiliary public Q & A and debugging help for your project.
    Bring us your coredumps, your stack overflows, your huddled parse problems, yearning to breath free.
    Auxiliary purpose of the lab: extended practice with the professional tools of the course.
    For some of you, this part will be review.
    Format of the Labs
    Usually a public Q & A followed by a lab exercise.
    Grading of the Labs
    The Labs are to help you complete your real homework (project) assignments and will be a miniscule portion of your grade, treated comparable to quizzes. Let's say, the aggregate of all labs (and quizzes, if any) will amount to 5%. A given lab might be 0.3-0.5% and is likely to be graded on a boolean (you did it or not) scsale.

    What is a "token" ?

    In compilers, a "token" is:
    1. a single word of source code input (a.k.a. "lexeme")
    2. an integer code that refers to (the category of) a single word of input
    3. a set of lexical attributes computed from a single word of input
    4. a struct (or object) that holds all those attributes
    Programmers think about all this in terms of #1. Syntax checking uses #2. Error reporting, semantic analysis, and code generation require #3. In a compiler written in C, for each token you allocate a C struct to store (3) for each token.

    Auxiliary data structures

    You were presented with the phases of the compiler, from lexical and syntax analysis, through semantic analysis, and intermediate and final code generation. Each phase has an input and an output to the next phase. But there are a few data structures we will build that survive across multiple phases: the literal table, the symbol table, and the error handler.
    lexeme table
    a table that stores lexeme values, such as strings and variable names, that may occur in many places. Only one copy of each unique string and name needs to be allocated in memory. (This is an optional memory-saving device.)
    symbol tables
    symbol tables store the names defined (and visible within) each particular scope. The most common scopes are: global, and procedure (local). Real languages have more scopes such as class (or record) and package.
    error handlers
    errors in lexical, syntax, or semantic analysis need a common reporting mechanism, that shows where the error occurred (filename, line number, and maybe column number are useful). This may entail helper functions, global variables, or entire data structures.

    lecture #3 began here

    Regular Expressions

    The notation we use to precisely capture all the variations that a given category of token may take are called "regular expressions" (or, less formally, "patterns". The word "pattern" is really vague and there are lots of other notations for patterns besides regular expressions). Regular expressions are a shorthand notation for sets of strings. In order to even talk about "strings" you have to first define an alphabet, the set of characters which can appear.
    1. Epsilon (ε) is a regular expression denoting the set containing the empty string
    2. Any letter in the alphabet is also a regular expression denoting the set containing a one-letter string consisting of that letter.
    3. For regular expressions r and s,
               r | s
      is a regular expression denoting the union of r and s
    4. For regular expressions r and s,
               r s
      is a regular expression denoting the set of strings consisting of a member of r followed by a member of s
    5. For regular expression r,
               r*
      is a regular expression denoting the set of strings consisting of zero or more occurrences of r.
    6. You can parenthesize a regular expression to specify operator precedence (otherwise, alternation is like plus, concatenation is like times, and closure is like exponentiation)

    Lex/Flex Extended Regular Expressions

    Although the basic regular expression operators given earlier are sufficient to describe all regular languages, in practice everybody uses extensions: Lex has a Lot of extended regular expressions. I have listed the minimum set that you are absolutely responsible for knowing. How many others do you remember from reading the Flex manual?







    Lex extended regular expressions

    Here is an almost-complete list of Lex's regular expressions:
    c
    normal characters mean themselves
    \c
    backslash escapes remove the meaning from most operator characters. Inside character sets and quotes, backslash performs C-style escapes.
    "s"
    Double quotes mean to match the C string given as itself. This is particularly useful for multi-byte operators and may be more readable than using backslash multiple times.
    [s]
    This character set operator matches any one character among those in s.
    [^s]
    A negated-set matches any one character not among those in s.
    .
    The dot operator matches any one character except newline: [^\n]
    r*
    match r 0 or more times.
    r+
    match r 1 or more times.
    r?
    match r 0 or 1 time.
    r{m,n}
    match r between m and n times.
    r1r2
    concatenation. match r1 followed by r2
    r1|r2
    alternation. match r1 or r2
    (r)
    parentheses specify precedence but do not match anything
    r1/r2
    lookahead. match r1 when r2 follows, without consuming r2
    ^r
    match r only when it occurs at the beginning of a line
    r$
    match r only when it occurs at the end of a line

    Avoid Common Regex Bugs

    Usually when doing your homework, you figure out if you messed up and fix it, but it can bite you on exams
    square bracket abuse
    trying to use regex operators inside square brackets; trying to use square brackets as if they were parentheses
    good old fashioned operator precedence problems
    when in doubt, use parentheses
    writing a regular expression that is too loose
    be careful especially when using a regex that matches everything; you can read the entire file with one match.

    Some Regular Expression Examples

    Regular expressions are the preferred notation for specifying patterns of characters that define token categories. The best way to get a feel for regular expressions is to see examples. Note that regular expressions form the basis for pattern matching in many UNIX tools such as grep, awk, perl, etc.

    What is the regular expression for each of the different lexical items that appear in C programs? How does this compare with another, possibly simpler programming language such as BASIC? What are the corresponding rules for our language this semester, are they the same as C?
    lexical category Kotlin BASIC C
    operators +
    -
    ..
    ..<
    ...
    ++
    ||
    ->
    =>
    +=
    the characters themselves
    such as + or -
    For operators that are regular expression operators we need mark them with double quotes or backslashes to indicate you mean the character, not the regular expression operator. Note several operators have a common prefix. The lexical analyzer needs to look ahead to tell whether an = is an assignment, or is followed by another = for example.
    reserved words if
    val
    var
    while
    break
    (In Kotlin, a buncha reserved words are only semi-reserved and usable as variable names except in certain contexts. In k0 we'll just fully reserve them.)
    the concatenation of characters; case insensitive Reserved words are also matched by the regular expression for identifiers, so a disambiguating rule is needed.
    identifiers ((Letter | '_') (Letter | '_' | UnicodeDigit)*)
    | "`"[^`\r\n]+"\`"
    We're not doing Unicode so UnicodeDigit can just become [0-9]. Should we do backtick identifiers?
    no _; $ at ends of some; 2 significant letters!?; case insensitive [a-zA-Z_][a-zA-Z0-9_]*
    numbers Kotlin has a lot, maybe extended from newer C/C++. Hex and binary, etc. We care about a 113-level amount of constants. ints and reals, starting with [0-9]+ 0x[0-9a-fA-F]+ etc.
    comments Kotlin allows nested comments, k0 we will settle for C-like non-nested. REM.* C's comments are tricky regexp's
    strings Kotlin seems to have special rules for """ triple quote strings; study those. Kotlin has unicode character escapes that k0 will not need to support. We will support \t\n etc. almost ".*"; no escapes escaped quotes
    what else?

    What are the Best Regular Expressions you can Write for k0 ?

    What corrections/improvements/additions are needed in the following? Note that the full ANSI C language would entail a lot of additions. What additions are needed for the your compiler's subset language this semester?

    Category Regular Expression
    Variable names [a-zA-Z_][a-zA-Z0-9_]*
    Integer constants "-"?[0-9]+ | "0x"[0-9A-Fa-f]+
    Real # Constants [0-9]*"."[0-9]+
    String Constants \"([^"\n]|("\\\""))*\"

    Lexing Reals

    C float and double constants have to have at least one digit, either before or after the required decimal. This is a pain:
    ([0-9]+.[0-9]* | [0-9]*.[0-9]+) ...
    
    You might almost be happier if you wrote
    ([0-9]*.[0-9]*)    { return (strcmp(yytext,".")) ? REAL : PERIOD; }
    
    Starring: C's ternary operator e1 ? e2 : e3 is an if-then-else expression, very slick. Note that if you have to support scientific/exponential real numbers (JSON does), you'll need a bigger regex.

    Lexical Attributes and Token Objects

    Besides the token's category, the rest of the compiler may need several pieces of information about a token in order to perform semantic analysis, code generation, and error handling. A lexical attribute is a piece of information about a token. These typically include:
    category an integer code used to check syntax
    lexeme actual string contents of the token
    line, column, file where the lexeme occurs in source code
    value for literals, the binary data they represent
    These are stored in an object instance of class Token, or in C, a struct. The fields are generally something like:
    struct token {
       int category;
       char *text;
       int linenumber;
       int column;
       char *filename;
       union literal value;
    }
    
    The union literal will hold computed values of integers, real numbers, and strings. In your homework assignment, I am requiring you to compute column #'s; not all compilers require them, but they are easy. Also: in our compiler project we are not worrying about optimizing our use of memory, so am not requiring you to use a union.

    Comments

    Observation

    Of the 50-100 or so categories of things that appear in the source code of most programming languages, maybe 90% of them are trivial and can be achieved by just putting the exact lexeme in double quotes, such as "while" or "<<". The remaining 10% are usually the identiers (variable names) and various literal constants. But what are the exact rules for those?

    lecture #4 began here

    Mailbag

    Tell me again, what up with K0 string literal constants?
    Well, let me see...
    I ran into a problem when I tried to return the struct token I created since yylex() returns an int and not a struct. Maybe I misinterpreted the homework instructions but can we go over how we get yylex to return a struct in class?
    yylex() returns an int. The token structure should be left in a global variable for main to pick up. In HW2 it says to use a variable named yytoken. Later we will leave it in a global variable dictated to us by the bison parser generator.

    Avoid These Common Bugs in Your Homeworks and Labs!

    1. yytext or yyinput were not declared global
    2. main() does not have its required argc, argv parameters!
    3. main() does not call yylex() in a loop or check its return value
    4. getc() EOF handling is missing or wrong! check EVERY all to getc() for EOF!
    5. opened files not (all) closed! file handle leak!
    6. end-of-comment code doesn't check for */
    7. yylex() is not doing the file reading
    8. yylex() does not skip multiple spaces, mishandles spaces at the front of input, or requires certain spaces in order to function OK
    9. extra or bogus output not in assignment spec
    10. = instead of ==

    lex(1) and flex(1)?

    These programs generally take a lexical specification given in a .l file and create a corresponding C language lexical analyzer in a file named lex.yy.c. The lexical analyzer is then linked with the rest of your compiler.

    The C code generated by lex has the following public interface. Note the use of global variables instead of parameters, and the use of the prefix yy to distinguish scanner names from your program names. This prefix is also used in the YACC parser generator.

    FILE *yyin;	/* set this variable prior to calling yylex() */
    int yylex();	/* call this function once for each token */
    char yytext[];	/* yylex() writes the token's lexeme to an array */
                    /* note: with flex, I believe extern declarations must read
                       extern char *yytext;
                     */
    int yywrap();   /* called by lex when it hits end-of-file; see below */
    

    The .l file format consists of a mixture of lex syntax and C code fragments. The percent sign (%) is used to signify lex elements. The whole file is divided into three sections separated by %%:

       header
    %%
       body
    %%
       helper functions
    

    The header consists of C code fragments enclosed in %{ and %} as well as macro definitions consisting of a name and a regular expression denoted by that name. lex macros are invoked explicitly by enclosing the macro name in curly braces. Following are some example lex macros.

    letter		[a-zA-Z]
    digit		[0-9]
    ident		{letter}({letter}|{digit})*
    

    A friendly warning: the UNIX/Linux/MacOS Flex tool is NOT good at handling input files saved in MS-DOS/Windows format, with carriage returns before each newline character. Some browsers, copy/paste tools, and text editors might add these carriage returns without you even seeing them, and then you might end up in Flex Hell with cryptic error messages for no visible reason. Download with care, edit with precision. If you need to get rid of carriage returns there are lots of tools for that. You can even build them into your makefile. The most classic UNIX tool for that task is tr(1), the character translation utility

    Flex Header Section syntax

    Flex Body Section

    The body consists of a sequence of regular expressions for different token categories and other lexical entities.

    Actually, your semantic actions in a compiler will do more than just return the category code.

    Note: other platforms with working Flex installs (including some versions of CENTOS) do not have a flex library, neither -ll nor -lfl. Use %option directives or provide your own functions instead of expecting -lfl to be present. Example:

    %{
    #include <stdio.h>
    %}
    %option noyywrap
    %%
    "abc"	{ printf("!!!"); }
    %%
    int main()
    {
       yyin = stdin;
       return yylex();
    }
    

    Lexical Error Handling

    lecture #5 began here

    Welcome to the lecture that tests the hypothesis that Jeffery could teach entire classes just by answering (current and past) students' questions.

    Questions for (and from) the floor

    How many operators does Kotlin have?
    How many reserved words does Kotlin have?
    How many other tokens does Kotlin have?








    ...you may be shocked to hear that

    OK, now, are there questions from zoom or from the floor today?

    Mailbag

    So do I (have to) count words, or not?
    (The individual part of) Lab #1 asks you to adapt just about the simplest possible Flex specification -- a word count program -- to turn it into a mini-scanner. The word count program counts words. HW #2 asks you to finish the job of writing a (Kotlin) scanner. HW #2 does NOT need the scanner to count words any more, but it asks you to count lines and store line numbers in each token.
    What should the lexical analyzer look like? where do I start?
    In Homework #2 you learn a declarative language called Flex which does almost all the work for you. The program should read source code from a named file and print a sequence of lines of human readable text to standard output. The key bit you need to learn and understand is: how does a flex-generated scanner interact with the rest of the compiler, i.e. its public interface. This is mostly hardwired by Flex. Your only customization option is in what form to make token information available to the later phases of the compiler.
    How should our output be visible?
    One human readable output line, per token, as shown in hw2.html Build the linked list first, then walk it (visit all nodes) to print the output. Figure out how to do this so output is in the correct order and not reversed!
    You mention storing the int and double as binary. That just means storing them in int and double variables, correct?
    for constants you construct/convert from the lexeme (string that actually appears in the source code) to the lexical value (int, double, etc.) and then store the result in the corresponding lexical attribute variable.
    When do you use extern and when do you use #include in C programming?
    extern can be done without an #include, to tell one module about global variables defined in another module. But if you are going to share that extern with multiple modules, it is best to put it in an #include. More generally, use #include in order to share types, externs, function prototypes, and symbolic #define's across multiple files. That is all. No code, which is to say, no function bodies.
    Can I add parameters to yylex()?
    No, you can't add your own parameters, yylex() is a public interface. You might be tempted to add some parameters to tell it what filename it is reading from, or other such stuff. But you can't. Leave yylex()'s interface alone, the parser will call it with its current interface.
    Can I change yylex()'s return type to return a struct token *?
    No, you can't change yylex()'s return type, yylex() is a public interface. You might be tempted to return a token structure pointer. Heck, the Java version (Jflex) of flex wants yylex() to return a class instance by default. But this isn't about what you or JFlex wants. It is about what the parser wants. The parser wants an integer category, which it interprets as a Terminal Symbol in its grammar. Give the parser what it wants.
    Do you want us to have a .h file for enumerating all the different kind of tokens for HW 2? I was looking into flex and bison and it looks like bison creates a tab.h file that does this automatically.
    Yes, in HW2 you create a .h file for these #defines from the group lab exercise; plan to throw it away in favor of the one Bison creates for you in HW#3.
    Will you always call "make" on our submissions?
    Yes. I expect you to use make and provide a makefile in each homework. Turn in the whole source, not just "changed" or "new" files for some assignments. My script will unpack your .zip file by saying "unzip" in some new test directory and then run "make" and then run your executable. If anything goes wrong (say, you unzipping into a subdirectory the script does not know the name of) you will lose a few points.

    On the other hand, I do not want the tool-generated files (lex.yy.c, cgram.tab.c) or .o or executables. The makefile should contain correct dependencies to rerun flex (and later, bison) and generate these files whenever source (.l, .y , etc.) files are changed.
    When creating the linked list I see that you have a struct token and a struct tokenlist. Should I create my linked list this way or can I eliminate the struct tokenlist and add a next pointer inside struct token (i.e. struct token *next) and use that to connect my linked list?
    The organization I specified - with two separate structs - was very intentional. Next homework, we need the struct tokens that we allocate from inside yylex(), but not the struct tokenlist that you allocate from outside yylex(). You can do anything you want with the linked list structure, but the struct token must be kept more-or-less as-is, and allocated inside yylex() before it returns each time.
    I was wondering if we should have a different code for each keyword or just have a 'validkeyword' code and an 'invalidkeyword' code.
    Generally, you need a different code for two keywords if and when they are used in different positions in the syntax. For example, int and float are type names and are used in the same situations, but the keywords fun and if, denoting the beginning of a function and the beginning of a conditional expression, have different syntax rules and need different integer codes.
    In the specification for the hw it says that if there is no extension given, we should add the required extension to the filename. Should we still accept and run our compiler on other file extensions that could be provided or should we return an error of some sort?
    Accept no other extensions besides .kt or no extension. If any file has another extension, you can stop with a message like: "Usage: k0 [options] filename[.kt] ..."
    how am I supposed to import the lexer into my main.c file?
    Do not import or #include your lexer. Instead, link your lexer into the executable, and tell main() how to call it, by providing a prototype for yylex() to your main.c file. If yylex() sets any global variables (it does), you'd declare those as extern. You can do prototypes and externs directly in main.c, but these things are exactly what header (.h) files were invented for.
    Is the struct token supposed to be in our main()? Do we use yylex() along with other variables within lex.yy.c to fill the struct token with the required information?
    Rather than overwriting a global struct each time, a pointer to a struct token should be in main(). Function yylex() should allocate a struct token, fill it, and make it visible to main(), probably by assigning its address to some global pointer variable. Function main() should build the linked list in a loop, calling yylex() each time through the loop. It should then print the output by looping through the linked list.
    When I compile my homework, my executable is named a.out, not what our compiler is supposed to be named! What do I do?
    Some of you who are less familiar with Linux should read the "manual pages" for gcc, make, etc. gcc has a -o option, that would work to allow you to specify the name of your output file. Or in your makefile you could rename the file after building it.
    Can I use flex start conditions?
    Yes, if you need to, feel free. start conditions, a.k.a. modes, can be tricky though. Use if you need to.
    Can I have an extension?
    Yeah, but the further you fall behind, the more zeroes you end up with for assignments that you don't do. Late homeworks are accepted with a penalty per day (includes weekend days) except in the case of a valid excused absence. The penalty starts at 10% per day (HW#2), and reduces by 2% per assignment (8%/day for HW#3, 6%/day for HW#4, 4%/day for HW#5, and 2%/day for HW#6). I reserve the right to underpenalize.
    Do you accept and regrade resubmissions?
    Submissions are normally graded by a script in a batch. Generally, if an initial submission was a fail, I might accept a resubmission for partial credit up to a passing (D) grade. If a submission fails for a trivial reason such as a missing file, I might ask you to resubmit with a lighter penalty.
    I have not been able to figure out how sscanf() will help me. Could you point me to an example or documentation.
    Yes. Note that sscanf'ing into a double calls for a %lg.
    What is wrong with my commandline argument code.
    If it is not that you are overwriting the existing arrays with strcat() instead of allocating new larger arrays in order to hold new longer strings, then it is probably that you are using sizeof() instead of strlen().
    Can you go over using the %array vs. the standard %pointer option and if there are any potential benefits of using %array? I was curious to see if you could use YYLMAX in junction with %array to limit the size of identifiers, but there is a probably a better way.
    After yylex() returns, the actual input characters matched are available as a string named yytext, and the number of input symbols matched are in yyleng. But is yytext a char * or an array of char? Usually it doesn't matter in C, but I personally have worked on a compiler where declaring an extern for yytext in my other modules, and using the wrong one, caused a crash. Flex has both pointer and array implementations available via %array and %pointer declarations, so your compiler can use either. YYLMAX is not a Flex thing, sorry. How do you think you should limit the length of identifiers? Incidentally: I am astonished, to read claims that the Flex scanner buffer doesn't automatically increase in size as needed, and might be limited by default to 8K or so per regex match. If you write open-ended regular expressions, it might be advisable in this day of big memory to say something like
    	i=stat(filename,&st);
    	yyin=fopen(filename,"r");
    	yy_current_buffer = yy_create_buffer(yyin, st.st_size);
    
    to set flex so that it cannot experience buffer overrun. By the way, Be sure to check ALL your C library calls for error returns in this class!
    Are we free to explore non-optimal solutions?
    I do not want to read lots of extra pages of junk code, but you are free to explore alternatives and submit the most elegant solution you come up with, regardless of its optimality. Note that there are some parts of the implementation that I might mandate. For example, the symbol table is best done as a hash table. You could use some other fancy data structure that you love, but if you give me a linked list I will be disappointed. Then again, a working linked list implementation would get more points than a failed complicated implementation.
    Is it OK to allocate a token structure inside main() after yylex() returns the token?
    No. In the next phase of your compiler, you will not call yylex(), the Bison-generated parser will call yylex(). There is a way for the parser to grab your token if you've stored it in a global variable, but there is not a way for the parser to build the token structure itself. However you are expected to allocate the linked list nodes in main(), and in the next homework that linked list will be discarded. Don't get attached.
    My tokens' "text" field in my linked list are all messed up when I go back through the list at the end. What do I do?
    Remember to make a physical copy of yytext each token, because it overwrites itself each time it matches a regular expression in yylex(). Typically a physical copy of a C string is made using strdup(), which is a malloc() followed by strcpy().

    Questions about Flex HW

    Brief Comments about Lab #1 Grading

    Mailbag

    Can you explain ival/dval/sval again? What should the output look like?
    How do I handle escapes in svals? Do I need to worry about more than \n \t \\ and \r?
    Escapes entail replacing two-or-more characters with a single, encoded character when building an sval from the yytext. '\\' followed by 'n' become a control-J character. We need \n \t \\ and \" -- these are ubiquitous. You can do additional ones like \r but they are not required and will not be tested.
    You mentioned that the next homework assignment, we won't be calling yylex() from main() (which is why you previously mentioned you cannot allocate the token structure in main()). I have followed that rule, but I question how will linked lists be set up in the next homework then?
    In the next HW, the linked list will be subsumed/replaced by you building a tree data structure. If you built a linked list inside yylex(), that would be a harmless waste of time and space and could be left in place. If you malloc'ed the token structs inside yylex() but built the linked list in your main(), your linked list will just go away in the next HW when we modify main() to call the Bison parser function yyparse() instead of the loop that repeatedly calls yylex().
    Can you test my scanner and see if I get an "A"?
    No.
    Can you post tests so I can see if my scanner gets an "A"?
    If you share tests that you devise, for example when you have questions, I will add them to a public collection for use by the class. BTW, test cases are a great example of something that one might find on the internet, and use with proper citation. Compilers often have validation test suites, a public one might exist for Kotlin. I started googling things like
        "Kotlin compiler" lexical test suite  
    and maybe there is something in there. Tests of some kind. CITE anything you find and use. Did you see the Kotlin language specification section on Syntax and Grammar, which includes a "lexical grammar"?
    So if I run OK on a few sample files, do I get an "A"?
    Maybe. You should devise coverage tests to hit all described features.
    Are we required to be using a lexical analysis error function lexerr()?
    The HW Specification says we are to use at least 2 separately compiled .c files. Does Flex's generated lex.yy.c count as one of them, or are you looking for yet another .c file, aside from lex.yy.c?
    lex.yy.c counts. You may have more, but you should at least have a lex.yy.c or other lex-compatible module, and a main function in a separate .c file
    For numbers, should we care about their size? What if an integer in the source file is greater than 2^64 ?
    We could ask: what do production compilers do? Or we could just say: good catch, your compiler would ideally range check and emit an error if a value that doesn't fit into 64-bits occurs, for either the integer or (less likely) float64 literals. Any ideas on how to detect an out of range literal?

    FYI here is what the Kotlin compiler does:

    % kotlinc hello.kt
    ...well, we'll have to try it to find out
    
    I was using valgrind to test memory leaks and saw that there is a leak-check=full option. Should I be testing that as well, or just standard valgrind output with no options?
    You are welcome to use valgrind's memory-leak-finding capabilities, but you are only being graded on whether your compiler performs illegal reads or writes, including reads from uninitialized memory.
    My C compiles say "implicit declaration of function"
    The C compiler requires a prototype (or actual function definition) before it sees any calls to each function, in order to generate correct code. On 64-bit platforms, treat this warning as an error, i.e. you really should fix it.
    For built-in library symbols (functions/methods etc.) that we need to reference, what should our compiler be doing and what will we need for this homework?
    Good question. These items are processed normally in the lexical and syntax analysis phases*. For semantic analysis, we will need a strategy for pre-initializing the symbol table based on what's been included/imported. *predefined type names such as system-introduced typedef's affect syntax analysis on some common language grammars.

    Toy compiler example adapted from the Flex Manpage

               /* scanner for a toy Pascal-like language */
    
               %{
               #include <stdio.h>
               /* need this for the call to atof() below */
               #include <math.h>
               %}
    
               DIGIT    [0-9]
               ID       [a-z][a-z0-9]*
    
               %%
    
               {DIGIT}+    {
                           printf( "An integer: %s (%d)\n", yytext,
                                   atoi( yytext ) );
                           }
    
               {DIGIT}+"."{DIGIT}*        {
                           printf( "A float: %s (%g)\n", yytext,
                                   atof( yytext ) );
                           }
    
               if|then|begin|end|procedure|function        {
                           printf( "A keyword: %s\n", yytext );
                           }
    
               {ID}        { printf( "An identifier: %s\n", yytext ); }
    
               "+"|"-"|"*"|"/"   { printf( "An operator: %s\n", yytext ); }
    
               "{"[^}\n]*"}"     {  /* eat up one-line comments */ }
    
               [ \t\n]+          { /* eat up whitespace */ }
    
               .         {  printf( "Unrecognized character: %s\n", yytext ); }
    
               %%
    
               int main(int argc, char **argv )
                   {
                   ++argv, --argc;  /* skip over program name */
                   if ( argc > 0 )
                           yyin = fopen( argv[0], "r" );
                   else
                           yyin = stdin;
    
                   yylex();
                   }
    

    Using character sets (square brackets) in Flex

    A student once sent me an example regular expression for comments that read:
       COMMENT [/*][[^*/]*[*]*]]*[*/]
    
    One problem here is that square brackets are not parentheses, they do not nest, they do not support concatenation or other regular expression operators. They mean exactly: "match any one of these characters" or for ^: "match any one character that is not one of these characters". Note also that you can't use ^ as a "not" operator outside of square brackets: you can't write the expression for "stuff that isn't */" by saying (^ "*/")

    Lab2 Questions

    What are valid variable names?
    k0 will use C variable names. Underscore is a letter. A name starts with a letter and then has letters or digits.
    Is there length enforcement in variable names?
    Can we control for order of flex operations? How do we ensure that we do not tokenize lets say a part of a variable x5xx as an invalid variable 5xx? Is this just automatically done by flex?
    In a declarative language you do not ensure or control, you just use Flex's rules. Flex's rules work left to right and match the longest they can at each point that they are invoked. If they are invoked at the beginning of "x5xx" and they can match all of "x5xx" there is no way they'll match "5xx" because it will have been eaten already.
    What does # do in kotlin?
    Great question. I am not yet aware of it being used except in #!
    What does #! Do in kotlin?
    It is a kind of comment called a shebang allowed and ignored, probably only at the top of files. It allows for the notion of writing "scripts", like shell scripts. It would be trivial to support but I don't care about it so k0 doesn't have to handle it.
    Which type of error should it (HW#2?) check?
    In HW#2 we are only looking for lexical errors, I think.
    For error checking, how descriptive should it (HW#2?) be?
    Filename, linenumber, brief message with the nature of the error as you understand it

    Mailbag Questions

    Integer literal values include 0. Since integer literal contains integer value, should we consider negative sign/value in our integer-literal RE or just the positive values including 0?
    In some languages the minus sign in -123 is not part of the integer literal, it is an application of the unary negation operator. So that's two tokens, two calls to yylex() and two codes returned. And all integer literals are non-negative. However I agree that my intuition begs -123 to be a single token.
    When there is a lexical error, should the scanner terminate with an error message in stdout and stop scanning or just output an error message and keep scanning the rest of the input(s) and at the end return with 1. I think it should output errors with error count and keep reading to find out all the errors and at the end return 1 if error count is more than 0. I just want to make sure about the HW requirement.
    Thank you for your opinion. A production compiler would probably try to recover from a lexical error and continue. You are not required or expected to do so. If a lexical error occurs, you are allowed to print an error message and then terminate with the error exit code.
    since Operators and Punctuations all are one character symbols, why are we dividing in two groups? When we parse using Flex, we are going to assign different token values for each of those allowed symbols? Just to understand it better, whether in HW, should we consider a special case besides just RE match?
    In lexical analysis the distinction of operators and punctuation as two groups is unnecessary. It is more of a syntax and semantics differentiation, and even there, since they all return different integer codes, the distinct groups just inform the organization of the document.
    Characters enclosed in apostrophes. i. e. '' ' ' 'a' 'abc' are valid string literal, right?
    Apostrophes are for character literals, which must always be of length 1. '' can't be empty; the stuff inside '' might be more than one letter if it is an escape sequence or whatever. Kotlin would allow Unicode in there but k0 does not.
    Do we want apostrophes (') or double quotes (“) to represent string literal?
    String literals are always and only enclosed in double quotes.

    Finite Automata

    Efficiency in lexical analyzers based on regular expressions is all about how to best implement those wonders of CS Theory: the finite automata. Today we briefly review some highlights from theory of computation with an eye towards implementation.

    A finite automaton (FA) is an abstract, mathematical machine, also known as a finite state machine, with the following components:

    1. A set of states S
    2. A set of input symbols E (the alphabet)
    3. A transition function move(state, symbol) : new state(s)
    4. A start state S0
    5. A set of final states F

    Implementing Finite Automata

    dfa v1

    The basic logic is almost a hardwired short loop like this:
       while ((c=getchar()) != EOF) S := move(S, c);
    
    The move() function, which handles transitions from state to state, might be some kind of table (2D array) lookup. What this "finite automaton algorithm" lacks in flexibility, it makes up in speed.

    dfa v2

    In real-life, there are some side effects or semantic action code associated with certain states, or it is not very useful.
       while ((c=getchar()) != EOF) { S := move(S, c); switch(S) { ... } }
    

    To go faster than this, you can stop representing the current state in some variable S, and instead make the state an implicit property of the instruction/program counter register.

    Questions

    Do we have to do augmented operations?
    Our guiding principle is: does an intro course use it? An intro course usually uses += and -=. Let's go ahead and include those two. Not so much *=, /=, etc.

    DFAs

    The type of finite automata that is easiest to understand and simplest to implement is called a deterministic finite automaton (DFA). The word deterministic here refers to the return value of function move(state, symbol), which goes to at most one state. Example:

    S = {s0, s1, s2}
    E = {a, b, c}
    move = { (s0,a):s1; (s1,b):s2; (s2,c):s2 }
    S0 = s0
    F = {s2}
    

    Finite automata correspond in a 1:1 relationship to transition diagrams; from any transition diagram one can write down the formal automaton in terms of items #1-#5 above, and vice versa. To draw the transition diagram for a finite automaton:

    The Automaton Game

    If I give you a transition diagram of a finite automaton, you can hand-simulate the operation of that automaton on any input I give you.

    DFA Implementation

    The nice part about DFA's is that they are efficiently implemented on computers. What DFA does the following code correspond to? What is the corresponding regular expression? You can speed this code fragment up even further if you are willing to use goto's or write it in assembler.
    state := S0
    for(;;)
       switch (state) {
       case 0: 
          switch (input) {
             'a': state = 1; input = getchar(); break;
             'b': input = getchar(); break;
    	 default: printf("dfa error\n"); exit(1);
             }
          break;
       case 1: 
          switch (input) {
             EOF: printf("accept\n"); exit(0);
    	 default: printf("dfa error\n"); exit(1);
             }
          }
    
    Flex has extra complications. It accepts multiple regular expressions, runs them all in parallel in one big DFA, and adds semantics to break ties. These extra complications might be viewed as "breaking" the strict rules of DFA's, but they don't really mess up the fast DFA implementation.

    lecture #6 began here

    Deterministic Finite Automata Examples

    A lexical analyzer might associate different final states with different token categories. In this fragment, the final states are marked by "return" statements that say what category to return. What is incomplete or wrong here?

    C Comments:

    Nondeterministic Finite Automata (NFA's)

    Notational convenience motivates more flexible machines in which function move() can go to more than one state on a given input symbol, and some states can move to other states even without consuming an input symbol (ε-transitions).

    Fortunately, one can prove that for any NFA, there is an equivalent DFA. They are just a notational convenience. So, finite automata help us get from a set of regular expressions to a computer program that recognizes them efficiently.

    NFA Examples

    ε-transitions make it simpler to merge automata:

    multiple transitions on the same symbol handle common prefixes:

    factoring may optimize the number of states. Is this picture OK/correct?

    Mailbag

    My team just turned in the ytab.h that we found amongst the Kotlin ANTLR code. Was that cool?
    Yeah sure. But the purpose of the exercise was to study and learn what are the lexical patterns and meaning associated with each type of token. If you just have an integer code and you don't know what character(s) or pattern(s) in the source code that token refers to, your job wasn't finished.
    What is the intended behavior of """ ? We could accept it as a string with content of " or as an empty string with a second " token.
    When in doubt, test on a Kotlin compiler. It is the start of a multi-line string, yes?
    For character literals, if any special kind of escaped character is typed that is technically two characters (for example, \n), that should be handled automatically by flex’s character encoding right? However, \’ is not one of those characters, and we think it should be allowed within a character literal. We did what we thought was correct in the above character regular expression, but are unsure if this is the right idea.
    Flex interprets \' as a single apostrophe character, but when you write the regular expression for character literals, you have to write a regular expression that explicitly accounts for escape sequences. To get a regular expression that accepts backslash followed by apostrophe you might be writing \\\' not just \'
    Are types, as in let k: i64 = 1;, identifiers?
    k is an identifier... in a lot of programming languages like C a built-in type like i64 is not an identifier, but maybe Kotlin is different. Let's look at that question further.
    Are we doing type paramters for lists
    If k0 has "arrays" they are a class type like
    val initArray = Array<Int>(3) { 0 }
    
    Does that look about right to you? Explicit element type and size.
    If we are, is the whole thing an identifier? Or are the brackets operators?
    type templates, the < and > are separate tokens, and what was inside them would be zero or more tokens (one, in this example).
    Relating to defining Multiple-byte operators and Punctuation, what would be the values associated with them. As each other expression we use their ascii representation and for the other keywords I have them continue after 257, would we just provide them their own ascii values? In order to not have any warning shown on ‘make’ I just assigned them their own values starting at 300. If in the future there is any required value please let us know, thanks!
    Multi-byte operators and punctuation don't have a clean ASCII code to use, so yes, return a unique integer greater than 257 for them. A number greater than 300 works, since it is a number greater than 257.

    NFA examples - from regular expressions

    Can you draw an NFA corresponding to the following?
    (a|c)*b(a|c)*
    
    (a|c)*|(a|c)*b(a|c)*
    
    (a|c)*(b|ε)(a|c)*
    

    Regular expressions can be converted automatically to NFA's

    Each rule in the definition of regular expressions has a corresponding NFA; NFA's are composed using ε transitions. This is called "Thompson's construction" ). We will work examples such as (a|b)*abb in class and during lab.
    1. For ε, draw two states with a single ε transition.
    2. For any letter in the alphabet, draw two states with a single transition labeled with that letter.
    3. For regular expressions r and s, draw r | s by adding a new start state with ε transitions to the start states of r and s, and a new final state with ε transitions from each final state in r and s.
    4. For regular expressions r and s, draw rs by adding ε transitions from the final states of r to the start state of s.
    5. For regular expression r, draw r* by adding new start and final states, and ε transitions
    6. For parenthesized regular expression (r) you can use the NFA for r.

    NFA's can be converted automatically to DFA's

    In: NFA N
    Out: DFA D
    Method: Construct transition table Dtran (a.k.a. the "move function"). Each DFA state is a set of NFA states. Dtran simulates in parallel all possible moves N can make on a given string.

    Operations to keep track of sets of NFA states:

    ε_closure(s)
    set of states reachable from state s via ε
    ε_closure(T)
    set of states reachable from any state in set T via ε
    move(T,a)
    set of states to which there is an NFA transition from states in T on symbol a

    NFA to DFA Algorithm:

    Dstates := {ε_closure(start_state)}
    while T := unmarked_member(Dstates) do {
    	mark(T)
    	for each input symbol a do {
    		U := ε_closure(move(T,a))
    		if not member(Dstates, U) then
    			insert(Dstates, U)
    		Dtran[T,a] := U
    	}
    }
    

    Practice converting NFA to DFA

    OK, you've seen the algorithm, now can you use it?

    ...







    ...did you get:





    OK, how about this one:

    HW#3

    Look at HW#3 (in progress)
    What do you mean, for ival/dval/sval, by telling us to "store binary value here"
    A binary value is the actual native representation that corresponds to the string of ASCII codes that is the lexeme, for example what you get when you call atoi("1234") for the token "1234".
    I am getting a lot of "unrecognized rule" errors in my .l file
    Look for problems with regular expressions or semantic actions prior to the first reported error. If you need better diagnosis, find a way to show me your code. One student saw these errors because they omitted the required space between their regular expressions and their C semantic actions.
    Do you have any cool tips to share regarding the un-escaping of special characters?
    Copy character-by-character from the yytext into a newly allocated array. Every escape sequence of multiple characters in yytext represents a single character in sval. Inside your loop copying characters from yytext into sval, if you see a backslash in yytext, skip it and use a switch statement on the next character. See below for additional discussion.
    Is a function name also an identifier?
    Yes.

    C Pointers, malloc, and your future

    For many of you success as a computer scientist may boil down to what it will take for you to master the concept of dynamically allocated memory, and whether you are willing to do that. In C this means pointers and the malloc() family of functions. Here are some tips:

    HW Tips

    These comments are based on historical solutions. I learned a lot from my older siblings when I was young. Consider this your opportunity to learn from your forebears' mistakes.
    better solutions' lexer actions looked like
    ...regex...      { return token(TERMSYM); }
    
    where token() allocates a token structure, sets a global variable to point to it, and returns the same integer category that it is passed from yylex(), so yylex() in turn returns this value.
    Put in enough line breaks.
    Use <= 80 columns in your code, so that it prints readably.
    Comment non-trivial helper functions. Comment non-trivial code.
    Comment appropriate for a CS professional reader, not a newbie tutorial. I know what i++ does, you do not have to tell me.
    Do not leave in commented-out debugging code or whatever.
    I might miss, and misgrade, your good output if I can't see it.
    Fancier formatting might calculate field widths from actual data and use a variable to specify field widths in the printf.
    You don't have to do this, but if you want to it is not that hard.
    Remind yourself of the difference between NULL and '\0' and 0
    NULL is used for pointers. The NUL byte '\0' terminates strings. 0 is a different size from NULL on many 64-bit compilers. Beware.
    Avoid O(n2) or worse, if at all possible
    It is possible to write bad algorithms that work, but it is better to write good algorithms that work.
    Avoid big quantities of duplicate code
    You will have to use and possibly extend this code all semester.
    Use a switch when appropriate instead of long chain of if-statements
    Long chains of if statements are actually slow and less readable.
    On strings, allocate one byte extra for NUL.
    This common problem causes valgrind trouble, memory violations etc.
    On all pointers, don't allocate and then just point the pointer someplace else
    This common student error results in, at least, a memory leak.
    Don't allocate the same thing over and over unless copies may need to be modified.
    This is often a performance problem.
    Check all allocations and fopen() calls for NULL return (good to have helper functions).
    C library functions can fail. Expect and check for that.
    Beware losing the base pointer that you allocated.
    You can only free() if you still know where the start of what you allocated was.
    Avoid duplicate calls to strlen()
    especially in a loop! (Its O(n2))
    Use strcpy() instead of strncpy()
    unless you are really copying only part of a string, or copying a string into a limited-length buffer.
    You can't malloc() in a global initializer
    malloc() is a runtime allocation from a memory region that does not exist at compile or link time. Globals can be initialized, but not to point at memory regions that do not exist until runtime.
    Don't use raw constants like 260
    use symbolic names, like LEFTPARENTHESIS or LP
    The vertical bar (|) means nothing inside square brackets!
    Square brackets are an implicit shortcut for a whole lot of ORs
    If you don't allocate your token inside yylex() actions...
    You'll have to go back and do it, you need it for HW#2.
    If your regex's were broken
    If you know it, and were lazy, then fix it. If you don't know it, then good luck on the midterm and/or final, you need to learn these, and devise some (hard) tests!

    lecture #7 began here

    Some questions

    What are the operators and keywords that we definitely need to support?
    What operators and keywords would an intro course use? +-*/= if/while/...
    Are we going to compile for x86_64 assembly or bytecode?
    Still thinking about it.
    What are some examples of code that our compiler should be able to handle?
    We should establish a body of K0-acceptable Kotlin. I may produce such a body, or may assign it as a lab exercise.
    Should our compiler be able to import and use Java libraries like regular Kotlin can?
    Depending on our reference kotlin base, we may have a limited number of Java imports with a limited number of methods we decide to support.
    Do we need to worry about null safety?
    It is pretty hard to be Kotlin without doing something about null safety. Possibilities include: allowing ? syntax but not enforcing anything, enforcing it in the type system (compile time type checks) but no runtime enforcement, or full enforcement. We have about two HW's to decide.
    Do we need to worry about optimizing the resulting code, and if so, what level of optimization is expected?
    Unfortunately, optimization is generally the subject of more advanced compiler courses. In this first compiler course, we will introduce some concepts from optimization but will not try to do much in that area.
    Do we need to have helpful error messages and warnings if the compiler encounters something strange during compilation?
    You need to have error messages that are not actively incorrect and/or misleading, if possible.

    On resizing arrays in C

    The sval attribute in the homework is a perfect example of a problem which a Business (MIS) major might not be expected to solve well, but a CS major should be able to do by the time they graduate. This is not to encourage any of you to consider MIS, but rather, to encourage you to learn how to solve problems like these.

    The problem can be summarized as: step through yytext, copying each character out to sval, looking for escape sequences.

    Space allocated with malloc() can be increased in size by realloc(). realloc() is awesome. But, it COPIES and MOVES the old chunk of space you had to the new, resized chunk of space, and frees the old space, so you had better not have any other pointers pointing at that space if you realloc(), and you have to update your pointer to point at the new location realloc() returns.

    There is one more problem: how do we allocate memory for sval, and how big should it be?

    Syntax Analysis

    Parsing is the act of performing syntax analysis to verify an input program's compliance with the source language. A by-product of this process is typically a tree that represents the structure of the program.

    Context Free Grammars

    A context free grammar G has:

    Example CFG

    Gonna use Bison syntax, not CSE 342 syntax. Arrows become colons. OR bars denote additional production rules. Semi-colons are superfluous.
      nt1 : nt2 nt3 nt4 ;
      nt2 : OPEN ;
      nt3 : nt3 X | nt3 ;
      nt4 : CLOSE ;
    

    Reading Assignment

    Amending the reading assignment from last lecture:

    Derivation

    A context free grammar can be used to generate strings in the corresponding language as follows:
    let X = the start symbol s
    while there is some nonterminal Y in X do
       apply any one production rule using Y, e.g. Y -> ω
    
    When X consists only of terminal symbols, it is a string of the language denoted by the grammar. Each iteration of the loop is a derivation step. If an iteration has several nonterminals to choose from at some point, the rules of derviation would allow any of these to be applied. In practice, parsing algorithms tend to always choose the leftmost nonterminal, or the rightmost nonterminal, resulting in strings that are leftmost derivations or rightmost derivations.

    Context Free Grammar Example (from BASIC)

    How many terminals and non-terminals does the grammar below use? Compared to the little grammar we started last time, how does this rate? What parts make sense, and what parts seem bogus?
    Program : Lines
    Lines   : Lines Line
    Lines   : Line
    Line    : INTEGER StatementList
    StatementList : Statement COLON StatementList
    StatementList : Statement
    Statement: AssignmentStatement
    Statement: IfStatement
     REMark: ... BASIC has many other statement types 
    
    AssignmentStatement : Variable ASSIGN Expression
    Variable : IDENTIFIER
     REMark: ... BASIC has at least one more Variable type: arrays 
    
    IfStatement: IF BooleanExpression THEN Statement
    IfStatement: IF BooleanExpression THEN Statement ELSE Statement
    
    Expression: Expression PLUS Term
    Expression: Term
    Term      : Term TIMES Factor
    Term      : Factor
    Factor    : IDENTIFIER
    Factor    : LEFTPAREN Expression RIGHTPAREN
     REMark: ... BASIC has more expressions 
    

    Kotlin Context Free Grammar Example

    Well, OK, so how much of the Kotlin language grammar can we come up with in class today? You can work on context free grammars "top down" or "bottom up", but perhaps bottom up is most tangible. Start with expressions, work on up to statements, and work there up to entire functions, and programs.
    insert your CFG ideas here 
    ...
    
    lecture #8 began here

    HW#2 Status

    Grammar Ambiguity

    The grammar
    E -> E + E
    E -> E * E
    E -> ( E )
    E -> ident
    
    allows two different derivations for strings such as "x + y * z".

       vs.   

    The grammar is ambiguous, but the semantics of the language dictate a particular operator precedence that should be used. One way to eliminate such ambiguity is to rewrite the grammar. For example, we can force the precedence we want by adding some nonterminals and production rules.

    E -> E + T
    E -> T
    T -> T * F
    T -> F
    F -> ( E )
    F -> ident
    
    Given the arithmetic expression grammar from last lecture:

    How can a program figure that x + y * z is legal?
    How can a program figure out that x + y (* z) is illegal?

    YACC (and Bison, etc.)

    YACC ("yet another compiler compiler") is a popular tool which originated at AT&T Bell Labs. YACC takes a context free grammar as input, and generates a parser as output. Several independent, compatible implementations (AT&T yacc, Berkeley yacc, GNU Bison) for C exist, as well as many implementations for other popular languages. There also exist other more "modern" parser generators, but they are often less portable and are heavily inspired/influenced by YACC so it is what we will study.

    YACC files end in .y and take the form

    declarations
    %%
    grammar
    %%
    subroutines
    
    The declarations section defines the terminal symbols (tokens) and nonterminal symbols. The most useful declarations are:
    %token a
    declares terminal symbol a; YACC can generate a set of #define's that map these symbols onto integers, in a y.tab.h file. Note: don't #include your y.tab.h file from your grammar .y file, YACC generates the same definitions and declarations directly in the .c file, and including the .tab.h file will cause duplication errors.
    %start A
    specifies the start symbol for the grammar (defaults to nonterminal on left side of the first production rule).

    The grammar gives the production rules, interspersed with program code fragments called semantic actions that let the programmer do what's desired when the grammar productions are reduced. They follow the syntax

    A : body ;
    
    Where body is a sequence of 0 or more terminals, nonterminals, or semantic actions (code, in curly braces) separated by spaces. As a notational convenience, multiple production rules may be grouped together using the vertical bar (|).

    Bottom Up Parsing (How Does Bison's yyparse() Work?)

    Bottom up parsers start from the sequence of terminal symbols and work their way back up to the start symbol by repeatedly replacing grammar rules' right hand sides by the corresponding non-terminal. This is the reverse of the derivation process, and is called "reduction".

    Example. For the grammar

    (1)	S->aABe
    (2)	A->Abc
    (3)	A->b
    (4)	B->d
    
    the string "abbcde" can be parsed bottom-up by the following reduction steps:
    abbcde
    aAbcde
    aAde
    aABe
    S
    

    Handles

    Definition: a handle is a substring that
    1. matches a right hand side of a production rule in the grammar and
    2. whose reduction to the nonterminal on the left hand side of that grammar rule is a step along the reverse of a rightmost derivation.

    Comments on HW#2

    I ran some tests on everyone's HW#2 submission. If you didn't "make" you got a fix-and-resubmit instruction on Canvas
    Testing is for basic functionality
    Testing will get more elaborate in later HW's
    The executable is to be named k0 and not anything else
    In a Makefile, CFLAGS is for -c compile steps, LDFLAGS is for link steps
    Typical to define CFLAGS in this class to be -c -g -Wall.
    Do not mix compile lines with link lines.
    The link line should link together .o files. The larger the program, the more important it is to not recompile all files whenever one changes.

    Mailbag

    Thank you to all of you who are sending me juicy questions by e-mail.
    I am trying to get a better picture of the communication that happens between Flex and Bison. From my understanding:
    1. main() calls yyparse()
    2. yyparse() calls yylex()
    3. yylex() returns tokens from the input as integer values which are enumerated to text for readability to the .y file
    4. yyparse() tries to match these integers against rules of our grammar.
      (if it is unsuccessful, it errors (shift/reduce, reduce/reduce, unable to parse))
    Is this correct?
    Does yylval have any use to us? I read that it is used to bring the value of the token into the parser.
    Yes, yylval is how yyparse() picks up lexical attributes. We will talk about it more in class.
    How would we add artificial tokens, such as to insert a token, perhaps an INDENT token, , without skipping real tokens when we return?
    Save the real, found token in a global or static, or figure out a way to push it back onto the input stream. Easiest is a one-token saved (pointer to a) token struct. One could build a whole stack of saved tokens if one wished. In a previous Lab #2 and its aftermath, we worked out how to generate the sequence of integer codes that were needed. This question gets more interesting once the token structures in (formerly yyoken and now) yylval must be considered.
    insertion?
    Two possibilities come to mind:
    1. Modify output of bison to replace the call to yylex() with myyylex(),
    2. Modify output of flex to change its declaration of yylex to realyylex()
    A traditional Linux tool for sneaky stuff like this is sed(1), which could be invoked from your makefile. -->
    Does our language support empty statements?
    Most mainstream languages allow these. Do they get used in a CSE 107 or 113 class? Unlikely. Occasionally I find them handy IRL. I would say they are optional in k0.
    I previously created one token to handle all cases of unsupported reserved words. Is there any reason I should keep the individual tokens for the unsupported reserved words in the .y file? Or can I just use my catch all token. Same goes for unsupported operators/punctuations.
    You can use your catch-all token if your lexical analyzer stops with a good error message (filename, linenumber, what's wrong). You don't have to have grammar rules for the things not our language, if those things would be caught in a fatal lexer error. This is why I am not making you "recover" from lexical errors; so your grammar can become smaller and you'll have less tree constructing to do in HW#3.
    How would you recommend "de-escaping" strings?
    I recommend the following algorithm:
    outputstring gets the empty string
    for every character inside the double quotes
       if the character is backslash then
          do a big switch statement on the next character
             (backslash followed by t, for example, means outputstring gets a '\t' appended to it
       else the character was not a backslash, so just append it to the outputstring
    It can look humorous, and very meta, to be switching on case 't' outputting a '\t', on case 'n' outputting a '\n' etc.
    Part 2, Step 6 of Lab 3 says: "Modify your HW2 main() function to call yyparse() one time in place of the while loop that called yylex() over and over again." In HW2, we used each call to yylex() to fill each node of our linked list. For Lab 3, are we supposed to keep the linked list?
    You don't need to build a linked list in Lab 3. If you did, you'd have to build it entirely inside yylex(), but instead for Lab 3 you are probably just ignoring the tokens. In Lab 4 we are going to be building tree nodes and putting the struct token * allocated in yylex() into the tree instead of into a linked list.
    I read we could use '{' and such as terminal symbols in the grammar, instead of names like LCURLY. What are the pros and cons. I know in C, '{' is just a small integer.
    I used to dislike grammars using character constants for their one-letter terminal symbols, instead of names. I found the mixture of character literals and name to be jarring, and found consistency to be comforting in my newbieness. But Bison starts named terminals above 256 for a reason. When a terminal is only one character long, using its ASCII code is the shortest, most human readable way to identify it. So I have taken to preferring that.
    What about no-op statements like 2+2?
    It is OK to not allow no-op statements like 2+2. You can impose a requirement to use the value somehow, like by writing it out or assigning it to a variable.
    The full language grammar has many symbols that have not been mentioned that we might not want in our subset language. Should we support them, or not? If we don't have to support them does that mean we don't have to include all the grammar rules that use them?
    Feel free to ask about specifics. The published grammar is for the whole language not our subset. You would need to delete from it en masse to get down to our subset. While that might be helpful from a code-management perspective, it would also leave you saying a more vague "parse error" message for many legal constructs for which a more helpful message is "this feature is not supported by XXX", where XXX is our language.
    How about a tool that would generate #define numbers for grammar rules automatically from our .y files?
    I wrote a cheap hack version 0 of such a tool awhile back. I have not tested it on our .y, it might be buggy.
    Will we get marked off for any reduce/reduce conflicts we have?
    Yes you will lose points if you turn in a HW#2 with reduce/reduce conflicts.
    How are we supposed to integrate the token names we created in the lexer with those token names in the Bison .y file?
    Any which way you can. Probably, you either rename yours to use their names, or rename theirs to use your names.
    what action should be taken in the case of epsilon statements
    HW#2 spec says to use $$=NULL. I could also imagine using $$=alctree(EPSILON, 0) to build an explicit epsilon leaf, for people who don't like to have to check for NULL everywhere.
    Will I be setting myself up for failure if I attempt to write my own grammar from scratch?
    Go right ahead and ignore the provided grammar if you want; feel free to instead derive your grammar from the reference manual.

    Shift Reduce Parsing

    A shift-reduce parser performs its parsing using the following structure
    Stack					Input
    $						ω$
    
    At each step, the parser performs one of the following actions.
    1. Shift one symbol from the input onto the parse stack
    2. Reduce one handle on the top of the parse stack. The symbols from the right hand side of a grammar rule are popped off the stack, and the nonterminal symbol is pushed on the stack in their place.
    3. Accept is the operation performed when the start symbol is alone on the parse stack and the input is empty.
    4. Error actions occur when no successful parse is possible.

    k0 clarification on ->

    Q: Does k0 have the token -> ??
    As far as I know, it does

    Type Names

    "Type names" include typedef names (C/C++), enumeration constants (C/C++/Java), and in OO languages, class names too.

    Type names are an example where in some languages if we are not careful, all the beautiful CFG parsing theory we've used up to this point breaks down and flex/bison compilers tend to cheat a little.

    typedef int foo;
    foo x;                    /* a normal use of typedef... */
    foo foo;                  /* try this on gcc! is it a legal global? */
    void main() { foo foo; }  /* what about this ? */
    
    On the other hand, some languages' grammars (yay Golang! Kotlin too, I think) avoid this problem entirely.

    What CSE 423 Should do About Type Names This Semester

    Although in the classic C language grammar type names had to return a different integer code than identifiers, I am not aware of any such issue in the Kotlin grammar. We will do nothing special (that is to say, names of types are identifiers) unless our careful reading of the Kotlin Grammar determines that something is necessary.

    The YACC Value Stack

    YACC Value Stack's Element Type: YYSTYPE

    Midterm Exam Date Discussion

    It will be an in-class (or in Student Access Services) midterm on Monday March 11. There will be an exam review Friday March 8.

    HW#2 Grading Status

    Old Mailbag

    During my "make" the linker complains about redefinition of yyparse() and missing main(). What's going on?
    If your main() function is in k0.c, you had better not name your grammar file k0.y, or be careful if you do -- on at least some platforms (probably this refers to Linux and GNU make) the "make" program has default rules that assume if a .c file has the same name as a .y file, it is supposed to build the .c from the .y by running yacc and renaming the foo.tab.c as foo.c!
    What is %prec about? What about %left and %right?
    Great question. %prec TERM directs Bison to apply the current grammar rule with the precedence of TERM. In a .y file, fake terminal symbols (like THEN) can be introduced to avoid shift/reduce conflicts. Note that neither %prec nor TERM are symbols on the righthand side of whatever production rule is being given -- if there aren't any other symbols then that precedence is being applied to an epsilon rule. %prec is used to apply some precedence rules specified via %left, %right etc. to production rules in the Bison grammar where there is not a symbol that can be declared as %left or %right. %left and %right are in turn, Bison's way of tie-breaking ambiguous grammar rules that would otherwise generate shift/reduce or reduce/reduce conflicts.
    I am working on the tree but I am confused as to how to approach it. For example package has the following rule:
    package: LPACKAGE sym ';'
    The tree struct shown on the HW assignment sheet has kids which are of type struct tree and a leaf which is of struct token. Since package has two tokens the LPACKAGE and ';' how should I approach saving this to the tree struct. Should everything be saved under kids? With how I have my %union right now, LPACKAGE and ';' are tokens and sym is struct tree.
    The example tree type in the HW spec, which you are not required to follow, illustrates one possible way to incorporate terminal symbols as leaves. If you follow it, separate from your struct token for each leaf you allocate a struct tree, with 0 children, whose prodrule is the token's terminal symbol #, and for a treenode with 0 children and a terminal symbol as a prodrule, the code that goes back through the tree afterwards would know to not visit the children array, but instead look at the leaf field for a token. To do all this with your current %union with pointer to struct token on the tree for terminal symbols, every time you are about to insert a tree node with terminal symbols, you would allocate a leaf node to hold the token *. So your rule for a package would allocate three tree nodes total, one for the parent and two for the two terminal symbols being placed into leaves. There are other ways that one can get it done, but this would work.

    lecture #9 began here

    Getting Lex and Yacc to talk

    Options:

    1. Declare that struct token may appear in the %union. In that case the value stack is a mixture of struct node and struct token. You still have to have a mechanism for how do tokens get wired into your tree. Are all children of type union YYSTYPE, and you use the prodrule R to tell which are which?
    2. For each terminal symbol, allocate a "leaf" tree node with 0 children and point its "leaf" field at your struct token. 0 children implies "don't use the kids field" and "a non-null leaf might be present"
    3. declare a tree type that allows tokens to include their lexical information directly in the tree nodes, perhaps tree nodes contain a union that provides EITHER an array of kids OR a struct token.
    If you have more than one %union type possible, be prepared to see type conflicts and to declare the types of all your nonterminals.

    Getting all this straight takes some time; you can plan on it. Your best bet is to draw pictures of how you want the trees to look, and then make the code match the pictures. Given pictures, I can help you make the code do what the pictures say. No pictures == "Dr. J will ask to see your pictures and not be able to help if you can't describe your trees."

    Declaring value stack types for terminal and nonterminal symbols

    Unless you are going to use the default (integer) value stack, you will have to declare the types of the elements on the value stack. Actually, you do this by declaring which union member is to be used for each terminal and nonterminal in the grammar.

    Example: in the cocogram.y that I gave you we could add a %union declaration with a union member named treenode:

    %union {
      nodeptr treenode;
    }
    
    This will produce a compile error if you haven't declared a nodeptr type using a typedef, but that is another story. To declare that a nonterminal uses this union member, write something like:
    %type < treenode > function_definition
    
    Terminal symbols use %token to perform the corresponding declaration. If you had a second %union member (say struct token *tokenptr) you might write:
    %token < tokenptr > SEMICOL
    

    Examples

    Chapter 5 of Building Your Own PL

    How to interpret your HW#2 Grade

    Mailbag

    Does Flex have "matching groups"?
    No.
    Well, Does Flex have a good way to match Kotlin strings literals like
          br##"foo #"# bar"##
    
    LOL, what do your CSE 342 instincts tell you about having to match an equal number of # signs before and after the double quotes?









    My compiler is complaining about strdup being missing. What up?
    It turns out -std=c99 removes strdup() because it is not part of that standard. Possibly solutions include: not using -std=c99 when compiling files that call strdup(), or writing/providing your own strdup().
    When I compile my .y file, bison complains spitting out a bunch of warnings about useless nonterminals and rules. how much attention should I pay to this?
    "Useless" warnings sound innocuous, but they mean what they say. You probably have something wrong that will have to be fixed. Everything but shift/reduce conflicts is potentially serious, until you determine otherwise. If you can't figure out what some Bison error is after giving it the reasonable college try, send it to me by e-mail or schedule an appointment. If I am not available or you are remote, we may schedule a Zoom meeting. You might have to learn some Zoom. I might have to setup a camera on my many machines, and remember my Zoom credentials.
    It seems we soon will have to implement a hash table. If this is the case, what would be a reasonable size (# buckets) of the table? n=20?
    Fixed-size tables should use a prime. For the size of inputs I will ever use in this class, probably a prime less than 100 would do. How about n=41 ?
    My syntax checker seems to be having some issues on control words such as for, while, etc.. Interestingly enough if I replace the return type of these statements in my .l file with an integer return type then the error goes away.
    The return type of yylex() definitely must be int and the integers returned definitely must be those yyparse() #define's in a .tab.h file generated by running "bison -d" on the .y file. To check if lexer and parser are talking OK, turn on global variable yydebug=1 in main() before calling yyparse() (and turn on YYDEBUG in the .y file if its not on already).
    My HW didn't get all the points. Can I resubmit with fixes?
    In many cases you need the fixes for use in your compiler the rest of the semester.
    Generally, if an initial submission was fail (for example, below 12/20), I accept a resubmission for partial credit up to a passing grade. If a submission fails for a trivial reason such as a missing file, I might ask you to resubmit with a lighter penalty.
    As a reminder, I do Not grade on a 90/80/70/60 scale.
    In the example A : b C d {$$ = tree(R,3,$1,$2,$3);} ; Suppose C is a terminal and b and d are non-terminals. Then $2 will be OK, but when will I be able to get the data that $1 and $3 need to be set to?
    Bison parsers are bottom up. You don't reduce this grammar rule or execute this code until sometime after the handle b C d has already been parsed, and in that process the production rules for b and d have already executed, just as surely as the shift of C which placed whatever yylval held at that time onto the value stack in $2. If the rules for b and d had actions that said {$$=...} at that point in the past, then $1 and $3 now will be holding what was assigned to $$ back in those rules' semantic actions.
    In the example A : b C d {$$ = tree(R,3,$1,$2,$3);} ; to what doth R refer?
    R was intended to be an integer code that allows you to tell, when walking through the tree later on, what production rule built that node. I would typically do a code for each nonterminal, gapped large enough that R can be (nonterminal+rule#forthatnonterminal). Suppose this was the first of three production rules that build an A. The integer might be (__A__ + 1) to denote the first A rule.
    I'm considering having some sort of stack that keeps track of the parent you should currently be attaching children to.
    You can do anything you want, but bison's value stack is that stack and and at each level you should allocate a node for $$ and attach all of its children. That is it.
    Should we be defining our own integer codes for token types or just use the ones in our *.tab.h file from HW#1?
    You can't define your own codes, you have to use the codes that bison generates. You'll have to modify your lexer to use bison's integers, or your flex and bison will not work together.
    Are yylval's types defined in the %union?
    Yes, yylval is of type YYSTYPE, the type bison generates for the %union.
    What is the actual value of a $n variable?
    Before your bison grammar's semantic action code triggers, $1, $2, ... etc. will be holding either (A) whatever you put in yylval if the corresponding symbol is a terminal, or (B) whatever you put in $$ for its rule if the symbol is a non-terminal.
    What is required of a for-statement?
    You don't have to do a declaration of a variable in a for initializer. Saying for(int i=1; ...) was a C++ thing.
    I have spent 30 hours and am not close to finishing adding the hundreds of grammar rules and tree construction semantic actions required for HW#2!
    As a professional programmer, you should invest time to master a powerful programmer's editor with a key-memorizing macro facility that can let you insert semantic actions (for example) very rapidly. If you've been typing them in by hand, ouch! Paste the right thing 500 times and then just tweak. Or paste all the constructors with the same number of children in batches, so you have less to tweak because you already pasted in the right number of kids.
    Do I need to add %type <treeptr> nonterm for every non-terminal symbol in the grammar in order to have everything work
    yes. If you are lucky, the %type's are in there in a comment, and all you have to do is uncomment them and add the <treeptr> (or whatever) part.
    Are we supporting (syntactically) nested classes/structs?
    no
    What do I do with epsilon rules? Empty tree nodes?
    I previously said to use either $$ = NULL or $$ = alctree(RULE, 0). Whether the latter is preferable depends on what will make your tree traversals easier later on, in HW#3, and maybe whether the encoding of an empty leaf with RULE would help you in reading the tree and knowing what to do with it. Saying $$=NULL implies you will have to check children to see if they are NULL before you try to visit them. Never setting $$ to NULL means you can "blindly" traverse child pointers if a parent has nkids > 0.
    All of the leaves in the tree structure are/can be made of lex tokens. To that point then, what are the non-leaves supposed to be? I think I may have over thought this point so I am not quite sure.
    Non-leaves (i.e. internal nodes) correspond to non-terminal symbols, built from particular production rules.
    For the structure of the tree, HW#2 provides a possible "setup".
    struct tree {
       int prodrule;
       int nkids;
       struct tree *kids[9];
       struct token *leaf;
    }
    
    While I understand nkids (number of kids this node has), *kids[9] (a pointer array to up to 9 kids), and leaf (the lex token), what exactly is the prodrule? I am fairly certain that this is the production rule, but I am not exactly sure what it associates with.
    The prodrule integer encodes what production rule was used to build this node, which includes (of course) what non-terminal it represents, and what syntactic role its children played. By the way, *kids[9] is an array of 9 pointers to kids, not a pointer to an array of nine kids.
    What exactly is in $1 or $2 or ... when I am at a reduction building a tree node in $$ for some non-terminal?
    I was wondering if it is ok to have a linked list of syntax trees, where the syntax tree for the current source file be inserted into a linked list (of syntax trees), then at the end of main(), after generating syntax trees for each file in command line argument, walk through the linked list and print out each syntax tree.
    What is expected is that for each file, you build the tree, return to main(), print it out, and then move on to the next filename. But building a linked list of trees and looping again over that to print things out would be fine. The main thing between each file on the command line is to clear out the type name table; each file is being compiled independently of whatever came before them during that compilation process.

    Conflicts in Shift-Reduce Parsing

    "Conflicts" occur when an ambiguity in the grammar creates a situation where the parser does not know which step to perform at a given point during parsing. There are two kinds of conflicts that occur.
    shift-reduce
    a shift reduce conflict occurs when the grammar indicates that different successful parses might occur with either a shift or a reduce at a given point during parsing. The vast majority of situations where this conflict occurs can be correctly resolved by shifting.
    reduce-reduce
    a reduce reduce conflict occurs when the parser has two or more handles at the same time on the top of the stack. Whatever choice the parser makes is just as likely to be wrong as not. In this case it is usually best to rewrite the grammar to eliminate the conflict, possibly by factoring.
    Example shift reduce conflict:
    S->if E then S
    S->if E then S else S
    

    In many languages two nested "if" statements produce a situation where an "else" clause could legally belong to either "if". The usual rule (to shift) attaches the else to the nearest (i.e. inner) if statement.

    Mailbag

    I don't get how to start tree traversal.
    yyparse() should build a syntax tree, but it returns an integer that indicates whether there were parse errors. You either traverse your tree inside yyparse() before it returns (a bit weird) or you traverse your tree outside yyparse() after it returns...by assigning the tree to a global variable.
    I have mysterious syntax errors, what do I do?
    I can't fix some of the shift/reduce conflicts, what do I do?
    Nothing. You do not have to fix shift/reduce conflicts.
    I can't fix some of the reduce/reduce conflicts, what do I do?
    These generally reflect a real bug and will cost you a few points on HW, but they mighty or might not cost you more points on test cases. It is only a deal breaker and has to be fixed if it prevents us from parsing correctly and building our tree. Sometimes epsilon rules can be removed successfully by adding grammar rules in a parent non-terminal that omit an epsilon-deriving child, and then modifying the child to not derive epsilon. This might or might not help reduce your number of reduce/reduce conflicts.
    With the default error handling, I am getting an error on the last line of the file: syntax error before '' token. It looks like an EOF error, but I cannot figure out how to fix it, as when I add an <<EOF>> rule to my lexer, it just hangs there, and still produces this error.
    Error on EOF might be because the grammar expects one more semi-colon, maybe your EOF regex should return one the first time it hits in each file. By the way, I usually don't have to write a <<EOF>> regex, yylex() returns the value on EOF that yyparse() expects. If you enable YYDEBUG and turn on yydebug you will get a detailed explanation of the parse and where it is failing when you run your parser, which may help you. Feel free to schedule a Zoom session.
    Example reduce reduce conflict:

    (1)	S -> id LP plist RP
    (2)	S -> E GETS E
    (3)	plist -> plist, p
    (4)	plist -> p
    (5)	p -> id
    (6)	E -> id LP elist RP
    (7)	E -> id
    (8)	elist -> elist, E
    (9)	elist -> E
    
    By the point the stack holds ...id LP id
    the parser will not know which rule to use to reduce the id: (5) or (7).

    Another Example Reduce Reduce Conflict

    The following grammar, based loosely on an expression grammar, illustrates a reduce reduce conflict, and how you have to exercise care when using epsilon productions. Epsilon productions were helpful for some of the grammar rewriting methods, such as removing left recursion, but used indiscriminately, they can cause much trouble.
    T : F | F T2 ;
    T2 : p F T2 | ;
    F : l T r | v ;
    
    The reduce-reduce conflict occurs after you have seen an F. If the next symbol is a p there is no question of what to do, but if the next symbol is the end of file, do you reduce by rule #1 or #4 ?

    lecture #10 began here

    Mailbag

    What up with the group assignments
    Everything is negotiable, but by default, these groups will persist through the rest of the semester. Do your best to contribute to your team. Each person individually turn in a ranking of 0..2 for each team member each lab and each homework from now on. One person turn in the lab or homework deliverables.
    Do we need to catch/handle the pound character #? Should it be an error?
    Any character not part of a legal token is a lexical error. Maybe # is used in shebang lines, but if it appears by itself not in a shebang regex, it is an error.
    Are we expected to account for composition and precedence for literals at this stage of lexical analysis?
    No, feel free to bring up specific examples, but the K0 subset of Kotlin is allowed to be simpler including not doing composition of literals other than things like println(" x is $x"). These ya-basics will be implemented via generated code.
    Chains of (often optional) NL tokens are causing me grammar hell! I have like hundreds of reduce/reduce conflicts! Can I drop them and require semi-colons like C does?
    The preferred options are to write a grammar that doesn't need semi-colons and works anyhow, or write a grammar that needs semi-colons, and perform "semi-colon insertion" (see below). If you require semi-colons where Kotlin would not, it will cost you a few points, around 10% on HW#3.

    A Short Primer on Semi-colon Insertion

    Semi-colons and similar statement separators (Pascal family) or statement terminators (C family) were in wide use by 1970 and plague us to this day. "Missing semi-colon" is one of the most frequent syntax errors worldwide resulting in countless recompiles.

    Without a marker token, it seems that it is hard to tell the end of one statement and the beginning of the next. And it seems that when newlines are used for this purpose, that creates its own problems, especially when a complex statement needs to span across multiple lines.

    Semi-colon insertion was invented by 1980 and "perfected" by 1983 but new languages seem to keep inventing alternative ways to do it. Kotlin seems to use newlines as explicit tokens, but getting that right seems to have eluded multiple CSE 4023 students in Spring 2025, and having explicit tokens for whitespace is generally to be avoided for exactly the reasons some of you are experiencing AND because it overcomplicates the grammar. One student suggested dropping the newline tokens and requiring semi-colons everywhere, and I can live with that, but if we break every Kotlin program for K0 that is a bummer*. So here are a couple ways to get rid of newline tokens by inserting semicolons, an attempt to have your cake and eat it too.

    The Go Way

    According to Michal Lowicki (non authoritative), Go inserts semi-colons at the point of a newline so long as the previous token was: an identifier, a constant, a keyword in
    break continue fallthrough return
    
    or one of
    ++ -- ) ] }
    
    The above rule is pretty simple to implement since it requires only one token of memory (look behind) to decide whether to insert a semi-colon upon seeing a newline. It would be enough for full credit to just do the Kotlin equivalent of this (keywords tweaked appropriately). Go has a special 2nd rule, which is the semi-colon can be omitted before a closing ) or } token. This would take one token of lookahead, slightly harder.

    Icon-style

    The Icon programming language semi-colon insertion rules are characterized as follows. They are slightly more sophisticated than the Go rules. Every token type is classified as to whether it is a "Beginner" and whether it is an "Ender" (two bit flags for each token type). One token of lookahead is recalled, so there is a PreviousToken (at least its terminal symbol category) each time yylex() is asked to read the next token. A global flag (call it NewlineSeen)is maintained that remembers whether a newline has occurred since the last token. Whenever the next token is requested,
        if (NewlineSeen && IsBeginner(CurrentToken) && IsEnder(PreviousToken)) {
           save the CurrentToken for next time
           return a Semi-colon token
        }
    
    With a little more dancing, you can rename flex's yylex() to be flexyylex() and write a shim function yylex() in between Flex's and Bison's parser:
      yylex(){
        if (SavedToken != NULL) {
          rv = SavedToken
          SavedToken = NULL
          yylval = rv
          return rv->category
          }
    
        CurrentToken = flexyylex()
    
        if (NewlineSeen && IsBeginner(CurrentToken) && IsEnder(PreviousToken)) {
           SavedToken = yylval
           construct and return a Semi-colon token
        }
    
        return CurrentToken->category
      }
    

    Optionals that aren't optional in K0: Simplify your grammar if need be

    In an ideal world, we would parse all of Kotlin and then reject everything that don't meet K0's restricted semantics. That way we could give good error messages for everything.

    In practice, I am a pragmatist, and if hacking out a lot of Kotlin that's not in K0 might make the difference between finishing HW3 or not, hack away. I will test it on fairly simple common Kotlin stuff. Here's a long list of stuff not in K0:

    On Debugging Bison Grammars

    There's debugging the C code generated by Bison when it runs, and debugging the actual grammar from which that C code is generated. The lecture notes probably already have comments on both, but here is a bit on debugging the actual grammar, to get rid of reduce/reduce conflicts for example.
    use bison -v
    learn to read its output file. Find the (pushdown automata) state where a reduce/reduce conflict is occurring. Look at what production rule(s) have reduces in that state. Rewrite one or more of those production rule(s)
    get rid of epsilon productions whenever possible.
    For an "optional" item it is often possible to go up one level in the grammar to where the optional item is used in a right hand side, and replace the production rule with two copies, one with the optional item present and non-optional, and another copy of the production rule with the optional item absent.
    add production rules incrementally
    Sometimes it is pretty hard to debug from the .output file. If you try to debug an entire language grammar there might be too many interacting parts. You may have more success working from the bottom up and/or adding one grammar rule at a time. Catch the offending production rule in the act.

    Grammar Conflict Examples, (cont'd)

    Another Example Shift Reduce Conflict

    A slightly different grammar is needed to demonstrate a shift-reduce conflict:
    T : F g;
    T : F T2 g;
    T2 : t F T2 ;
    T2 : ;
    F : l T r ;
    F : v ;
    
    This grammar is not much different than before, and has the same problem, but the surrounding context (the "calling environments") of F cause the grammar to have a shift-reduce instead of reduce-reduce. Once again, the trouble is after you have seen an F and dwells on the question of whether to reduce the epsilon production, or instead to shift, upon seeing a token g.

    The .output file generated by "bison -v" explains these conflicts in considerable detail. Part of what you need to interpret them are the concepts of "items" and "sets of items" discussed below.

    Mailbag

    I am running into an issue with my grammar when trying to allocate a tree node that uses a precedence rule. For example, I have a grammar:
    expr_nostruct: lit {$$ = allocateTree("expr_nostruct", 1, $1); }
        | %prec IDENT path_expr {$$ = allocateTree("expr_nostruct", 2, $1, $2); }
    	
    When I try to build the grammar I get the error:
    [_]image.png
    
    I noticed that if a grammar rule contains a %prec before a terminal symbol, Bison does not count it as a node that can be passed into the allocateTree function. Do we just ignore terminal symbols if they contain %prec, or is there a way of pulling the information?
    Yeah, so %prec XYZ means "apply this rule using the same precedence as if you had an XYZ terminal". The %prec and the XYZ are not symbols on the stack and do not have $n associated with them, so reducing a path_expr to a expr_nostruct with the precedence of an IDENT will only have one symbol (path_expr) on the righthand side. You can allocateTree with one child or let the default ({ $$=$1; }) take effect.
    What do we do with Kotlin-not-k0 statements? Do we just remove them completely from the grammar?
    What should be the value for (int) production rule in alctree()? Should I just give it some make-up integer? Like 1004, 1005, 1006,...
    Yes, you can just make up integer codes, I recommend codes that are not overlapping with tokens, so starting at 1000 is not a bad idea.
    I can't understand how I can correctly make the root of the tree - how do we supposed to start the root in the start symbols?
    In Bison there is really only one start symbol. If there are multiple production rules that build the start symbol, you would assign the root in each of them. If you assigned to root everywhere(!), it would be silly but OK/harmless in that the last one assigned is probably the one that is for the "real" start symbol.
    In some long grammar rules, I'm confused how the OR ( the vbar | ) works. Does it just use the OR on one word before and after the vbar? For example, I have this grammar:
    optional_comma_test: ',' test | {  $$=NULL; };
    
    So in which way the parser will understand this?
    optional_comma_test:          ( ',' test ) | {  $$=NULL; }     ;
    optional_comma_test:   ','    ( test        | {  $$=NULL; } )      ;
    
    And should I just put parenthesis in cases like this?
    Vertical bar is the lowest possible precedence and actually denotes the start of another production rule, so your interpretation #1 is used without need for parentheses. Parentheses are not used in Bison to force precedence.
    Should we make a tree node for optional type of grammar? An example is
    optional_expr: expr | {  $$=NULL; } ;
    
    This example shows two production rules. The first one has only one child and does not require a new tree node. The second one has zero production rules and can be represented by either NULL or by a special leaf that represents the epsilon.
    How are we supposed to make tree for a prime_type_grammar: for example, I did:
    statements: stmt stmt_prime;
    stmt_prime: stmt_prime stmt;
    stmt_prime: {  $$=NULL; } ;
    
    (which is equivalent to the following)
      statements: stmt statements;
      statements: stmt ;
    
    How do we make tree for this? (do alctree for the first two statements?) And am I doing this correctly at all? I saw your file_input_prime in the lab and I just copied the way. As of typing it right now, I saw 1 of your mailbox's answer saying to just make a epsilon leaf node tree for epsilon rule, so if I just do that for all grammar rules, will that solves this question?
    In this example stmt_prime has two production rules. The first one has two symbols on the righthand side and needs a binary tree node for those two children. The second production rule is an epsilon and needs either a NULL or a special leaf that represents the epsilon.

    YACC precedence and associativity declarations

    YACC headers can specify precedence and associativity rules for otherwise heavily ambiguous grammars. Precedence is determined by increasing order of these declarations. Example:
    %right ASSIGN
    %left PLUS MINUS
    %left TIMES DIVIDE
    %right POWER
    %%
    expr: expr ASSIGN expr
        | expr PLUS expr
        | expr MINUS expr
        | expr TIMES expr
        | expr DIVIDE expr
        | expr POWER expr
        | IDENT
        ;
    

    YACC error handling and recovery

    Improving (YACC, Bison) Syntax Error Reporting

    yyerror(s) overrides the default error message, which usually just says either "syntax error" or "parse error", or "stack overflow".

    You can easily add information in your own yyerror() function. For example, for decades GCC emitted messages that look like:

    goof.c:1: parse error before '}' token
    

    You can do at least that good using a yyerror() function that looks like

    void yyerror(char *s)
    {
       fprintf(stderr, "%s:%d: %s before '%s' token\n",
    	   yyfilename, yylineno, s, yytext);
    }
    
    Reminders:

    You could instead, use the error recovery mechanism to produce better messages. For example

    lbrace : LBRACE | { error_code=MISSING_LBRACE; } error ;
    
    Where LBRACE is an expected token {
    This uses a global variable error_code to pass parse information to yyerror().

    Another related option is to call yyerror() explicitly with a better message string, and tell the parser to recover explicitly:

    package_declaration: PACKAGE_TK error
    	{ yyerror("Missing name"); yyerrok; } ;
    

    But, using error recovery to perform better error reporting runs against conventional wisdom that you should use error tokens very sparingly. What information from the parser determined we had an error in the first place? Can we use that information to produce a better error message?

    lecture #11 began here

    Mailbag

    I've got reduce/reduce errors! What do I do?
    What do we do with the warning of "nonterminal useless in grammar"? I'm guessing it means those grammars will not be used during parsing? Is that correct? Are we ignoring them or are we supposed to fix them somehow?
    Nonterminal useless in grammar means what it says: there is no path from the start symbol that ever uses those non-terminals so they are dead code and will be unused. If it is for a feature that is not in PunY it might be harmless, but if it is a feature that normal Python programs would use then it indicates a bug where some production rule is missing or wrong, that should have referenced that non-terminal from someplace higher in the grammar.
    I got my trees printing for toy examples, am I done?
    You are responsible for thoroughly testing your code, including constructing test cases. For homeworks from now on, try and provide tests so that each tree-constructor code fragment that you write gets used by at least one test case (achieve statement-level test coverage). We will have one or more labs on constructing test cases.
    When we are calling alctree(), does the order that we pass the non-terminals to alctree matter?
    Hmm, in theory, no. In practice, I recommend that you Keep It Simple and stick to the order given in the production rule.
    For error reporting, do we have to use yyerror? I was thinking of implementing another argument in my alctree() function to accept error types (no error, error Kotlin stmt not in k0, anything else I can think of) to do so, but I do not know if you would allow it and/or if it would be more difficult to do that instead of yyerror.
    yyerror() is the official syntax error reporting mechanism of Bison. It is called implicitly by the parser unless you say yerrok inside a rule with an error recovery token. It would be difficult to subvert this in the general case. But
    Can I delete some rules and tokens in Kotlin not in k0? I am quickly running out of time because there are many hundreds of production rules in k0gram.y and adding a call to alctree() for each one is taking a lot of time and I am getting all sorts of weird errors.
    Do whatever you have to do. I mean, long-term I recommend you get really proficient at a powerful code editor that has a good keyboard-replay-macros capability and such. But short-term you can do what is necessary. I recommended pasting in a call to an error function as the semantic action for many unused Kotlin features, but deleting those grammar rules would at least turn those things into syntax errors instead of allowing unsupported features to result in undefined behavior from your compiler. So yeah, I will take what I can get. A small number of points will be associated with Kotlin-not-k0 errors.
    Is there a way to debug syntax errors in our grammar .y file? We are really having a hard time to solve the syntax error ( incomplete sequence error) that we are getting from our bison file.
    You may use a combination of valgrind + gdb, but your best bet may be to turn on bison debugging:
    1. %{ #define YYDEBUG 1 %} in your .y file header section
    2. extern int yydebug; ... and yydebug=1; in your main() module
    I did that yesterday for a stuck student, and it showed that they were returning an IDENT token for one of their reserved words.

    What questions have you got for me, regarding HW#3?









    When Leaves in a Syntax Tree are Needed...and when NOT

    In case it didn't get said elsewhere:

    Clarification on Error Recovery Behavior

    One might wonder why
    statement: error ';' ;
    
    would let you recover from errors when you get to the end of a statement. As a student noted, it looks like it would only work if the error was the last thing before a semi-colon, but usually errors happen in the middle, and a bunch of input symbols have to be discarded before we would get to a semi-colon. I have adapted the answer here from the Bison manual.

    What happens if a syntax error occurs in the middle of an expression? The error recovery rule, interpreted strictly, says to build a statement from the precise sequence of an error followed by a semicolon. If an error occurs in the middle of an expression, there will probably be some additional tokens and subexpressions on the stack after the last stmts, and there will be tokens to read before the next newline. So the rule is not applicable in the ordinary way.

    But Bison can force the situation to fit the rule, by discarding part of the semantic context and part of the input. First it discards states and objects from the stack until it gets back to a state in which the error token is acceptable. (This means that in performing error recovery, the current statement's subexpressions that are already parsed are discarded, back to the last complete statement.) At this point the error token can be shifted. Then, if the old lookahead token is not acceptable to be shifted next, the parser reads tokens and discards them until it finds a token which is acceptable. In this example, Bison reads and discards input until the next newline so that the fourth rule can apply. Note that discarded symbols are possible sources of memory leaks, see Freeing Discarded Symbols, for a means to reclaim this memory.

    This massive throwing-away-of-stuff is a good argument in support of Levine's advice that the error token should be used sparingly: the error token is a bit of a loaded cannon.

    LR Syntax Error Messages: Advanced Methods

    The pieces of information that YACC/Bison use to determine that there is an error in the first place are the parse state (yystate) and the current input token (yychar). These are exactly the pieces of information one might use to produce better diagnostic error messages without relying on the error recovery mechanism and mucking up the grammar with a lot of extra production rules that feature the error token.

    Inside __yyerror(msg, yystate) you can use a switch statement or a global array to associate messages with specific parse states. But, figuring out which parse state means which syntax error message would be by trial and error.

    Merr

    So, your parse state (yystate) and current token (yychar) are how Bison discovers that there is a syntax error. If you have access to that same information, you can write a different message for each parse state, or in the extreme case you could write a different message with each unique (yystate,yychar) tuple, and provide wayyy better syntax error messages than just saying "parse error".

    To solve this problem, I created a tool called Merr. Merr can generate your yyerror() function with good syntax error messages based on examples: you supply the sample syntax errors and messages, and Merr figures out which parse state integer goes with which message. Merr also uses the yychar (current input token) to refine the diagnostics in the event that two of your example errors occur on the same parse state. See the Merr web page.

    Example merr input. The format is:
          error-fragment:::message\n

    procedure p(); 1 := 2 end
    ::: missing semicolon or operator
    procedure main()
    every x do { }
    }
    end
    ::: too many closing curly braces
    procedure main()
      y := X(1)
      z := "a","b"
    end
    ::: missing semicolon or operator
    global::: unexpected end of file
    global x y::: invalid global declaration
    global x, , y::: missing identifier
    procedure p(x) end::: missing semicolon
    link procedure p(x)
    end
    ::: link list expected
    

    Recursive Descent Parsing

    Perhaps the simplest parsing method, for a large subset of context free grammars, is called recursive descent. It is simple because the algorithm closely follows the production rules of nonterminal symbols.

    Recursive Descent Parsing Example #1

    E -> E + T
    E -> T
    T -> T * F
    T -> F
    F -> ( E )
    F -> ident
    
    Consider the grammar we gave above. There will be functions for E, T, and F. The function for F() is the "easiest" in some sense: based on a single token it can decide which production rule to use. The parsing functions return 0 (failed to parse) if the nonterminal in question cannot be derived from the tokens at the current point. A nonzero return value of N would indicate success in parsing using production rule #N.
    int F()
    {
       int t = yylex();
       if (t == IDENT) return 6;
       else if (t == LP) {
          if (E() && (yylex()==RP) return 5;
          }
       return 0;
    }
    
    Comment #1: if F() is in the middle of a larger parse of E() or T(), F() may succeed, but the subsequent parsing may fail. The parse may have to backtrack, which would mean we would have to be able to put tokens back for later parsing. Add a memory (say, a gigantic array or link list for example) of already-parsed tokens to the lexical analyzer, plus backtracking logic to E() or T() as needed. The call to F() may get repeated following a different production rule for a higher nonterminal.

    Comment #2: in a real compiler we need more than "yes it parsed" or "no it didn't": we need a parse tree if it succeeds, and we need a useful error message if it did not.

    Question: for E() and T(), how do we know which production rule to try? Option A: just blindly try each one in turn. Option B: look at the first (current) token, only try those rules that start with that token (1 character lookahead). If you are lucky, that one character will uniquely select a production rule. If that is always true through the whole grammar, no backtracking is needed.

    Question: how do we know which rules start with whatever token we are looking at? Can anyone suggest a solution, or are we stuck?

    Below is an industrious start of an implementation of the corresponding recursive descent parser for non-terminal T. Now is student-author time, what is our next step? What is wrong with this picture?

    int T()
    {  // save where the current token is
       if (T() && (yylex()==ASTERISK) && F()) return 3;
       // restore the current input pointer to the saved location
       if (F()) return 4;
       return 0;
    }
    

    Removing Left Recursion

    E -> E + T | T
    T -> T * F | F
    F -> ( E ) | ident
    
    We can remove the left recursion by introducing new nonterminals and new production rules.
    E  -> T E'
    E' -> + T E' | ε
    T  -> F T'
    T' -> * F T' | ε
    F  -> ( E ) | ident
    
    Getting rid of such immediate left recursion is not enough, one must get rid of indirect left recursion, where two or more nonterminals are mutually left-recursive. One can rewrite any CFG to remove left recursion (Algorithm 4.19).
    for i := 1 to n do
       for j := 1 to i-1 do begin
          replace each Ai -> Aj γ with productions
             Ai -> δ1γ | δ2γ | ... | δkγ, where
                Aj -> δ1 | δ2 | ... | δk are all current Aj-productions
          end
       eliminate immediate left recursion
    

    lecture #12 began here

    Mailbag

    Can I have an extra day to complete the lab?
    Labs are almost credit/non-credit and should reflect ~3 hours work. They are not graded for correctness. I grab them as a batch at some point after they are due and often won't notice ones that are a little late that are submitted by the time I grab the batch. Also, illnesses and other documented university-excused reasons may absolve lateness.
    Can I have an extra day to complete the homework?
    Homeworks are more serious than labs, but illnesses and other documented university-excused reasons still absolve lateness. Other than that, please get help when needed and do your best.
    I have reduce/reduce conflicts on simple identifiers that might either be an expression or a type name, encountered on a closing right parenthesis. What do I do?
    Maybe K0 does not need the rule
    	type: LPAREN type RPAREN
          
    What do you all think? More generally: delete or alter the Kotlin syntax grammar as necessary to get regular Kotlin examples parsing OK. Don't think of it as something you can't change. Own your version of it, without deliberately introducing not-Kotlin syntax.
    My group member doesn't respond to e-mails. What do I do?
    It is reasonable for us to modify and/or reconstitute groups for future assignments. If you would like to be in a new group or additional members for an existing group, please let me know. If you are working alone by choice: on the one hand I've taught compilers many years with all homeworks as individual assignments and that has been fine. On the other hand, we have a course learning outcome for group work, and writing a compiler is a Big job.
    How do we get from the list of token rules into storing them in a tree.
    Wrap your struct token in a struct tree, and stick it in yylval.treeptr... all from inside of yylex(). yyparse() will copy yylval onto the value stack. You will later see it as a $1 or $2 or whatever.
    Where is this tree supposed to be stored?
    How it is supposed to be built?
    These trees get built bottom-up one node at a time. One node gets created for each shift or reduce operation.
    How to prepare the tree for the next layer of the compiler.
    After assembling the whole tree, stick it (the root) in a global variable.
    In what location do we store the head pointer to the tree?
    In the last "reduce" operation, it is normal to finally assemble the whole syntax tree and either stick it in a global variable for use after yyparse() returns, or else to pass it as a parameter in a call to the rest of the compiler, from within yyparse().
    where do we store the current branch we are adding onto in the tree?
    It goes on the top of the value stack. when we are reducing a production this is done by assignment to $$. When we are shifting a token, it is done by assignment to yylval.
    In what order does the tree get built and how is it structured?
    It gets built bottom-up from leaves (terminals) to internal nodes (non-terminals). It is structured as structures, some of which contain pointers to their children.
    I have my lex file set up and it calls a function that looks at the current yytext and then allocates a new tree node with that token. Is there some yy variable that is intended to represent this current token?
    yylval is copied onto the value stack, where it is subsequently referred to as $1, $2, or $3, etc.
    I assume the values that bison is looking at are the ones that are returned from the lexer.
    For terminal symbols that is correct.
    So (yylex() returns) this integer value, and bison looks though its rules and attempts to parse each of the rules as if it works in parallel...
    Well, bison analyzed your grammar and generated a pushdown automaton that tracks all possible shifts/reduces at each point...
    then when the parser reduces for one production rule, it runs the code with the $$ = ... to assign it to... where?
    Assigning to $$ is specifying the value to be pushed onto the value stack in the slot that corresponds to the non-terminal that we have just built on the parse stack using this grammar rule.
    how do we understand when a string of tokens should be put all together as children, compared to when we have a string of tokens that need to be a chain?
    That will depend precisely on how the context free grammar production rules are organized. If you have a production
      A: B C D E F G
      
    then you have to see six things previously parsed successfully before you can reduce them down to an A node. If you have
      A : B C
      C : C D | ;
      D : D E | ;
      E : E F | ;
      F : F G | ;
      
    Your "tree" might look more like a linked list when you get finished.
    What exactly gives the shape of the tree? I know that it is formed from the rules defined in bison, but I am having trouble visualizing it.
    At each node of the tree, the shape (a.k.a. "fan-out", or # of children) is defined by the # of symbols on the righthand side of the production rule used to construct that node. As an example, for the Go language, the distribution at one point in time was about as follows:

    Size of RHS # of Rules of that size
    0 19
    1 98
    2 46
    3 67
    4 17
    5 21
    6 2
    7 1
    8 2
    Kotlin would have a similar distribution, but different (LOL).

    I totally have an example where a shift-reduce conflict was a Real Problem even though you said we could ignore shift-reduce conflicts!
    Ouch! A real shift-reduce problem! You might fix it by simply changing a right recursion to a left recursion! Very cool, we finally know why Bison warns of this kind of ambiguity in the grammar: sometimes it is really a problem. I have taken the liberty of reducing your example to just about its simplest form:
    %%
    Program:	DeclarationList ProgramBody ;
    ProgramBody: 	Function SEMICOLON ProgramBody	| ;
    Function:	Declaration OPEN_PAREN CLOSE_PAREN ;
    DeclarationList:Declaration SEMICOLON DeclarationList | ;
    Declaration:		    INT IDENTIFIER ;
    
    The corresponding input that dies on this is:
    int x;
    int main();
    
    How about a tool that would generate numbers automatically from our grammar .y files? It should perhaps use negative numbers (to avoid overlap/conflicts with Bison-generated numbers for terminal symbols).
    We looked again to see if Bison had an option to generate that, but I am not aware of one. Awhile back I wrote a cheap hack version 0 of such a tool...feel free to adapt it or rewrite something similar.

    Where We Are

    Removing Left Recursion, part 2

    Left recursion can be broken into three cases

    case 1: trivial

    A : A α | β
    
    The recursion must always terminate by A finally deriving β so you can rewrite it to the equivalent
    A : β A'
    A' : α A' | ε
    
    Example:
    E : E op T | T
    
    can be rewritten
    E : T E'
    E' : op T E' | ε
    

    case 2: non-trivial, but immediate

    In the more general case, there may be multiple recursive productions and/or multiple non-recursive productions.
    A : A α1 | A α2 | ... | β1 | β2
    
    As in the trivial case, you get rid of left-recursing A and introduce an A'
    A :  β1 A' | β2 A' | ...
    A' : α1 A' | α2 A' | ... | ε
    

    case 3: mutual recursion

    1. Order the nonterminals in some order 1 to N.
    2. Rewrite production rules to eliminate all nonterminals in leftmost positions that refer to a "previous" nonterminal. When finished, all productions' right hand symbols start with a terminal or a nonterminal that is numbered equal or higher than the nonterminal no the left hand side.
    3. Eliminate the direct left recusion as per cases 1-2.

    Left Recursion Versus Right Recursion: When does it Matter?

    A student came to me once with what they described as an operator precedence problem where 5-4+3 was computing the wrong value (-2 instead of 4). What it really was, was an associativity problem due to the grammar:
    E : T + E | T - E | T
    
    The problem here is that right recursion is forcing right associativity, but normal arithmetic requires left associativity. Several solutions are: (a) rewrite the grammar to be left recursive, or (b) rewrite the grammar with more nonterminals to force the correct precedence/associativity, or (c) if using YACC or Bison, there are "cheat codes" we will discuss later to allow it to be majorly ambiguous and specify associativity separately (look for %left and %right in YACC manuals).

    Recursive Descent Parsing Example #2

    The grammar
    S -> A B C
    A -> a A
    A -> ε
    B -> b
    C -> c
    
    maps to pseudocode like the following. (:= is an assignment operator)
    procedure S()
      if A() & B() & C() then succeed # matched S, we win
    end
    
    procedure A()
      if yychar == a then { # use production 2
         yychar := scan()
         return A()
         }
      else
         succeed # production rule 3, match ε
    end
    
    procedure B()
       if yychar == b then {
          yychar := scan()
          succeed
          }
       else fail
    end
    
    procedure C()
       if yychar == c then {
          yychar := scan()
          succeed
          }
       else fail
    end
    

    Backtracking?

    Could your current token begin more than one of your possible production rules? Try all of them, remember and reset state for each try.
    S -> cAd
    A -> ab
    A -> a
    
    Left factoring can often solve such problems:
    S -> cAd
    A -> a A2
    A2-> b
    A2-> (ε)
    
    One can also perform left factoring to reduce or eliminate the lookahead or backtracking needed to tell which production rule to use. If the end result has no lookahead or backtracking needed, the resulting CFG can be solved by a "predictive parser" and coded easily in a conventional language. If backtracking is needed, a recursive descent parser takes more work to implement, but is still feasible. As a more concrete example:
    S -> if E then S
    S -> if E then S1 else S2
    
    can be factored to:
    S -> if E then S S'
    S'-> else S2 | ε
    

    Reading Assignment

    We are heading into semantic analysis.

    Some More Parsing Theory

    Automatic techniques for constructing parsers start with computing some basic functions for symbols in the grammar. These functions are useful in understanding both recursive descent and bottom-up LR parsers.

    First(α)

    First(α) is the set of terminals that begin strings derived from α, which can include ε.
    1. First(X) starts with the empty set.
    2. if X is a terminal, First(X) is {X}.
    3. if X -> ε is a production, add ε to First(X).
    4. if X is a non-terminal and X -> Y1 Y2 ... Yk is a production, add First(Y1) to First(X).
    5. for (i = 1; if Yi can derive ε; i++)
              add First(Yi+1) to First(X)
      

    First(a) examples

    by the way, this stuff is all in section 4.3 in your text.

    Last time we looked at an example with E, T, and F, and + and *. The first-set computation was not too exciting and we need more examples.

    stmt : if-stmt | OTHER
    if-stmt:  IF LP expr RP stmt else-part
    else-part: ELSE stmt | ε
    expr: IDENT | INTLIT
    
    What are the First() sets of each nonterminal?

    Follow(A)

    (The helper function that goes along with First(X))

    Follow(A) for nonterminal A is the set of terminals that can appear immediately to the right of A in some sentential form S -> aAxB... To compute Follow, apply these rules to all nonterminals in the grammar:

    1. Add $ to Follow(S)
    2. if A -> aBβ then add First(β) - ε to Follow(B)
    3. if A -> aB or A -> aBβ where ε is in First(β), then add Follow(A) to Follow(B).

    Follow() Example

    For the grammar:
    stmt : if-stmt | OTHER
    if-stmt:  IF LP expr RP stmt else-part
    else-part: ELSE stmt | ε
    expr: IDENT | INTLIT
    
    It can get pretty muddy on the Follow() function, for even this simple grammar. It helps if you follow the algorithm, instead of just "eyeballing it".
    For all non-terminals X in the grammar do
       1. if X is the start symbol, add $ to Follow(X)
       2. if N -> αXβ then add First(β) - ε to Follow(X)
       3. if N -> αX or N -> αXβ where ε is in
           First(β) then add Follow(N) to Follow(X)
    
    Since the algorithm depends on First(), what are First sets again?
    First(stmt) = {IF, OTHER}
    First(if-stmt) = {IF}
    First(else-part) = {ELSE, ε}
    First(expr) = {IDENT, INTLIT}
    
    Because each non-terminal has three steps, and our toy grammar has 4 non-terminals, there are 12 steps. When you just apply these twelve steps, brute force, it is clear that the statement of what to do to compute them was not an algorithm, it was only a declarative specification, and there is an ordering needed in order to compute the result.
       1. stmt is the start symbol, add $ to Follow(stmt)
       2. if N -> α stmt β then add First(β) - ε to Follow(stmt)
    	---- add First(else-part)-ε to Follow(stmt)
       3. if N -> α stmt or N -> α stmt β where ε
    	 is in First(β) then add Follow(N) to Follow(stmt)
    	---- add Follow(else-part) to Follow(stmt)
       4. if-stmt is not the start symbol (noop)
       5. if N -> αif-stmtβ then add First(β) - ε to Follow(if-stmt)
    	---- n/a
       6. if N -> αif-stmt or N -> αif-stmtβ where ε is in
           First(β) then add Follow(N) to Follow(if-stmt)
    	---- add Follow(stmt) to Follow(if-stmt)
       7. else-part is not the start symbol (noop)
       8. if N -> αelse-partβ then add First(β) - ε to Follow(else-part)
    	---- n/a
       9. if N -> αelse-part or N -> αelse-partβ where ε is in
           First(β) then add Follow(N) to Follow(else-part)
    	--- add Follow(if-stmt) to Follow(else-part)
       10. expr is not the start symbol (noop)
       11. if N -> αexprβ then add First(β) - ε to Follow(expr)
    	---- add RP to Follow(expr)
       12. if N -> αexpr or N -> αexprβ where ε is in
           First(β) then add Follow(N) to Follow(expr)
    	---- n/a
    
    What is the dependency graph? Does it have any cycles? If it has cycles, you will have to iterate to a fixed point.
    Follow(stmt) depends on Follow(else-part)
    Follow(if-stmt) depends on Follow(stmt)
    Follow(else-part) depends on Follow(if-stmt)
    
    If I read this right, there is a 3-way mutual recursion cycle.

    Can we First/Follow Anything Else

    Like preferably, a real-world grammar example? Please remember that real world grammars for languages like ANSI C are around 400+ production rules, so in-class examples will by necessity be toys. If I pick a random* (*LOL) YACC grammar, can we First/Follow any of its non-terminals?

    LR vs. LL vs. LR(0) vs. LR(1) vs. LALR(1)

    The first char ("L") means input tokens are read from the left (left to right). The second char ("R" or "L") means parsing finds the rightmost, or leftmost, derivation. Relevant if there is ambiguity in the grammar. (0) or (1) or (k) after the main lettering indicates how many lookahead characters are used. (0) means you only look at the parse stack, (1) means you use the current token in deciding what to do, shift or reduce. (k) means you look at the next k tokens before deciding what to do at the current position.

    Midterm

    The midterm exam March 11 will be an in-class exam. There will be a midterm review on Thursday March 6.

    LR Parsers

    LR denotes a class of bottom up parsers that is capable of handling virtually all programming language constructs. LR is efficient; it runs in linear time with no backtracking needed. The class of languages handled by LR is a proper superset of the class of languages handled by top down "predictive parsers". LR parsing detects an error as soon as it is possible to do so. Generally building an LR parser is too big and complicated a job to do by hand, we use tools to generate LR parsers.

    The LR parsing algorithm is given below.

    ip = first symbol of input
    repeat {
       s = state on top of parse stack
       a = *ip
       case action[s,a] of {
          SHIFT s': { push(a); push(s') }
          REDUCE A->β: {
             pop 2*|β| symbols; s' = new state on top
             push A
             push goto[s', A]
             }
          ACCEPT: return 0 /* success */
          ERROR: { error("syntax error", s, a); halt }
          }
       }
    

    On Trees

    Trees are classic data structures.

    Parse trees are k-ary, where there is a variable number of children bounded by a value k determined by the grammar. You may wish to consult your old data structures book, or look at some books from the library, to learn more about trees if you are not totally comfortable with them.

    #include <stdarg.h>
    
    struct tree {
       short label;			/* what production rule this came from */
       short nkids;			/* how many children it really has */
       struct tree *child[1];	/* array of children, size varies 0..k */
    				/* Such an array has to be the LAST
    				   field of a struct, and "there can
    				   be only ONE" for this to work. */
    };
    
    struct tree *alctree(int label, int nkids, ...)
    {
       int i;
       va_list ap;
       struct tree *ptr = malloc(sizeof(struct tree) +
                                 (nkids-1)*sizeof(struct tree *));
       if (ptr == NULL) {fprintf(stderr, "alctree out of memory\n"); exit(1); }
       ptr->label = label;
       ptr->nkids = nkids;
       va_start(ap, nkids);
       for(i=0; i < nkids; i++)
          ptr->child[i] = va_arg(ap, struct tree *);
       va_end(ap);
       return ptr;
    }
    

    lecture #13 began here

    Having Trouble Debugging?

    To save yourself on the semester project in this class, you should learn gdb (or some other source level debugger) as well as you can. Sometimes it can help you find your bug in seconds where you would have spent hours without it. But only if you take the time to read the manual and learn the debugger.

    To work on segmentation faults: recompile all .c files with -g and run your program inside gdb to the point of the segmentation fault. Type the gdb "where" command. Print the values of variables on the line mentioned in the debugger as the point of failure. If it is inside a C library function, use the "up" command until you are back in your own code, and then print the values of all variables mentioned on that line.

    After gdb, the second tool I recommend strongly is valgrind. valgrind catches some kinds of errors that gdb misses. It is a non-interactive tool that runs your program and reports issues as they occur, with a big report at the end. I recommend you run valgrind periodically in between debugging sessions, just to catch and fix things early. I recommend that you have things valgrind-clean before coming to me for help, although I will help you fix things in valgrind if you can't figure them out.

    Reading Tree Leaves

    In order to work with your tree, you must be able to tell, preferably trivially easily, which nodes are tree leaves and which are internal nodes, and for the leaves, how to access the lexical attributes.

    Options:

    1. encode in the parent what the types of children are
    2. encode in each child what its own type is (better)
    How do you do option #2 here?

    There are actually nonterminal symbols with 0 children (nonterminal with a righthand side with 0 symbols) so you don't necessarily want to use an nkids of 0 is your flag to say that you are a leaf. Perhaps the best approach to all this is to unify the tokens and parse tree nodes with something like the following, where perhaps an nkids value of -1 is treated as a flag that tells the reader to use lexical information instead of pointers to children:

    struct node {
    int code;		/* terminal or nonterminal symbol */
    int nkids;
    union {
       struct token { ...  } leaf; // or: struct token *leaf;
       struct node *kids[9];
       }u;
    } ;
    

    Tree Traversals

    Besides a function to allocate trees, you need to write one or more recursive functions to visit each node in the tree, either top to bottom (preorder), or bottom to top (postorder). You might do many different traversals on the tree in order to write a whole compiler: check types, generate machine- independent intermediate code, analyze the code to make it shorter, etc. You can write 4 or more different traversal functions, or you can write 1 traversal function that does different work at each node, determined by passing in a function pointer, to be called for each node.
    void postorder(struct tree *t, void (*f)(struct tree *))
    {
       /* postorder means visit each child, then do work at the parent */
       int i;
       if (t == NULL) return;
    
       /* visit each child */
       for (i=0; i < t-> nkids; i++)
          postorder(t->child[i], f);
    
       /* do work at parent */
       f(t);
    }
    
    You would then be free to write as many little helper functions as you want, for different tree traversals, for example:
    void printer(struct tree *t)
    {
       if (t == NULL) return;
       printf("%p: %d, %d children\n", t, t->label, t->nkids);
    }
    

    Observations on Debugging the ANSI C++ Grammar to be more YACC-able

    Expectation
    not that you pick it up by magic and debug it all yourself, but rather that you spend enough time monkeying with yacc grammars to be familiar with the tools and approach, and to ask the right questions.
    Tools
    YYDEBUG/yydebug, --verbose/--debug/y.output
    Approach

    *why? why not?

    Suggestions on HW

    Did you Test your Work on login.cs.nmt.edu?
    Lots of folks doing work on lots of OSes, but if it doesn't run well on the test machine, you won't get many points.
    Warnings are seldom OK
    shift/reduce warnings are "usually" OK (not always). Get rid of other warnings so that when warning of a real issue shows up, you do not ignore it like "the boy who cried Wolf!".
    Using { $$ = $4; } is probably a bad idea
    Q: Why? Q: under what circumstances is this fine?
    Using { $$ = $1; } goes without saying
    It is the default... but epsilon rules had better not try it.
    passing an fopen() or a malloc() as a parameter into a function is probably a bad idea
    usually, this is a resource leak. It gives you no clean and safe way to close/free.
    Some of you are still not commenting to a minimum professional level needed for you to understand your own code in 6 months

    Mailbag

    Since this will matter later, do we support generic types?
    Only for Array and only for five built-in types of array elements.
    You marked me down for Valgrind, but I did not have illegal memory reads or writes! What gives?
    From hw1.html:
    For the purposes of this class, a "memory error" is a message from valgrind indicating a read or write of one or more bytes of illegal, out-of-bounds, or uninitialized memory.
    The uninitialized memory part includes messages such as:
    ==25504== Conditional jump or move depends on uninitialised value(s)
    
    You have been told that the valgrind header and summary, including memory leaks, are not going to cost you points, I am only interested in valgrind error messages reported for behavior at runtime. Any stuff you see in between the valgrind header and summary is either your output, or valgrind messages that may point at bugs in your code.

    Lab#5 Comments

    Semantic Analysis

    Semantic ("meaning") analysis refers to a phase of compilation in which the input program is studied in order to determine what operations are to be carried out. The two primary components of a classic semantic analysis phase are variable reference analysis and type checking. These components both rely on an underlying symbol table.

    What we have at the start of semantic analysis is a syntax tree that corresponds to the source program as parsed using the context free grammar. Semantic information is added by annotating grammar symbols with semantic attributes, which are defined by semantic rules. A semantic rule is a specification of how to calculate a semantic attribute that is to be added to the parse tree.

    So the input is a syntax tree...and the output is the same tree, only "fatter" in the sense that nodes carry more information. Another output of semantic analysis are error messages detecting many types of semantic errors.

    Two typical examples of semantic analysis include:

    variable reference analysis
    the compiler must determine, for each use of a variable, which variable declaration corresponds to that use. This depends on the semantics of the source language being translated.
    type checking
    the compiler must determine, for each operation in the source code, the types of the operands and resulting value, if any.

    Notations used in semantic analysis:

    syntax-directed definitions
    high-level (declarative) specifications of semantic rules
    translation schemes
    semantic rules and the order in which they get evaluated

    In practice, attributes get stored in parse tree nodes, and the semantic rules are evaluated either (a) during parsing (for easy rules) or (b) during one or more (sub)tree traversals.

    Two Types of Attributes:

    synthesized
    attributes computed from information contained within children. These are generally easy to compute, even on-the-fly during parsing.
    inherited
    attributes computed from information obtained from elsewhere in the tree, such as a parent or siblings. These are generally harder to compute. Compilers may be able to jump through hoops to compute some inherited attributes during parsing, but depending on the semantic rules this may not be possible in general. Compilers resort to tree traversals to move semantic information around the tree to where it will be used.

    Look at HW#4

    Attribute Examples

    Semantic attributes are used for a lot of things from now on in the course, such as type checking, but here are a couple easier examples.

    Isconst and Value

    Not all expressions have constant values; the ones that do may allow various optimizations.

    CFG Semantic Rule
    E1 : E2 + T E1.isconst = E2.isconst && T.isconst
    if (E1.isconst)
        E1.value = E2.value + T.value
    E : T E.isconst = T.isconst
    if (E.isconst)
        E.value = T.value
    T : T * F T1.isconst = T2.isconst && F.isconst
    if (T1.isconst)
        T1.value = T2.value * F.value
    T : F T.isconst = F.isconst
    if (T.isconst)
        T.value = F.value
    F : ( E ) F.isconst = E.isconst
    if (F.isconst)
        F.value = E.value
    F : ident F.isconst = FALSE
    F : intlit F.isconst = TRUE
    F.value = intlit.ival

    Mailbag

    Should we have 1 symbol table for each scope? For example a symbol table for global, then another for functions?
    1 symbol table for global scope, one symbol table for each function, possibly some additional as needed for built-ins tbd.
    What are the built-ins? Where are they described? How are we supposed to implement them?
    Good questions. What do you think should be in the minimal set of Kotlin built-ins?
    println()
     
    Sure, now what about input? 3 (out of many) options include:
    std::io::stdin().read_line(&mut input) // booo!
    
    console...
    
    text_io:
      let i : i64 = read!();
    
    
    After an in-class discussion, we decided hey maybe macro read!(); would be the simplest and most useful.

    Symbol Table Module

    The following very abstract API description from the red dragon book might give you an idea of some of the operations you would want to implement.

    mktable(parent)
    creates a new symbol table, whose scope is local to (i.e. inside) a parent symbol table (the parameter). The outermost table has a null parent -- usually the "global" symbol table.
    enter(table, symbolname, type, offset)
    insert a symbol into a table. Type and offset information will be discussed a little later. There might be more info about the symbol (i.e. more parameters) that you need to store.
    lookup(table, symbolname)
    lookup a symbol in a table; returns a pointer to a structure (or object) that has type and offset information. lookup operations are often chained together progressively from most local scope on out to global scope.
    compute_width(table)
    sums the widths of all entries in the table. Worry not about method compute_width() until code generation you wish to implement.
    These symbol tables are attached to scopes, which in most languages are nested inside enclosing scopes, forming a tree of scopes. As one walks around the parse tree one needs to remember which symbol table you are in at any given time. You can implement a stack of symbol tables that you push and pop in parallel with entering and exiting tree nodes that introduce scopes, or you can implement a semantic attribute in your tree nodes so each node knows its scope/symbol table. The "stack" push and pop symbol table API might have
    newscope = enterscope(table, name)
    enters (pushes) the local scope of the named entity, looking up name in table
    newscope = exitscope()
    exits/pops a local scope, returning to the previous scope
    currentscope()
    return the current scope, i.e. the top of the stack

    lecture #14 began here

    The k0 spec

    OK, so some of you found your way to a k0.html language specification file when it was mostly an old Java subset specification. It is now almost complete as a definition of k0. Questions and bug reports are welcome.

    Symbol Table Basics

    The goal of a symbol table is to allow storage and retrieval of variable (and related) information by name. We are associating a data payload with that name, so we need a struct with the name and the data payload, and lookup and insert functions on it. What is the data payload?
    struct symtab_entry {
       char *sym;
       struct typeinfo *type; /* what type is this variable? forthcoming */
       /* ... more stuff added later ... */
    }
    
    We have to be able to look up stuff by name. We could just do this with a linked list:
    struct elem {
       struct symtab_entry *ste; // information about a symbol
       struct elem *next;
       };
    struct elem *theEntireSymbolTable;
    struct symtab_entry *lookup(struct elem *st, char *name) {
       if (st==NULL) return NULL;
       if (!strcmp(st->ste->sym, name)) return st->ste;
       return lookup(st->next, name);
    }
    struct elem *insert(struct elem *st, char *name, struct typeinfo *t) {
       struct elem *n;
       struct symtab_entry *ste = lookup(st, name);
       if (ste != NULL) {
          fprintf(stderr, "symbol is already inserted\n");
          exit(3);
          }
       /* ste was NULL, make a new one */
       ste = malloc(sizeof (struct symtab entry));
       ste->sym = strdup(name);
       ste->type = t;
       n = malloc(sizeof (struct elem));
       n->ste = ste;
       n->next = theEntireSymbolTable;
       theEntireSymbolTable = n;
    }
    

    Aside on malloc()

    malloc() can fail and return NULL. Consider something like the following, that you can use everywhere in place of malloc():
    void *ckalloc(int n) // "checked" allocation
    {
      void *p = malloc(n);
      if (p == NULL) {
         fprintf(stderr, "out of memory for request of %d bytes\n", n)
         exit(4);
      }
      return p;
    }
    

    Hash Functions and Hash Tables for Symbol Tables

    int hash(char *s) { return 0; }   // linked list, O(n)
    int hash(char *s) { return s[0];} // hash using first char, x1 x2 x3 hash same
    int hash(char *s) {               // what does this one do?
       int len = strlen(s);
       return s[0] + (len>1 ? s[len-1] : 0);
    }
    int hash(char *s) {               // "good enough"; what is weak here?
       int i=0, sum = 0, len = strlen(s);
       for( ; i<len; i++) sum += s[i];
       return sum;
    }
    

    Looking at More Symbol Table Examples

    Lessons From the Godiva Project

    By way of comparison, it may be useful for you to look at some symbol tables and type representation code that were written for the Godiva programming language project. Check out its hash function. Being a dialect of Java, Godiva has compile-time type checking and might provide relevant ideas for OOP languages.

    Mailbag

    is it appropriate to make a new symbol table when the parser begins (global symbol table) and then when it encounters a new block as defined in k0gram.y (new scope definition)? And then to make a sym_entry when an identifier is encountered?
    Yes and yes. Well, we said it would be sufficient in k0 to do one global and one for each function.
    [Is it appropriate] to make a sym_entry when an identifier is encountered?
    Not all identifiers. When an identifier appears in a syntactic context in which the identifier is being declared, such as a function name, or a formal parameter name, or a let.
    What is the "struct sym_entry **tbl" in the last definition of the sym_table?
    struct sym_entry **tbl in some (lab 6?) reference code is a declaration of an array of pointers -- the array of heads of linked lists -- declared in such a way that different symbol tables can be malloced with different sizes (== different numbers of buckets). Not required for k0.
    How can symbol tables be made correctly if the parser works bottom up and sees identifiers first?
    You can do what makes sense for you, but semantic analysis is perhaps easiest and best conducted as one or more tree traversals after parsing has been completed. You may have to think hard about what order to do things in. You can make as many passes over the tree as you need.
    when should we be declaring a new symbol table?
    During a tree traversal you are looking for the node(s) that introduce scopes. What string label or not production rule # would you use to recognize these nodes? Answer will vary depending on your grammar.
    I may ... always add the first symbol (which should always be the function name) to the parent symbol table. Is that okay to implement?
    I would characterize this in terms of specific tree shapes where the function name is declared vs. where other symbols local to the function are declared.
    Why is the "struct sym_entry **tbl" of type sym_entry? Should it not be of type sym_table since it will point to other symbol tables? And how should it be initialized?
    A symbol table contains a collection of elements that are the information known about the various symbols in a program scope. These elements are stored in type struct sym_entry, which are often chained together into linked list of all the symbols that happen to hash to the same index in the array of linked lists. The struct sym_entry **tbl is literally intended to be a pointer to an array of buckets (linked list heads) that would typically be initialized like
       st->tbl = calloc(num_buckets, sizeof(struct sym_entry *));
    

    Symbol Table Example

    Consider this simple nonsense program:
    int foo(int x, char *y) {
       return x;
       }
    
    int main()
    {
       int z;
       z = foo(5, "funf");
       return 0;
    }
    
    or, in Python
    def foo( x:int, y:str) -> int :
       return x
    
    z : int
    z = foo(5, "funf")
    
    What does its syntax tree look like? Can we (by hand) simulate the symbol table population process? Even a small example like this is a pain to do by hand!

    Mailbag

    ival, sval and dval are lexical attributes: compile-time things that apply to literal constant tokens only. For variables with type any, we will need a strategy for runtime, and we will need to emit instructions to read and write values at runtime. The virtual machine we will target by default is in fact very well-suited for a lot of this, since it is one of Python's ancestors. From a recent addition to punyref.html:
    The rules are: if you assign a value to a variable without a declaration, the variable is implicitly declared to be of type Any. For example you can say
    x  = 5
    
    Since anything can be assigned to variable of type Any, it is fine receiving a 5. Suppose now that you declare a variable of type int, and assign x to it.
    y : int
    y = x
    
    If the compiler does not know the type of x, the generated code for assignment to y must look like
    y = int(x)
    
    Note that the conversion-to-int is more tied to the assignment-to-an-int rather than to the x that is to be converted. One almost wants to define typed assignment operators like =int and to modify your syntax tree to make it look like:
    y =int x
    
    But I am a pragmatist. You could encode this by modifying the syntax tree to explicitly call the int() function if you want.

    lecture #15 began here

    Midterm Schedule

    Old Mailbag

    How should I focus my midterm studying? There is a lot of material in this class and I would like to try to optimize my study time. Should I focus on the lecture notes? Should I be studying the textbooks?
    The best way to study for the exams in this course is to do your homework assignments. I try to write exam questions that you should know if you have done your assignments. Having said that, if I were picking and choosing between the lecture notes or the books I would hit the lecture notes the hardest, referring to the books when more (or different) explanations are needed.
    I still have no idea where to start my symbol table HW! What do I do?
    The symbol table HW (and next few HW's) are tree traversals. There is but a single, powerful magic tool at your disposal: recursion. Start with basis cases, at leaves. Make it work for the MOST SIMPLE CASES POSSIBLE before you worry about anything bigger. Work your way up the tree.
    Where do I store my symbol tables?
    If I were you, I'd just get a single (global) symbol table working before worrying about local scopes, but... the obvious alternative answers to your question are:
    Do we have to check whether array (or map, if writing a language with maps/dictionaries) subscripts are legal (correct type, in-range) in this homework?
    Checking "legality" is the next (type checking) HW's job. Even then, some checks are not do-able by the compiler and if required, entail generating code that does the checks at runtime.
    My tree's attributes aren't propagating from parent to child, why not?
    If there are tree nodes in the middle, there have to be semantic rules that copy attributes up or down in order for the information to get from source to destination.
    What is wrong with this hash?
    for(i=0; i < strlen(s); i++) {
       sum += s[i];
       sum %= ARRAYSIZE;
       }
    
    How many potential problems can you find in this code?
    I was wondering if it would be better for the hash table to build it based on the terminals I find in the tree or the non-terminals?
    the keys you are inserting and looking up in hash tables are the variable names declared in the program you are parsing -- those names came into your tree as terminals/leaves, and not all the leaves -- only leaves that are names of things (identifier, or LNAME, or whatever you are calling them), and only when those leaves appear in particular subtrees/production rules where new variables or functions (or type names) are being introduced.
    As I am traversing the tree, should I be looking for NAMEs and other terminal symbols, and then to determine if I should insert them, or should I look for nonterminals and then as I see those non terminals grab the LNAME and the other important data
    Sorta the latter, you are usually looking for non terminals or specific production rules, and then traversing selected children within which you know you have a list of names being declared.

    Variable Reference Analysis

    The simplest use of a symbol table would check:

    Semantic Analysis in Concrete Terms

    Broadly, we can envision the semantic analysis as two passes:
    Pass 1: Symbol Table Population
    Symbol table population is a syntax tree traversal in which we look for nodes that introduce symbols, including the creation and population of local scopes and their associated symbol tables. As you walk the tree, we look for specific nodes that indicate symbols are introduced, or new local scopes are introduced. What are the tree nodes that matter (from cgram.y) in this particular example?
    1. create a global symbol table (initialization)
    2. each function_declarator introduces a symbol.
    3. each init_declarator introduces a symbol.
    4. oh by the way, we have to obtain the types for these.
    5. "types" for functions include parameter types and return type
    6. "types" for init_declarators come from declaration_specifiers, which are "uncles" of init_declarators
    Pass 2: Type Checking
    Type checking occurs during a bottom up traversal of the expressions within all the statements in the program.

    Discussion of a Semantic Analysis Example

    Old Mailbag

    I feel like I am just staring at a wall... I am just kinda lost as to how to start.
    Start by copying modifying your tree printer to only print out the names of variables at the point at which they are declared.
    As I am creating new symbol tables, how should I keep track of them? Should I include like a *next pointer?
    There is logically a tree of symbol tables. Parent symbol tables (in our case the symbol table for our "global" package, package main) contain entries for symbols defined within them, such as functions, so the parent should be able to reach the children's symbol tables by looking them up by name within the parent symbol table. On the other hand, children's symbol tables might want to know their parent enclosing symbol table. For general nested programming language the child symbol table should contain a parent pointer. For the special case that is our language this semester, there isn't much nesting and you could just have a global variable that knows the root symbol table (for package main) and every symbol table that is not the root, can rest assured that its parent is the root.
    So, how many symbol tables do we have to have? More than one?
    One for each package, one for each function, one for each struct type.
    How do I know what symbol table I am using?
    One could implement this as an inherited attribute, or one can track it with an auxiliary global as one walks around the tree. If you don't do the inherited attribut, you may have to maintain a stack of scopes. When you walk into something that has a more local scope, make that scope current and use enclosing scopes when more local scopes don't find a particular symbol you are looking for.

    Midterm Exam Review

    The Midterm will cover lexical analysis, finite automatas, context free grammars, syntax analysis, parsing, and semantic analysis*. *except type checking

    Q: What is likely to appear on the midterm?

    A: questions that allow you to demonstrate that you know

    Sample problems:

    1. Write a regular expression for numeric quantities of U.S. money that start with a dollar sign, followed by one or more digits. Require a comma between every three digits, as in $7,321,212. Also, allow but do not require a decimal point followed by two digits at the end, as in $5.99
    2. Write a non-deterministic finite automaton for the following regular expression, an abstraction of the expression used for real number literal values in C.
           (d+pd*|d*pd+)(ed+)? 
    3. Write a regular expression, or explain why you can't write a regular expression, for Modula-2 comments which use (* *) as their boundaries. Unlike C, Modula-2 comments may be nested, as in (* this is a (* nested *) comment *)
    4. Write a context free grammar for the subset of C expressions that include identifiers and function calls with parameters. Parameters may themselves be function calls, as in f(g(x)), or h(a,b,i(j(k,l)))
    5. What are the FIRST(E) and FOLLOW(T) in the grammar:
           E : E + T | T
           T : T * F | F
           F : ( E ) | ident
    6. What is the ε-closure(move({2,4},b)) in the following NFA? That is, suppose you might be in either state 2 or 4 at the time you see a symbol b: what NFA states might you find yourself in after consuming b?
      (automata to be written on the board)
    7. (20 points) (a) Explain why a compiler might be less able to recover and continue from a lexical error than from a syntax error. (b) Explain why a compiler might be less able to recover and continue from a syntax error than from a semantic error.
    8. (30 points) (a) Write a regular expression (you may use Flex extended regular expression operators) for declarations of the form given by the grammar below. You may use the usual regular expression for C/C++ variable names for IDENT. (b) Under what circumstances is it better to use regular expressions, and under what circumstances is it better to use context free grammars?
      declaration : type_specifier decl_list ';' ;
      type_specifier : INT | CHAR | DOUBLE ;
      decl_list : decl | decl ',' decl_list ;
      decl: IDENT | '*' IDENT | IDENT '[' INTCONST ']' ;
      

    9. (30 points) Some early UNIX utilities, like grep and lex, implemented a non-deterministic finite automata interpreter for each regular expression, resulting in famously slow execution. Why is Flex able to run much faster than these early UNIX tools?

    10. (20 points) Perhaps the most important thing to learn in homework #2 about Flex and Bison was how the two tools communicate information between each other. Describe this communications interface.

    11. (30 points) Perhaps the second most important thing to learn in homework #2 was how and when to build internal nodes in constructing your syntax tree.
      (a) Describe how and when internal nodes need to be constructed, in order for a Bison-based parser to end up with a tree that holds all leaves/terminal symbols. (b) Under what circumstances might a new non-terminal node construction site be skipped? (c) Under what circumstances might some of the leaves/terminal symbols not be needed later during compilation?
    12. (40 points) Consider the following grammar for C variable declarations, given in YACC-style syntax. sm stands for semi-colon. cm stands for comma. id stands for identifier. lb stands for left square bracket. intconst stands for integer constant. rb stands for right square bracket.
      VD : CL T DL sm ;
      CL : static | register | /* epsilon */ ;
      T : int ;
      DL : D | D cm DL ;
      D : id | AST D | D lb intconst rb ;
      
      a) What are the terminal symbols? b) What are the nonterminal symbols? c) Which nonterminals have recursive productions? d) Remove left recursive rules from this grammar if there are any.
    13. (30 points) Write C code that is error free and produces no warnings, which performs the following tasks: a) declare a variable of type pointer to struct token, where struct token has an integer category, a string lexeme, and an integer lineno, b) allocate some memory from the heap large enough to hold a struct token and point your variable at it, and c) initialize your memory to all zero bits. You may assume struct token with its field definitions, has already been defined earlier in the C file before your code fragment.
    14. (20 points) In looking at your yydebug output, several of you noticed that it appeared like the same terminal symbol (for example, a semi-colon) was repeated over and over again in the output, even through sections of parsing where no syntax error occurred. Why might the same terminal symbol appear on the input repeatedly through several iterations of a shift-reduce parser. (30 points) What are semantic attributes? Briefly define and give an example of the major kinds of semantic attributes that might be used in semantic analysis for a language such as C++.
    15. (30 points) Symbol tables play a prominent role in semantic analysis. How are symbol tables used? Give an example, with a brief explanation of how the symbol table is involved.

    Representing Types

    This material is similar/redundant to the discussion from the lab on representing types, but is left in here in case it explains anything differently and/or better than the lab.

    In statically-typecheck'ed languages, the target language's type system must be represented using data structures in the compiler's implementation language. In the symbol table and in the parse tree attributes used in type checking, there is a need to represent and compare source language types. You might start by trying to assign a numeric code to each type, kind of like the integers used to denote each terminal symbol and each production rule of the grammar. But what about arrays? What about structs? There are an infinite number of types; any attempt to enumerate them will fail. Instead, you should create a new data type to explicitly represent type information. This might look something like the following:

    struct type {
       /*
        * Integer code that says what kind of type this is.
        * Includes all primitive types: 1 = int, 2=float,
        * Also includes codes for compound types that then also
        * hold type information in a supporting union...
        * 7 = array, 8 = struct, 9 = pointer etc. */
       int base_type;
       union {
          struct array {
             int size; /* allow for missing size, e.g. -1 */
    	 struct type *elemtype; /* pointer to type for elements in array,
    	 				follow it to find its base type, etc.*/
          } a;
          struct struc {		/* structs */
             char *label;
    	 int nfields;
             struct field **f;
    	 } s;
          struct type *p;		/* pointer type, points at another type */
       } u;
    }
    
    struct field {			/* members (fields) of structs */
       char *name;
       struct type *elemtype;
    }
    
    
    Given this representation, how would you initialize a variable to represent each of the following types:
    int [10][20]
    struct foo { int x; char *s; }
    

    For PunY, this might look like:

    struct type {
       /*
        * Integer code that says what kind of type this is.
        * Includes all primitive types:
        * 1 = int, 2=float, 3=string, 4=bool,
        * Also includes codes for compound types that then also
        * hold type information in a supporting union...
        * 5=list, 6=dict., 7=func, 8=class */
       int base_type;
      /* gone away! for PunY */
      union {
       struct funcdef {
          struct type *return_type;
          int nparams;
          struct params **p;
          } f;
    /* maybe we can get away with just "knowing" for only predefined class info
       struct classdef {
          struct methods **meth;
          struct members **mem;
          } f;
     */
       } u;
    }
    
    struct field {			/* members (fields) of structs */
       char *name;
       struct type *elemtype;
    }
    

    Mailbag

    It seems like my approach to finding different cases of variable declarations is not ideal. I'm just looking for some more of the "parse tree theory" behind the variable declarations, and then a little bit of guidance for putting that theory into efficient code.
    Spend some time getting familiar with your trees. Mostly they will reflect your grammar, which you should also get (more) familiar with. And while there does not exist "parse tree theory" or "syntax tree theory", we can certainly talk about "tree theory". It is all about recursion, since trees are recursively defined: what do we do for basis cases (leaves) and recursion steps (internal nodes)?
    "Grok recursion deeply and you can do anything that needs doing on computers"
    is a grotesque paraphrase of my grad school algorithms prof Udi Manber, former VP of Search at Google.

    However you represent it in the actual bits, you have to be able to tell for each tree node what production rule built it. And you have to be able to look for specific nodes, and traverse the unique shapes that they have underneath them.

    Other than that, the main way for me to help you might be to draw the shapes of your (syntax tree) subtrees for various grammar rules, but of course, that depends on what you did in your version of cgram.y

    Does our language support calling a function inside another call, as in
       foo(bar(t1, baz(99,1)));
    
    Yes.
    How to deal with includes/imports: what does -in general- a compiler do? For example, when it includes math.h (or in other languages, imports a math package), does it copy the functions' declarations? Where can I find the math package file? Are there in Bison built-in function to copy these stuff and added them to the source file, or should I do copy content and edited the source file in C at the main function?
    Great question. In a real compiler, an include or import is a pretty big operation. C/C++ #include textually inserts another file, while import is probably reading from a database to get the declarations of the package being imported. For our language if we do anything with this at all, we are doing a hardwired special case, treating these as "built-ins", so the compiler can do whatever it wants in order to get the runtime library functionality it needs. For example, what do we do with this:
    extern crate text_io;
    fn main() {
    let a: i64 = read!();
    }
    Are we supposed to create a separate symbol table for each function? Or just a symbol table for functions in general?
    You are supposed to create one "global" symbol table, one local symbol table for each function, and one local symbol table for each struct type.
    I am struggling on figuring out how to detect undeclared variables.
    We should talk about this in detail looking at the non-terminals used in your grammar. With any big vague software task, it is wise to break it up into smaller, well-defined pieces. Before you try to find all undeclared variables, you could:
    1. write a tree traversal that just lists all the uses of a variable (in expressions, where values are read or written), showing the variable name and line number. These are the things that must be checked.
      Still too big a job? Break it into even smaller pieces:
      • write a tree traversal that just lists the names of functions for which you have a function body, and therefore a compound statement that contains executable expressions.
      • are there anything besides function bodies where you would have to check for undeclared variables?
    2. write a tree traversal that inserts all the variable declarations. print out the whole symbol table when finished, to show what you've got.
    3. modify the tree traversal #1 to lookup within the symbol table(s) and print semantic errors if any lookup fails.

    lecture #16 began here

    Reading Assignment

    What to do about predefined things

    Discussion of k0 built-ins

    Discussion of "Import", and more Generally, Packages

    Not all languages have them, but when you do...

    Suggested approaches for implementing semantic analysis of packages/imports:

    Option A: treat import like a special "include"
    Option B: respond to "import" by inserting some symbol table entries (hardwired to the package name)

    Mailbag

    I'm trying to make a given node's symbol table into an inherited attribute. However, whenever I #include "symtab.h" in my tree.h file so that I can add a SymbolTable attribute to my tree nodes, I get function redeclaration errors for everything that's in my symtab.h file. Basically, I think the problem is that symtab.h is getting included in too many places. Do you have any suggestions on how to share symtab.h into tree.h?
    Consider which case each .h file falls into: The typical way of protecting a file foo.h from being multiply included is:
    #ifndef FOO_H
    #define FOO_H
    ...rest of your .h file
    #endif
    
    I've seen folks manage to get into trouble despite such a protection, or implement such a protection incorrectly, but it takes work to do so.

    HW#5

    Lab #6 Comment

    Comments on Past HW

    Building a Type struct from a Syntax Tree Fragment

    /*
     * Build Type From Prototype (syntax tree) Example
     */
    void btfp(nodeptr n)
    {
       if (n==NULL) return;
       for(int i = 0; i < n->nkids; i++) btfp(n->child[i]);
       switch (n->prodrule) {
       case INT:
          n->type = get_type(INTEGER);
          break;
       case CHAR:
          n->type = get_type(CHARACTER);
          break;
       case IDENTIFIER:
          n->type = get_type(DONT_KNOW_YET);
          break;
       case '*':
          n->type = get_type(POINTER);
          break;
       case PARAMDECL_1:
          n->type = n->child[0]->type;
          break;
       case THINGY:
          n->type = n->child[0]->type;
          break;
       case PARAMDECL_2:
          n->type = clone_type(n->child[1]->type);
          n->type->u.p.elemtype = n->child[0]->type;
          break;
       case PARAMDECLLIST_2:
          n->type = get_type(TUPLE);
          n->type->u.t.nelems = 1;
          n->type->u.t.elems = calloc(1, sizeof(struct typeinfo *));
          n->type->u.t.elems[0] = n->child[0]->type;
          break;
       case PARAMDECLLIST_1:
          n->type = get_type(TUPLE)
    
          /* consider whether left child, guaranteed to be a PARAMDECLLIST,
             is guaranteed to be a tuple.  Maybe its not. */
          n->type->u.t.nelems = n->child[0]->type->u.t.nelems + 1;
          n->type->u.t.elems = calloc(n->type->u.t.nelems,
    				      sizeof(struct typeinfo *));
          for(i=0;i < n->child[0]->type->u.t.nelems; i++)
             n->type->u.t.elems[i] = n->child[0]->type->u.t.elems[i];
          n->type->u.t.elems[i] = n->child[1]->type;
    
          break;
       case INITIALIZER_DECL:
          n->type = get_type(FUNC)
          n->type->u.f.returntype = get_type(DONT_KNOW);
          n->type->u.f.params = n->child[1].type;
          break;
       case SIMPLE_DECLARATION_1:
          n->type = clone_type(n->child[1]->type);
          n->type->u.f.returntype = n->child[0]->type;
       }
    }
    

    Mailbag

    At this point in order to implement proper typechecking I need to implement expression type evaluation, which would come from starting to append information about expressions to the tree.
    Yup. You are supposed to stick all the type information for all the symbols, somehow. Then you are to write a tree traversal that walks the tree and for the code bodies, do a bottom-up post-order traversal that assigns a .type value to each tree node until you hit things that logically have no value (and hence no type). The sequence of nodes visited and .type's assigned includes:
      x (look up, find INT_TYPE in symbol table)
      1 (INT_TYPE is a lexical attribute or property of this constant)
      add_expr (one applies typecheck rules, verifies that +
         can add the two children types, and computes result type INT_TYPE)
      return_stmt (check type of return matches type of function)
    
    Can you give me a more concrete example of how type information is produced and placed in symbol table entries?
    Let's look at
    var x : Int
    It takes some digging to even figure out where in the grammar this is. YOUR grammar by default would have it differently than mine, but in some grammar it might be:
                  __asgn___
                 /         \
               pat    maybe_ty_ascription
                /           /    \
             NAME        ':'      ty
                 x                  \
                                    NAME
                                        Int
    
    To place type INT_TYPE as the type of x, you would
    1. do a pass in which types are synthesized up (type of Int is INT_TYPE, type of ty is INT_TYPE, type of maybe_ty_ascription is INT_TYPE, type of asgn is INT_TYPE), and then
    2. do a tree traversal in which symbol tables are populated and at the pat or maybe the NAME that would insert x with type INT_TYPE.

    From Types to Type Checking

    To summarize type checking:

    Syntax Trees to Type Representation Example

    Let's consider an example parse tree for a function header. In C it would be
    int f(int x, int y, float z)
    
    We are studying tree shapes in order to write traversal functions. We might broach questions such as:

    Where we are at

    Constructing Type Representation for a Function (cont'd)

    The function:

    int f(int x, int y, float z);

    The syntax tree:

    The type struct we are supposed to fill out from this tree:
    C type where in the tree?
    typedef struct typeinfo {
       int basetype;
       union { /* ... along with other union members */
          struct funcinfo {
    	 char *name; /* ? */
    	 struct sym_table *st;
    	 struct typeinfo *returntype;
    	 int nparams;
    	 struct param *parameters;
    	 }f;
       }u;
    }
    
    FUNC_TYPE
    
    
    n->kids[1]->kids[0]->leaf->text
    new_st(...);
    n->kids[0]->type
    nparams_helper(n->kids[1]->kids[1])
    params_builder(n->kids[1]->kids[1])
    
    
    
    

    Parameter "list" vs. "tuple"

    In-Class Exercise: helper Functions to construct param typeinfo

    calc_nparams

    Practice helper function, just count how many parameters there are
    int calc_nparams(struct tree *n)
    {
       if (n->prodrule == PD) {
          return 1;
          }
       else if (n->prodrule == PL) {/* recursion/induction step */
          return 1 + calc_nparams(n->kids[0]);
          }
       return 0;
    }
    

    Old Mailbag

    What is the difference between a function declaration and a variable declaration, when it comes to adding the symbols to the table? as far as the tree is concerned they are almost exactly the same, with the exception of which parent node you had. Is there (or should there be) a line in the symbol entry which states the entry as a function vs a variable?
    You add the symbols to the same table. For HW#4 they are thus treated basically identically. For HW#5 you put in different type information for functions (whose basetype says they are a function, and whose typeinfo includes their parameters and return type) than for simple variables.
    I have code written which (hopefully) creates the symbol table entry for variables. This code uses a function which spins down through non-terminals to get the identifier. Can I use this same function to get the identifier for a function? A function is
    direct_function_declarator: direct_declarator LP ... RP ...
    so after the direct_declarator it has other useful things that I'm not sure need to be in the symbol table entry.
    You can sometimes re-use tree traversal functions that work through similar subtrees, either as-is (if the subtrees really use the same parts of the grammar) or by slightly generalizing or making generic the key decisions about what to do based on production rule. For example, you might add a flag parameter to a function that spins through nonterminals, indicating whether this was in a function declaration or not; that might allow you to tweak the tree traversal to adjust for minor differences. Or maybe you just add cases to the switch statement to account for additional production rules in the various similar kinds of subtrees.

    mk_params

    Helper function: construct type representation for parameters. After agonizing a bit over the "linked list" versus "tuple" question...
    struct param * mk_nparams(struct tree *n)
    {
       if (n->prodrule == PD) {      /* basis case */
          struct param *p = malloc(sizeof (struct param)); /* check for NULL */
    
          p->type = n->kids[0]->type; /* if we synthesized it already */
    /*or*/
          if (n->kids[0]->prodrule == VOID)
             p->type = void_typeptr; /* ...and more like this, INT etc. */
          else
             p->type = synthesize_type_from_declspecs(n->kids[0]);
    
          if (n->kids[1]->prodrule == IDENTIFIER)
             p->name = n->kids[1]->leaf->text;
          else { /* non-simple declarator, pointer or something */
             p->name = get_name_from_declarator(n->kids[1]);
             /* type is modified by declarator; it's a pointer or something */
             p->type = inherit_type_into_declarator(p->type, n->kids[1]);
             }
     
          p->next = NULL;
          return p;
          }
       else if (n->prodrule == PL) { /* recursion/induction step */
          struct param *p1 = mk_nparams(n->kids[0]);
          struct param *p2 = mk_nparams(n->kids[1]);
          struct param *p3 = p1;
          while (p3->next != NULL) p3 = p3->next;
          p3->next = p2;
          return p1;
          }
       return NULL;
    }
    

    Recursing through Trees Built from some weird grammar that was not cgram.y

    codetreecomments and/or symbol table
    // empty file
    
    n/a is empty file a syntax error? or fine? In any case, create empty symbol table
    (nothing to insert)
    int x;
    
              FILE
                   \
                   XDCLLIST_2
             	|      \
                    ε     COMMON_DCL
                            /     \
                          LVAR   VARDCL
                                 /     \
                               INT    IDENT
                              "int"    "x"
    
    Construct a type from INT "int" because it is known type basis case.
    Synthesize it up to VARDCL. Inherit it down into declarator list (including x) Insert "x" into current (global) symbol table because it is VARDCL kid #2.
    int main() { }
    
              FILE
                   \
                   XDCLLIST_2
             	|      \
                    ε     XFNDCL
                         /  |   \
                    FNRES FNDCL  FNBODY_2
                   /     /  |  \       \
                INT  IDENT ATL        STMTLIST
              "int" "main"  |    |       |
                            ε    ε       ε
    
    Construct a FUNC type from FNDCL
    • Construct an TUPLE of length 0 from empty ATL
    • Construct a INT type from FNRES
    Insert "main" into global symbol table
    Create a local symbol table
    Insert parameters into local symbol table
    Insert local variables into local symbol table

    Mailbag

    My code segfaults! Help!
    I used to debug my segfaults with gdb. Then I found that valgrind was more effective, including cases gdb missed, and saves me a lot of time. Of course, the real secret is building a mental model (literally, pictures in your head) of the memory, and getting to the point where that mental model somewhat corresponds to reality.
    In some example code on the class site, you have the type checking done at the same time as the symbol table entry. Is there any reason not to break these out into 2 separate functions?
    No, no reason at all. In the old days there were reasons.
    Is it normal to feel like my code for adding to and checking the symbol tables is messy, gross, and more hard-coded than I'd like?
    The Flex and Bison homeworks had an unfair advantage: they were using a declarative language. Later homeworks will be messy and gross by comparison, because from here on out we are using the imperative (and OO, which is "imperative with style") paradigm. Walking trees and getting the details all in there will require a lot of code. How gross it is, "Zis is all up to you" (from a Geronimo Stilton book).
    Does our language really require comma-separated lists of variables in declarations? It would be so much easier if it only did one variable per declaration.
    Don't exaggerate. It would not be that much easier. This is now a j0 Level Two thing and Level One folks are off the hook, But (even for Level Two/Three folks) it is FINE to start by just getting it working for one-variable declarations, then detect and handle two-variable declarations as a special case, then generalize to 3+ variables. You might pass some test cases with only one-variable declarations working.

    Connecting Trees to Traversals

    /*
     * Semantic analysis from syntax tree, Go (subset) Edition
     */
    void semantic_anal(struct symtab *current_st, nodeptr n)
    {
       if (n==NULL) return;
       for(int i = 0; i < n->nkids; i++) semantic_anal(current_st, n->child[i]);
       switch (n->prodrule) {
       case XFNDCL : /* whole function */
          n->symtab = n->child[1]->symtab;
          populate_locals(n->symtab, n->child[2]);
          /*
           * visit body to check for undeclared/redeclared
           */
          check_variable_uses(n->child[2]);
          break;
       case FNDCL :  /* function header */
          char *name = get_func_name(n->child[0]);
          n->symtab = mk_symtab();
    
          n->type = get_type(FUNC);
          n->type->u.f.returntype = n->child[4]->type;
          n->type->u.f.params = n->child[2]->type;
          n->type->u.f.symtab = n->symtab;
          st_insert(current_st, name, n->type);
    
          populate_params(n->symtab, n->child[2]);
          break;
       }
    }
    

    Discussion of Tree Traversals that perform Semantic Tests

    This example illustrates just one of the myriad-of-specialty-traversal-functions that might be used. This mindset is one way to implement semantic analysis.

    Suppose we have a grammar rule

    AssignStmt : Var EQU Expr
    
    We want to detect if a variable has not been initialized, before it is used. We can add a boolean field to the symbol table entry, and set it if we see, during a tree traversal, an initialization of that variable. What are the limitations or flaws in this approach?

    We can write traversals of the whole tree after all parsing is completed, but for some semantic rules, another option is to extend the C semantic action for that rule with extra code after building our parse tree node:

    AssignExpr : LorExpr '=' AssignExpr { $$ = alctree(..., $1, $2, $3);
    	lvalue($1);
    	rvalue($3);
    	}
    
    void lvalue(struct tree *t)
    {
       if (t->label == IDENT) {
          struct symtabentry *ste = lookup(t->u.token.name);
          ste->lvalue = 1;
       }
       for (i=0; i<t->nkids; i++) {
          lvalue(t->child[i]);
          }
    }
    void rvalue(struct tree *t)
    {
       if (t->label == IDENT) {
          struct symtabentry *ste = lookup(t->u.token.name);
          if (ste->lvalue == 0) warn("possible use before assignment");
       }
       for (i=0; i<t->nkids; i++) {
          rvalue(t->child[i]);
          }
    }
    

    lecture #17 began here

    Midterm Distribution

      294
      273 276
    __265_268_________
      256 256 257
      242 242 245
    __234_236_236_238_
      220 224
      214 215
    __194_198_________
      185
      178
      170 171 174
    __163_____________
      148 151
      140 141 145
      75
    

    Comments on Midterm Exam Solutions

    Real Life vs. toy lvalue/rvalue example

    This example illustrated walking through subtrees looking for specific nodes where some information was inserted into the tree. In real life... For example, if the program starts by calling a subroutine at the bottom of code which initializes all the variables, the flow graph will not be fooled into generating warnings like you would if you just started at the top of the code and checked whether for each variable, assignments appear earlier in the source code than the uses of that variable.

    (x, y, z int) vs. var a, b, c int

    In a parameter list In a variable declaration
             ATL
           /  |  \
        ATL   ,    ARGTYPE
       / | \        |     \
    ATL  , ARGTYPE  LNAME  LNAME
     |       |	 "z"   "int"
    ARGTYPE LNAME
     |       "y"
    LNAME
     "x"
    
         COMMONDCL
        /       \
      LVAR    VARDCL
              /    \
           DNL     NTYPE
           /|\       \
        DNL , LNAME  LNAME
        /|\    "c"   "int"
     DNL , LNAME
      |     "b"
    LNAME
     "a"
    
    OK, how do we get type information down to the tree nodes where "x"? Specialized subtraversals and/or multiple passes. This sample is probably a duplicate of some earlier sample code, just tied to the non-terminal names of the go.y grammar a bit.
    void populate(struct tree *n, struct symtab *st)
    {   int i;
        if (n==NULL) return;
        for(i=0; i<n->nkids; i++)
           populate(n->kids[i], st);
        switch (n->prodrule) {
        case VARDCL:
           n->type = n->kid[1].type;         // synthesiz
           n->kid[0].type = n->type;         // inherit
           insert_w_typeinfo(n->kid[0], st);
           break;
        case NTYPE:
           n->type = n->kid[0].type;
           break;
        case LNAME:
           if (!strcmp(n->token->text, "int"))
              n->type = T_INTEGER;
           break;
        case ARG_TYPE_LIST_1: /* ATL: arg_type */
           break;
        case ARG_TYPE_LIST_2: /* ATL: ATL ',' arg_type */
           break;
        case ARG_TYPE_1: /* AT: name_or_type */
           break
        case ARG_TYPE_2: /* AT: sym name_or_type */
           break
        case ARG_TYPE_3: /* AT: sym dotdotdot */
           break
        case ARG_TYPE_4: /* AT: dotdotdot */
           break
        }
    }
    
    /*
     * "inherited attribute" for type could go down by copying from
     * parent node to child nodes, or by passing a parameter. Which is better?
     */
    void insert_w_typeinfo(struct tree *n, struct symtab *st)
    { int i;
      if (n == NULL) return;
      for(i=0; i<n->nkids; i++) {
         if (n->kids[i]) {
            n->kids[i]->type = n->type;
            insert_w_typeinfo(n->kids[i], st);
    	}
         }
      switch (n->prodrule) {
      case DNL: /* ?? nothing needed */
        break;
      case LNAME:
        st_insert(st, n->token->text, n->type);
        break;
      }
    }
    

    Type Checking

    Type Systems

    Type systems are a "CS theory" topic, semi-redundant to what we talked about when we introduced struct typeinfo. The point then and now is that you need to add whatever is needed to represent the types that occur in the language for which you are writing a compiler. A type system might start with rules like: In addition, a type system includes rules for assigning these types to the various parts of the program; usually this will be performed using attributes assigned to grammar symbols.

    Example Semantic Rules for Type Checking

    grammar rule semantic rule
    E1 : E2 PLUS E3 E1.type = check_types(PLUS, E2.type, E3.type)
    Where check_types() returns a (struct type *) value. One of the values it can return is TypeError. The operator (PLUS, probably the binary addition operator) is passed in to the check types function because typecheck rules depend on the operator -- the result type for array subscripting works different than the result type for the arithmetic operators, which may work different (in some languages) than the result type for logical operators that return booleans.

    In-class brainstorming

    Consider the class project. What else will we need to check during semantic analysis, and specifically during type checking?
    What should the type rules for k0's PLUS be?
    The following table is probably WRONG for Kotlin since it came from some other language. What should k0's answers be? In the table below, blank entries are compile time errors. Are there any uses of PLUS that need runtime checks?

    row PLUS colIntDoubleStringBooleanArrayfunc
    Int Int Double
    DoubleDoubleDouble
    String String String String
    Boolean Boolean
    Array Array
    func

    lecture #18 began here

    Problems with weak Semi-colon Insertion

    A student recently asked whether it was acceptable for semi-colon insertion, if you use that strategy, to "break" useful Kotlin code like
      f(
      x
      ,
      y
      )
    
    or
      import foo
         .bar
         .baz
    
    In either case, semi-colons will be inserted after identifiers, probably causing syntax errors. Here is an approach that would fix it.

    Semi-Colon Insertion in Icon and Unicon

    This technique appeared in the Icon programming language[Griswold et al] around 1980. Algorithm Pseudo-Code:
    1. Add a static array with two booleans for each token category:
    2. Remember previous token's Ender status, update each call to yylex()
    3. Set a flag every newline to mark the next token First_on_a_line; update each call to yylex(); unset this flag after each token.
    4. Algorithm for new yylex() shim function in between yyparse() and old yylex:
       
         IF we saved the last token (i.e., returned Semi-Colon last time) THEN
            return saved token
         ELSE {
            call old yylex()
            IF First_on_a_line AND Beginner AND last.Ender THEN {
                save_token; return Semi-Colon
                }
            ELSE return token
            }
      

    Typechecking, cont'd

    Codewise, all this might look like:
    struct typeinfo *check_types(int operator,
                                 struct typeinfo *e1,
                                 struct typeinfo *e2) {
       switch(operator) {
          case PLUS: {
             /* implement row PLUS col table here */
             break;
             }
       }
    }
    
    What other type-check rules for k0 can we derive?

    Let's look at a few lines of a typecheck example.

    In the break between lecture and lab, a student reported that I have previously stated that redeclarations at the same level are OK so long as they are the same type. I believe I said that. Now: should that be true for PunY???

    Type Checking Example

    Work through a type checking example for the function call to foo() shown in bold in the example below. This is a C language example. The Kotlin would be pretty similar; how much would be different?

    C Kotlin
    int foo(int x, char *y) {
       return x;
       }
    
    int main()
    {
       int z;
       z = foo(5, "funf");
       return 0;
    }
    
    fun foo(x : Int, y : String) : Int {
       return x
       }
    
    fun main() : Int
    {
       var z : Int
       z = foo(5, "funf")
       return 0
    }
    

    After parsing, the symbol table (left) and syntax tree for the call (right) looks like:

    The typecheck of this tree proceeds as a post-fix traversal of the tree. Type information starts from leaves, which either know their type if they have one (constants) or look up their type in the symbol table (identifiers). Can you hand-simulate this in the correct order, filling in .type fields for each tree node?

    Mailbag

    I was wondering if names that are declared as function should be treated similarly to other names. If I define a function with the name "stuff", can I, in the same scope, declare "stuff" as a variable, or is that illegal? What about the inverse, trying to declare a function with the same name as an existing variable?
    Follow the rules defined for your language. In at least one major language (do you know which?), function names tracked differently in the symbol table from variables: there is a function slot and a variable slot for each name. But in most languages, the function stuff() should get inserted into the module/file/global scope. Declaring a variable of the same name is an error in that scope, but just fine in a local scope.

    Type Promotion, Type Equivalence, and Automatic Conversion

    When is it legal to perform an assignment x = y?

    Type Promotion / C Example:
    Implicit Explicit
    int x;
    long y;
    y = y + x;
    
    int x;
    long y;
    y = y + (long)x;
    

    Type Conversion / int to float

    In some languages, it is not about different sizes of integers. What about floats/doubles vs. ints? What happens? Could PunY users ever write code like this?
    Implicit Explicit
    x : int
    y : float
    y = y + x
    
    x : int
    y : float
    y = y + float(x)
    

    This may come about in real life in parameters and return values, such as an int parameter passed into a function that requires a float.

    int x, y;
        ...
    y = sqrt(y*y + x*x);
    

    Name equivalence vs. structure equivalence

    For classes/records/structures, some languages use name equivalence, while others use structure equivalence.
    Features like typedef complicate matters.
    If you have a new type name MY_INT that is defined to be an int, is it compatible to pass as a parameter to a function that expects regular int's? In C, the answer is: yes.
    Object-oriented languages also get interesting. Do OOP languages use name equivalence or structure equivalence?
    Name+inheritance. Subclasses usually are allowed anyplace their superclass would be allowed.

    Implementing Structs

    Some years, CSE 423 implements classes, not structs. But some languages such as C or Go have structs. And structs seem to be a preliminary step towards classes. How much of the following "structs" discussion is similar to what classes would need, and how much needs to be different?
    1. In C/C++, storing and retrieving structs by their label -- the struct label is how structs are identified. In Go/VGo, the reserved word type is more like a C/C++ typedef. If you were doing C/C++-style struct labels, the labels can be keys in a separate hash table, similar to the global symbol table. You can put them in the global symbol table so long as you can tell the difference between them and regular symbol names, for example by storing them as "struct foo" (not a legal name) instead of just storing them as "foo".
    2. You have to store fieldnames and their types, from where the struct is declared. This is conceptually a new local scope. In a production language you would use a hash table for each struct, but in CSE 423 a link list would be OK as an alternative. Then again, if you've built a hash table data type for the global and local symbol tables, why not just use it for struct field scopes?
    3. You have to use the struct's type information to check the validity of each dot operator like in rec.foo. To do this you'll have to lookup rec in the symbol table, where you store rec's type. rec's type must be a struct type for the dot to be legal, and that struct type should include the hash table or link list that gives the names and types of the fields -- where you can lookup the name foo to find its type.

    "Old Mailbag"

    I am having a problem with a segmentation fault. Can you help?
    When sticking print statements into your program no longer cuts it, your two best hopes are GDB and Valgrind. Yes, I can usually help.
    The problem is, there is just so much to type check (I mean, literally everything has a type!); can you suggest any ways to go about this in a quicker manner, or anything that could be pruned/ignored?

    Need More Help with Type Checking?

    struct typeinfo with struct params

    Note that if you stick parameter lists in the type representation, this means your struct typeinfo holds some things that are not C types in and of themselves, but are used in larger composite types. I guess a TUPLE was a similar pseudo-type option.

    Type Checking Function Calls

    void typecheck(nodeptr n)
    {
       if (n==NULL) return;
       for(int i; i < n->nkids; i++) typecheck(n->child[i]);
       switch(n->prodrule) {
       ...
       case POSTFIX_EXPRESSION_3: {
          n->type = check_types(FUNCALL, n->child[0]->type, n->child[2]->type);
          }
       }
    }
    
    ...
    
    typeptr check_types(int operand, typeptr x, typeptr y)
    {
       switch (operand) {
       case  FUNCALL: {
          if (x->basetype != FUNC)
             return type_error("function expected", x);
          if (y->basetype != TUPLE)
             return type_error("tuple expected", y);
          if (x->u.f.nparams != y->u.t.nelems)
             return type_error("wrong number of parameters", y);
    
          /*
           * for-loop, compare types of arguments
           */
          for(int i = 0; i < x->u.f.nparams; i++)
             if (check_types(PARAM, x->u.f.params[i], y->u.t.elems[i]) ==
    	     TYPE_ERROR) return TYPE_ERROR;
          /*
           * If the call is OK, our type is the function return type.
           */
          return x->u.f.returntype;
          break;
          }
       }
    }
    

    Semantic Analysis for Structs and Classes

    What work is performed during the semantic analysis phase, to support structs and classes?

    How to TypeCheck Square Brackets

    This is about the grammar production whose right-hand side is:
    postfix_expression LB expression RB
    
    or its equivalent in your grammar. In mainstream languages this syntax may be reserved for an array type, but in a language like Python it might be for lists, strings, dictionaries, etc.
    1. recursively typecheck $1 and $3 ... compute/synthesize their .type fields.
    2. What type(s) does $1 have to be? LIST/ARRAY, or STRING (or DICT/TABLE, if a table type exists)
    3. What type(s) does $3 have to be? INTEGER (or e.g. STRING/ARRAY OF CHAR, for tables with those keys)
    4. What is the result type we assign to $$? Lookup the element type from $1, possibly Any
    Pseudo-code fragment. Goal is to find errors and determine n's type.
    int typecheck_squarebrackets(struct tree *n)
    {
       struct tree *n1 = n->child[0];
       struct tree *n3 = n->child[2];
       /*
        * Typecheck children. This is a mutual recursion.
        */
       if (typecheck(n1) == TYPE_ERROR ||
           typecheck(n3) == TYPE_ERROR) {
          n->type = TYPE_ERROR;
          return TYPE_ERROR;
          }
    
       /*
        * Given the children's types, see whether n1[n3] is legal.
        * For PunY, you need to allow for Any at various points.
        */
       switch (n1->type->basetype) {
       case STRING:
          /* ... insert string typecheck code here */
          /* check if n3's type is integer */
          if (n3->type->basetype != BT_INTEGER) {
              bad_type("list must be subscripted with integers", n3);
    	  return TYPE_ERROR;
              }
          /* a subscript of a string is still a string, right? */
          n->type = n1->type;
          break;
       case LIST:
          /* check if n3's type is integer */
          if (n3->type->basetype != BT_INTEGER) {
              bad_type("list must be subscripted with integers", n3);
    	  return TYPE_ERROR;
              }
          /* assign n's type to be n1's element type */
          n->type = n1->type->u.l.elemtype;
          break;
       case DICT:
          /* check if n3's type is n1's index type */
          if (n3->type->basetype != n1->type->u.t.indextype->basetype) {
              bad_type("table must be subscripted with its declared index type", n3);
    	  return TYPE_ERROR;
             }
          /* assign n's type to be n1's element type */
          n->type = n1->type->u.t.elemtype;
          break;
       default:
          bad_type("list or table expected in [] operation", n1);
          /* what does n's type field hold, then */
          n->type = /* ?? */
          return TYPE_ERROR;
          }
    





    Did we get something like:
          if (n3->type->basetype != INTEGER) {
             bad_type("index must be integer in [] operation", n3);
             }
          n->type = n1->type->elemtype;
    
    and
          if (n3->type->basetype != n1->type->indextype) {
             bad_type("index type must be compatible in [] operation", n3);
             }
          n->type = n1->type->elemtype;
    

    What other type checking examples should we be doing?

    So far in lecture we have seen possibly-too-handwavy examples of What else would help you wrap your brains around type checking?

    Mailbag

    I have homework due in other classes. Can we have an extension?
    The bad news is: we don't have room in the semester for much extension. The good news is: by this point, the late fees are very low.
    I'm still working on hw4 since I can't get my assignment 4 working clean enough for hw5. What should I do? I do not wanna fail this class and its not like I haven't been putting lots of effort into it.
    You are graded relative to your peers, and many of your peers are also far behind. If you have been putting in appropriate effort, and you are getting instructor help when you get stuck, you should be able to make progress on your HW. If you stay in, and give it your best shot, odds are good that relative to your peers, you will be OK.
    When int and double mix, C-based languages promote int to double. Do we ever have to demote from double to int?
    What should happen if the program contains:
       i : int
       r : float
       r = 27.0
       i = 3.1415 * r * r # implicit demotion to int
       i = int(3.1415 * r * r) # explicit demotion to int
    

    "Drop day" commentary

    Thoughts on HW

    lecture #19 began here

    Another Typecheck Example

    Consider the following small program.
    C Kotlin
    void write(int);
    int fib(int);
    void main()
    {
       int i;
       i = 0;
       while (i < 5) {
          write(fib(i));
          i += 1;
          }
    }
    int fib(int n)
    {
      if (n < 3) { return 1; }
      return fib(n-2) + fib(n-1);
    }
    
    
    
    
    fun fib(n : Int) : Int {
      if (n < 3) { return 1 }
      return fib(n-2) + fib(n-1)
    }
    fun main() {
       var i : Int = 1
       while (i < 7) {
          println(fib(i))
          i += 1
          }
    }
    

    See below for a syntax tree

    From Type Checking on to Code Generation

    Type Checking Fibonacci Example, Cont'd

    As a reminder, our code looked like this:

    void write(int);
    int fib(int);
    void main()
    {
       int i;
       i = 0;
       while (i < 5) {
          write(fib(i));
          i += 1;
          }
    }
    int fib(int n)
    {
      if (n < 3) { return 1; }
      return fib(n-2) + fib(n-1);
    }
    

    The full syntax tree looks like the following. Still barely legible on my office monitor

    Now, let's crank the tree traversal for the main typecheck of the executable code.

    lecture #20 began here

    Mailbag

    My team did not get the grade we deserved on HW#4
    I want to award points for implemented and demonstrable functionality. Feel free to make a show-and-tell appointment, or please point me at test cases that show functionality, and I will adjust grades accordingly.
    I have been having a bug... I am at the point that I'm thinking I should start over with my code entirely from scratch and try to rebuild it. But if I do this I really have no hope of finishing HW5 which I am assuming will carry on to HW6. I recently found an error with some of the stuff I wrote in HW3 which was causing problems with my HW4... I am wondering if there is anyways to have stuff that we are currently working on now graded separately. Otherwise I either will have to rewrite my entire code again being late on everything. I would rather not do this as I did it in HW3 when my prior code was not working. But otherwise all of my homeworks moving forward is not going to run properly because of errors in prior homework
    Yup, the later homeworks depend upon earlier ones. You should not try and go on to a later homework if an earlier one has not been completed to a passing level. If an earlier homework was not at a passing level, you should fix and resubmit, and you are encouraged to seek help if that is needed.

    Run-time Environments

    How does a compiler (or a linker) compute the addresses for the various instructions and references to data that appear in the program source code?
    To generate code for it, the compiler has to "lay out" the data as it will be used at runtime, deciding how big things are, and where they will go.

    Scopes and Bindings

    This last item is generally done via a symbol table. In most compiled languages it happens at compile time (in contrast, for example ,with LISP).

    Environment and State

    Environment maps source code names onto storage addresses (at compile time), while state maps storage addresses into values (at runtime). Environment relies on binding rules and is used in code generation; state operations are loads/stores into memory, as well as allocations and deallocations. Environment is concerned with scope rules, state is concerned with things like the lifetimes of variables.

    Runtime Memory Regions

    Operating systems vary in terms of how the organize program memory for runtime execution, but a typical scheme looks like this:
    code
    static data
    stack (grows down)


    big hole


    heap (may grow up, from bottom of address space)

    HW#6

    Questions to ask about a language, before writing its code generator

    May procedures be recursive?
    All but toy languages require this and it entails a stack region
    What happens to locals when a procedure returns?
    Lazy deallocation would be exceedingly rare
    May a procedure refer to non-local, non-global names?
    nested procedures? dynamic scoping? implicit object/struct field names?
    How are parameters passed?
    Many styles possible. Different declarations for each? Rules hardwired by type?
    May functions be passed as parameters?
    Not too awful. Function pointers are popular with both the functional programming and operating systems communities!
    May procedures be return values?
    Adds complexity for non-local names
    May storage be allocated dynamically (like, on the heap)
    All modern languages require this, but some languages do it with syntax (new) others with library (malloc))
    Is storage deallocated explicitly
    Do we have a garbage collector? Does that affect code generation?

    "Modern" Runtime Systems

    Modern languages' runtime systems have extra properties compared with that of a traditional language like C. Here are a few examples.

    A "self" or "this" in non-static method calls.
    Possibly implemented via a dedicated register, or an implicit, extra parameter. Either way, OO slightly alters the activation record.
    Garbage collection.
    Automatic (heap) storage management is one of the most important features that makes programming easier. The Basic problem in garbage collection is: given a piece of memory, are there any pointers to it? (And if so, where exactly are all of them please). Approaches: There is a fine paper presenting a "unified theory of garbage collection"
    Reflection.
    Objects can describe themselves, via a set of member functions. This plays a central role in Visual GUI builders, IDE's, component architectures and other uses.
    Serialization
    Objects can convert into a string of bits and vice-versa. If you can do this, you can store on disk, or move across a network to another machine. Its easy to move data-only objects, but difficult to move other OS resources besides memory.
    Just-in-time compilation.
    A virtual machine ("byte code") execution model...can be augmented by a compiler built-in to the VM that converts VM instructions to native code for frequently executed methods or code blocks.
    Security model.
    Modern languages may attempt to guarantee certain security properties, or prevent certain kinds of attacks.
    For what its worth, goal-directed programs in languages such as Unicon have an activation tree each instant, due to suspended activations that may be resumed for additional results. The lifetime view is a sort of multidimensional tree, with three types of nodes.

    Mailbag

    So, really, how would/do I handle classes (or in some languages: structs)?
    Well... to answer this question:
    1. In k0, we are only supporting a tiny # of 107-required "built-in to k0" class types. Maybe just strings? Anything else?
    2. there should be a basetype (integer code) for class (or struct) in your type representation.
    3. In k0 all variables of a declared class type start with a initializer or a null value (ensure no uninitialized data).
    4. the same class/structure name/label cannot be declared twice. Pre-Insert the classname as a global symbol table entry for built-in class types.
    5. each unique class type should be able to tell you its fields' and methods' names and types.
    6. each unique class type should be able to lookup a field/method by name and tell you its type. i.e. classes define a local scope
    7. Variables declared to be of a class type can be assigned from any value of that type, and not any other type.
    8. The same exact-match type check semantics are applied e.g. to passing a class instance as a parameter, it is like an assignment
    9. Do objects need to support any other operators besides assignment and being passed as a parameter? Method calls, for a limited number of methods.
    10. For languages with structs only:
      • check your language. In many/most languages struct values cannot be assigned to other named struct types, even if they are "equivalent" in size and shape
      • while you cannot declare a struct variable for an undeclared struct type, you can declare a pointer to an unknown struct type.
      • if we were doing typedefs we would have extra complications like distinguishing a variable of a struct type from a typedef name for a struct type.

    Activation Records

    Activation records organize the stack, one record per method/function call.


    earlier
    activation
    record

    earlier
    activation
    record
    return value
    parameter
           ⋮
    parameter
    previous frame pointer (FP)
    saved registers
           ⋮
    %rbp →saved PC
    local
        ⋮    
    local
    temporaries
    %rsp →        ⋮
    current
    activation
    record
    "top" of stack

    grows down by subtracting from %rsp
    calls create new
    activation
    records here

    At any given instant, the live activation records form a chain and follow a stack (push/pop) discipline for allocation/deallocation. Since each activation record contains a pointer to the previous one, it is really pretty much a linked list we are talking about, with a base pointer register holding the pointer to the top of the stack.

    Over the lifetime of the program, all these activation records, if saved, would form a gigantic tree. If you remember all prior execution up to a current point, you have a big tree in which its rightmost edge are live activation records, and the non-rightmost tree nodes are an execution history of prior calls. (Program Monitoring and Visualization could allow us to depict and inspect this history tree.)

    Variable Allocation and Access Issues

    Given a variable name, how do we compute its address?
    globals
    easy, symbol table lookup... once we have figured out how to allocate addresses in a process that does not exist yet.
    locals
    easy, symbol table gives offset in (current) activation record
    objects
    Is it "easy"? If no virtual semantics*, symbol table gives offset in object, activation record has pointer to current object in a standard location. (This is the reason C++ does not use virtual semantics by default.)
    For virtual semantics, generate code to look up offset in a table at runtime, based on the current object's type/class.
    locals in some enclosing block/method/procedure
    ugh. Pascal, Ada, and friends offer their own unique kind of pain. Q: does the current block support recursion? Example: for procedures the answer would be yes; for nested { { } } blocks in C the answer would be no.

    *What are "Virtual" Semantics?

    C++ is (just about) the only major object-oriented language that has to compete with C in the performance arena. For this reason, it chose early on to be different than every other OO language. By default, if you are working on a class Foo object, you can find Foo's member variables and call Foo's methods by compile-time-determinable memory offsets and addresses. So a class is basically no worse than a struct to generate code for.

    If you say the keyword "virtual" in C++, or if you use just about any other OOP language, subclassing and interfacing semantics mean that the address referred to by o.x or o.m() has to be calculated at runtime by looking up o's actual class, using runtime type information.

    Sizing up your Regions and Activation Records

    Add a size field to every symbol table entry. Many types are not required for your project but we might want to discuss them anyhow.

    You do this sizing up once for each scope. The size of each scope is the sum of the sizes of symbols in its symbol table.

    Break Questions

    Jackson's Question: what about x = if ...
    Kotlin is more expression-oriented than C. Things like if's and while's produce values that can be used in surrounding expressions. We are mostly not bothering with this "parlor trick"...but are there any gotcha's?
    Hunter's question: what about dot? Do we have anything
    I thought we'd need dot for some built-ins, but maybe not.

    Intermediate Code Generation

    Goal: list of machine-independent instructions for each procedure/method in the program. Basic data layout of all variables.

    Can be formulated as syntax-directed translation

    How does the compiler talk at compile-time about addresses that will exist at runtime?

    Mailbag

    These HW are big. Is it possible to reintegrate group work?
    Yes. But you all are responsible for understanding how your compiler works. See HW #6.

    Functions for Intermediate Code Generation

    These helper functions might really be best described as returning "addresses". Intermediate code addresses are described down below.

    Semantic Rules for Intermediate Code Generation

    Code for evaluating traditional expressions can be synthesized via bottom-up traversal. If the location where an expression's computed result can be read is called addr and the sequence of intermediate code instructions that compute that value is called icode, you can postulate the following semantic rules. This is a simplifying abstraction of what would be needed for an entire modern programming language expression grammar, and your non-terminal names might be different. Operator ||| is a list concatenation.

    ProductionSemantic Rules
    Assignment : IDENT '=' AddExpr
    Assignment.addr = IDENT.addr
    Assignment.icode = AddExpr.icode |||
                       gen(ASN, IDENT.addr, AddExpr.addr)
    AddExpr : AddExpr1 '+' MulExpr
    AddExpr.addr = genlocal()
    AddExpr.icode = AddExpr1.icode ||| MulExpr.icode |||
                    gen(ADD,AddExpr.addr,AddExpr1.addr,MulExpr.addr)
    AddExpr : AddExpr1 '-' MulExpr
    AddExpr.addr = genlocal()
    AddExpr.icode = AddExpr1.icode ||| MulExpr.icode |||
                    gen(SUB,AddExpr.addr,AddExpr1.addr,MulExpr.addr)
    MulExpr : MulExpr1 '*' UnaryExpr
    MulExpr.addr = genlocal()
    MulExpr.icode = MulExpr1.icode ||| UnaryExpr.icode |||
                    gen(MUL,MulExpr.addr,MulExpr1.addr,UnaryExpr.addr)
    MulExpr : MulExpr1 '/' UnaryExpr
    MulExpr.addr = genlocal()
    MulExpr.icode = MulExpr1.icode ||| UnaryExpr.icode |||
                    gen(DIV,MulExpr.addr,MulExpr1.addr,UnaryExpr.addr)
    UnaryExpr : '-' UnaryExpr1
    UnaryExpr.addr = genlocal()
    UnaryExpr.icode = UnaryExpr1.icode |||
                      gen(NEG,UnaryExpr.addr,UnaryExpr1.addr)
    UnaryExpr : '(' AddExpr ')'
    UnaryExpr.addr = AddExpr.addr
    UnaryExpr.icode = AddExpr.icode
    UnaryExpr : IDENT
    UnaryExpr.addr = IDENT.addr
    UnaryExpr.icode = emptylist()

    lecture #21 began here

    Three-Address Code

    Basic idea: break down source language expressions into simple pieces that: You can literally just make up this intermediate code file format. It should be human readable and resemble assembler code.

    Instructions: with the exception of immediate mode, operands are addresses and instructions implicitly dereference values in memory located at those addresses. Words are understood to be a 64-bit size unit. Offsets are in bytes (nwords * 8).

    opcode/mnemonicC equivalent description
    ADD,SUB,MUL,DIVx = y op z store result of binary operation on y and z to x
    NEGx = -y store result of unary operation on y to x
    ASNx = y store y to x
    ADDRx = &y store address of y to x (*)
    LCONx = *y store contents pointed to by y to x (*)
    SCON *x = y store y to location pointed to by x (*)
    GOTOgoto L unconditional jump to L
    BLT,BLE,BGT,BGEif (x rop y) goto L test relation and conditional jump to L
    BIFif (x) goto L conditional jump to L if x==0
    BIFNif (!x) goto L conditional jump to L if x!=0
    PARMparam x store x as a parameter
    CALLcall p,n,x call procedure p with n words of parameters, store result in x
    RETreturn x return from procedure, use x as the result

    (*) Not all languages will use all TAC instructions.

    Mailbag

    How do I get started on code generation?
    Start at the leaves, a.k.a. basis cases. After you have done these things (you may need to print trees that show), you are ready to start allocating .icode
    What do we do about strings?
    There might be material about this elsewhere in the lecture notes, but here are a few thoughts:
    Three-Address Code Declarations (Pseudo instructions):

    declarationdescription
    glob x,n    
    declare a global named x that has n words of space
    proc x,n1,n2
    declare a procedure named x with n1 words of parameter space and n2 words of local variable space
    loc x,n
    use name x to refer to offset n in the local region (the procedure frame); replaces any prior definiton of x that may exist.
    lab Ln
    designate that label Ln refers to the next instruction
    end
    declare the end of the current procedure

    TAC for Composites/Containers and Object Oriented Code

    The sketchiness of the following table is pretty good evidence that we are just making this up as we go along.

    mnemonicequivalentdescription
    MEMBER
    x = y.zlookup field named z within y, store address to x
    NEW
    x = new Foo,ncreate a new instance of class named x, store address to x.
    Constructor is called with n parameters (previously pushed on the stack).
    class
    class x,n1,n2pseudoinstruction to declare a class named x with n1 bytes of class variables and n2 bytes of class method pointers
    field
    field x,npseudoinstruction to declare a field named x at offset n in the class frame

    Note: no new instructions are introduced for a member function call. In a non-virtual OO language, a member function call o.m(x) might be translated as Foo__m(o,x), where Foo is o's class. Other translation schemes are possible.

    Variable Reference, Dereference, and Assignment Semantics

    Given, say, x having a value of 2, what does the following compute?
       int y = x + (x = x + 1) + x;
    
    OK, what about
       int y = x + x + (x = x + 1) + x;
    
    In order to get the answers right, one has to understand the moment at which a variable reference is computed versus the moment at which it is dereferenced to obtain its value, versus the moment at which it is assigned a new value.

    Operator precedence (and parentheses) determine what order the expressions are evaluated. But evaluating something as simple as expr+expr can give surprise results if variables' values can change between their reference construction and dereferencing operation.

    Tree Traversals for Moving Information Around

    Like with semantic analysis, the intermediate code generation game is largely one of moving information around in the tree. The alternative to writing one huge recursion consisting of gigantic switch statements is the "swarm" model: write a suite of mutually-recursive functions that know how to do work for each different rule or each different type of non-terminal node for that traversal.

    Intermediate Code Generation: Traversal Code Example

    The following code sample illustrates a code generation tree traversal. Note the gigantic switch statement. A student once asked the question of whether the link lists might grow longish, and if one is usually appending instructions on to the end, wouldn't a naive link list do a terrible O(n2) job. To which the answer was: yes, and it would be good to use a smarter data structure, such as one which stores both the head and the tail of each list.
    void codegen(nodeptr t)
    {
       int i, j;
       if (t==NULL) return;
    
       /*
        * this is a post-order traversal, so visit children first
        */
       for(i=0;i<t->nkids;i++)
          codegen(t->child[i]);
    
       /*
        * back from children, consider what we have to do with
        * this node. The main thing we have to do, one way or
        * another, is assign t->code
        */
       switch (t->label) {
       case PLUS: {
          t->icode = concat(t->child[0]->icode, t->child[1]->icode);
          g = gen(PLUS, t->address,
                  t->child[0]->address, t->child[1]->address);
          t->icode = concat(t->icode, g);
          break;
          }
       /*
        * ... really, a bazillion cases, up to one for each
        * production rule (in the worst case)
        */
       default:
          /* default is: concatenate our children's code */
          t->icode = NULL;
          for(i=0;i<t->nkids;i++)
             t->icode = concat(t->icode, t->child[i]->icode);
       }
    }
    

    Intermediate Codegen Example

    CKotlin
    void write(int);
    void main()
    {
       int i;
       i = 5;
       i = i * i + 1;
       write(i);
    }
    
    fun main() {
       var i : Int;
       i = 5;
       i = i * i + 1;
       println("$i");
    }
    
    For the code of the main() function, we want something like
    proc main,0,32
    	ASN	loc:0,const:5
    	MUL	loc:8,loc:0,loc:0
    	ADD	loc:16,loc:8,const:1
    	ASN	loc:0,loc:16
    	PARAM	loc:0
    	CALL	write,1,loc:24
    	RETURN
    
    How how do we get there?

    Code Gen Example, cont'd

    After lexical and syntax analysis, the syntax tree (C version) is something like

    The prerequisites for intermediate codegen also include semantic analysis (symbol tables and typechecking), and we need to simulate that, or present it as already accomplished, in order for the actual codegen tree traversal to make sense. In particular, for codegen we need .type and .addr, which for variables come from the symbol table. I kinda promised you a code generation example last Friday, but decided that we need to talk about control flow first! And before that, here's a few more thoughts on built-ins and strings.

    More About Built-ins, and Strings

    Previously we have said we'd need a few minimal things as built-ins.

    What is the bare minimum for strings in Kotlin?

    Compute the Offset of Each Variable

    Add an address field to every symbol table entry.
    The address contains a region plus an offset in that region.
    No two variables may occupy the same memory at the same time.
    At the intermediate code level, do not bother to re-use memory. In optimization and then in final code, re-use will be a big thing.

    Locals and Parameters are not Contiguous!

    For each function you need either to manage two separate regions for locals and for parameters, or else you need to track where in that region the split between locals and parameters will be. This may become more "interesting" if parameters are passed in registers.

    lecture #22 began here

    Announcements

    Basic Blocks

    A basic block is a sequence of 1+ instructions in which there are no jumps into or out of the middle. In the most extreme case, every instruction is a basic block. Start from that perspective and then lump adjacent instructions together if nothing can come between them.

    What are the basic blocks in the following 3-address code? ("read" is a 3-address code to read in an integer.)

    	read x
    	t1 = x > 0
    	if t1 == 0 goto L1
    	fact = 1
    	label L2
    	t2 = fact * x
    	fact = t2
    	t3 = x - 1
    	x = t3
    	t4 = x == 0
    	if t4 == 0 goto L2
    	t5 = addr const:0
    	param t5		; "%d\n"
    	param fact
    	call p,2
    	label L1
    	halt
    

    Discussion of Basic Blocks

    Basic blocks are often used in order to talk about specific types of optimizations.
    For example, there are optimizations that are only safe to do within a basic block, such as "instruction reordering for superscalar pipeline filling".
    So, why introduce basic blocks here?
    our next topic is intermediate code for control flow, which includes gotos and labels, so maybe we ought to start thinking in terms of basic blocks and flow graphs, not just linked lists of instructions.
    view every basic block as a hamburger
    it will be a lot easier to eat if you sandwich it inside a pair of labels:
    	label START_OF_BLOCK_7031
    	...code for this basic block...
    	label END_OF_BLOCK_7031
    
    the label sandwich lets you:
    For example, for an if-then statement, you may need to jump to the beginning of the statement in the then-part...or you may need to jump over it, the choice depending on the outcome of a boolean.
    Yeah, these lecture notes repeat themselves about the label sandwich, almost immediately. That must be on purpose.

    Operators

    In case you were fuzzy on the operators, you need to support.
    Essential (C) Non-essential In k0
    = += -= *= /= %= <<= >>= &= ^= |= = += -=
    postfix ++ --
    + - * / % >> << prefix ++ -- ^ + - * / % (?)
    && || ! & | ~ && || !
    < <= > >= == != ternary x ? y : z ?:
    expr[expr] &x x->y *x x.y ?

    Old Mailbag

    What do I have to do to get a "D"?
    You are graded relative to your peers. In previous semesters the answer to this has been something like: pass the midterm and final, and convince me that you really did semantic analysis. If you did poorly on the midterm, you might want to try and do better on the final, and you might want to get some three address code working. Do you really want to settle for a "D"? Almost everyone who was "D" material dropped the class already.
    Are we supporting comparisons of doubles with ==, !=, <=, and >=?
    Yes.
    Is it possible/legal to have variables of type Unit / void?
    No.
    Do we have to support NULL? To what types of variable is it legal to assign a NULL?
    CSE 113 probably does have explicit uses of null, mainly in comparisons.
    Do we need separate intermediate code instructions for floating point and for integer operations?
    Good question. What do you think?

    Intermediate Code for Control Flow

    Code for control flow (if-then, switches, and loops) consists of Each chunk of code that is executed together (no jumps into or out of it) is called a basic block.

    Depending on your source language's semantic rules for things like "short-circuit" evaluation for boolean operators, the operators like || and && might be similar to + and * (non-short-circuit) or they might be more like if-then code.

    A general technique for implementing control flow code:

    Mailbag

    I am encountering a very interesting segfault when testing my hw. as seen in pic below.

    This bug only comes up with a 'char' declaration, otherwise it runs fine on an int or String of the same name. This 'char' decl for some reason parses fine by yydebug, however when populating the sym table, it is interpreted as a ClassDecl (pic2).

    I have looked at my grammar and cannot figure out why this would happen or how I can figure out why this is happening. Any tips or ideas on what I might be missing here?
    Wow, thank you for this excellent question, and for the wonderful screenshots. The first screenshot tells us we are 12 levels deep in recursion at the TOD. But it doesn't give us the line numbers at all these levels. Are you compiling with -g on your gcc -c lines? You should be. The second picture leaves us guessing. A wrong internal tree node could be the result of building the tree structure/shape wrong, or putting the wrong information into the tree. We might start by looking at your suspicious .y file, but when you seek help by e-mail or appointment, I will probably need to see your flex file and/or the code that prints the symbol tables, in order to find this bug.

    If-Then and If-Then-Else

    The labels have to actually be allocated and attached to instructions at appropriate nodes in the tree corresponding to grammar production rules that govern control flow. An instruction in the middle of a basic block need neither a first nor a follow.

    ProductionSemantic Rules
    IfThenStmt :
          if '(' Expr ')' Stmt
    Expr.onTrue = Stmt.first
    Expr.onFalse = IfThenStmt.follow
    Stmt.follow = IfThenStmt.follow
    IfThenStmt.icode = (Expr.icode != null) ? Expr.icode
                                                        : gen(BIF, Expr.onFalse, Expr.addr, con:0)
    IfThenStmt.icode |||:= gen(LABEL, Expr.onTrue) ||| Stmt.icode
    IfThenElseStmt :
          if '(' Expr ')' Stmt1 else Stmt2
    Expr.onTrue = Stmt1.first
    Expr.onFalse = Stmt2.first
    Stmt1.follow = IfThenElseStmt.follow;
    Stmt2.follow = IfThenElseStmt.follow;
    IfThenElseStmt.icode = (Expr.icode != null) ? Expr.icode
                                                        : gen(BIF, Expr.onFalse, Expr.addr, con:0)
    IfThenElseStmt.icode |||:= gen(LABEL, Expr.onTrue) ||| Stmt1.icode |||
          gen(GOTO, IfThenElseStmt.follow) ||| gen(LABEL, Expr.onFalse) ||| Stmt2.icode

    Generating Code for Conditions

    The big picture on code generation for control structures such as if's and while's requires an understanding of how to generate code for the boolean expressions that control these constructs.

    Alternative Models of Control Flow

    Different languages have different semantics for booleans.

    Grade Jitters?

    Canvas is certainly not a viable measuring stick for your grade. Your measuring sticks are your midterm grade, graded HWs, whether you are making regular progress, and whether you are seeking help.
    If you are struggling in this class, that is normal and to a certain degree, healthy.
    The goal is for everyone to get xp and levels out of the course proportional to what you put in. Even the smartest persons in the class.
    Get as far as you can by end of semester.

    Mailbag

    Can we assume all types are 8 bytes?
    We can almost assume that. Types that are not 8 bytes can be an 8 byte pointer to something that is not 8 bytes.
    When calling gen() for an ASN instruction, can I simply leave the third source addr empty?
    You can pass it as an {R_NOTUSED, 0} address or something similar of your own advising. If, based on the opcode, it is never read, it will not matter what you put in that field.
    How should we use ASN with array initializations?
    Options include
    1. Initialize array elements one at a time, as if they were assignments.
    2. Creating a new intermediate code instruction, named AASN for example.
    3. Defining an internal runtime system function for array initialization, and calling it when neeeded
    For these, or any other options, there is also the issue of how the elements in the initialer are stored in the code. Option #1 might store them in immediate mode operands of multiple instructions. Options #2 or #3 might involve writing them out in a "constant region" similar to the "string region" postulated for storing string constants.
    For intermediate code generation, I sorta understand that we are supposed to get labels added at key places in our tree as semantic properties for the next step. What I do not understand is what I will be doing in more detail. What should this be looking like.
    Tree traversals that compute more attributes. A lot of small do-able pieces of work that will be combined. Steps to intermediate code generation include:
    Traditionally, % only works on integer arguments. Do I need to ensure that, or do I need to worry about modulus for other types?
    % requires integer arguments. You need to type check it.
    How do I do float constants?
    Most CPUs do not have float immediate instructions. We need an actual constant region; we need that for string constants too. Depending on your target code, you might create a region for each constant type, perhaps R_FLOAT and R_STRING, with byte offsets starting at word boundaries. With other targets such as Unicon ucode, you probably should just do one combined region, perhaps R_RODATA.

    Boolean Expression Example

    a<b || c<d && e<f
    
    Compare three intermediate code solutions given below.

    conditional branches relop instructions short circuit
    100:	if a<b goto 103
    	t1 = 0
    	goto 104
    103:	t1 = 1
    104:	if c<d goto 107
    	t2 = 0
    	goto 108
    107:	t2 = 1
    108:	if e<f goto 111
    	t3 = 0
    	goto 112
    111:	t3 = 1
    112:	t4 = t2 AND t3
    	t5 = t1 OR t4
    
    t1 := a LT b
    t2 := c LT d
    t3 := e LT f
    t4 := t2 AND t3
    t5 := t1 OR t4
    
        if a<b goto E.true
        if c<d goto L1
        goto E.false
    L1: if e<f goto E.true
        goto E.false
    

    Short circuit semantics is short, fast, and can be used to play parlor tricks.

    lecture #23 began here

    Mailbag

    Someone told me one time that a language is only useful if it supports interfacing with C libraries, as most of the useful libraries in the world are written in C; If it is unable to interface with these libraries (such as stdio.h in this instance), programmers will seldom find it useful unless it has as robust a standard library of its own as the C standard library. What do you think? Is this a naïve solution that has a better alternative?
    Wow, what a C-supremacist worldview. This is largely true when writing software for well-known and established domains. It is a less useful argument for those in the business of exploring new application domains. Python is popular in large part because its C-calling interface was quite good. Java had to go to considerable effort to make C calling not too awful, and then over time Java has improved its performance and developed its own libraries to the point where it is at least somewhat competitive in some domains. Kotlin is standing on the shoulders of the giant called Java, which is standing on the shoulders of the giant called C. Fine.

    For k0, we are not contemplating whether k0 source code should be able to call C code, but we can get a lot of mileage out of making our generated code be able to call C code to do the work for us in our runtime system.

    Arrays

    Example Code

    Consider the following example code, adapted from a k0 test case.
    fun main() {
      var i : Int
      var a : Array<Int> = Array<Int>(8) {0}
      a[5] = 5
      i = a[5]
      println("i is $i!!!")
    }
    
    The array of 8 integers requires 64 bytes. It is initialized to zeroes Element 5, the sixth element, is set to 5. It is printed out, by first copying it into a simple variable for which a k0 println() command should work.

    Array Creation

    As a student mentioned in class, one of the big issues is: in what memory region are arrays allocated? A C-style implementation would allocate them on the stack, in a local region. A slightly more Java-like implementation would allocate them on the heap.

    In the preceding example, allocating 64 bytes in the local region on the stack might be as simple as inserting a symbol table entry for local variable "a", specifying its address loc:8 (offset 8, after variable i) and size (64) in the symbol table entry. The local region total size is larger due to temporary variables. If you were doing heap-allocation, the actual address of the array is not known until runtime, what is allocated for variable a is 8 bytes for a pointer.
    C arrays array "Objects"
    proc main,0,104		// 8 + 64 + 8 + 8 + 8 + 8
    	...
    	end
    
    proc main,0,48		// 8 + 8 + 8 + 8 + 8 + 8
    	PARAM  const:64
    	CALL   malloc,1,loc:8
    	...
    	end
    

    Array Initialization

    k0 simplifies the number of ways to specify initial values of an array down to one way. Initialization of local variable i at (loc:0) might be similar (not shown). (Q: is initializing i upon creation required by k0?)

    In the heap-allocated array, the SCON instruction stores/copies through a pointer. The array arithmetic needed foreshadows what we will need for the subscript operator, computing the address at an offset from a base pointer known only at runtime.
    C arrays array "Objects"
    proc main,0,120		// 8 + 64 + 8 + 8 + 8 + 8 + ...
    	ASN loc:8,const:0
    	ASN loc:16,const:0
    	ASN loc:24const:0
    	ASN loc:32,const:0
    	ASN loc:40,const:0
    	ASN loc:48,const:0
    	ASN loc:56,const:0
    	ASN loc:64,const:0
    	end
    
    proc main,0,48		// 8 + 8 + 8 + 8 + 8 + 8
    	PARAM  const:64
    	CALL   malloc,1,loc:8
    	ASN  loc:16, loc:8
    	SCON loc:16, const:0	
    	ADD  loc:16, const:8
    	SCON loc:16, const:0	
    	ADD  loc:16, const:8
    	SCON loc:16, const:0	
    	ADD  loc:16, const:8
    	SCON loc:16, const:0	
    	ADD  loc:16, const:8
    	SCON loc:16, const:0	
    	ADD  loc:16, const:8
    	SCON loc:16, const:0	
    	ADD  loc:16, const:8
    	SCON loc:16, const:0	
    	ADD  loc:16, const:8
    	SCON loc:16, const:0	
    	end
    

    Assigning to a[i]

    In general, the address of a[i] is i*elemsize bytes from the base address. For a constant index such as a[5] we could compute the offset at compile time, but the code below is for the more general case.
    C arrays array "Objects"
    proc main,0,120		// 8 + 64 + 8 + 8 + 8 + 8
    	...
    	ADDR	loc:72, loc:8			// t1 = &(a[0])
    	MUL	loc:80, const:5, const:8	// t2 = 40
    	ADD	loc:88, loc:72, loc:80		// t3 = t1 + t2
    	SCON	loc:88, const:5			// *t3 = 5
    	ASN	loc:96, loc:8			// t4 = &(a[0])
    	MUL	loc:104, const:5, const:8	// t5 = 40
    	ADD	loc:112, loc:72, loc:80		// t6 = t4 + t5
    	LCON	loc:0, loc:112			// i = *t6
    	end
    
    proc main,0,120		// 8 + 8 + 8 + 8 + 8 + 8 + ...
    	PARAM  const:64
    	CALL   malloc,1,loc:8
    	...
    	ASN	loc:16, loc:8			// t1 = &(a[0])
    	MUL	loc:24, const:5, const:8	// t2 = 40
    	ADD	loc:32, loc:16, loc:24		// t3 = t1 + t2
    	SCON	loc:32, const:5			// *t3 = 5
    	ASN	loc:40, loc:8			// t4 = &(a[0])
    	MUL	loc:48, const:5, const:8	// t5 = 40
    	ADD	loc:56, loc:40, loc:48		// t6 = t4 + t5
    	LCON	loc:0, loc:56			// i = *t6
    
    	end
    

    Strings

    Would you be happier just implementing k0 strings as C strings, or defining your own representation? For k0, C strings are fine for an underlying representation. We will use that for the discussion today, but you can use a different representation if you prefer.

    k0 string concatenation

    String s3 = s1 + s2
    
    literally, use strcat. You can add a three address code for it, or generate three address code where it is a function call.
       // ... t1 = strlen(s1)
       // ... t2 = strlen(s2)
       // ... s3 = malloc(t1+t2)
       // ... strcpy(s3,s1)
       PARAM s2
       PARAM s3
       CALL  strcat, 2, s3
    

    Kotlin string literals with embedded variables, e.g. $x

    This is for things like "i is $i!!!". If it contains one or more occurrences of $ it is really "i is " || string(i) || "!!!" where || is string concatenation and string(i) is a conversion-to-string function that our generated code must call.

    ... Back to Control Flow / Boolean Eval Topics

    Do we know enough now to write the code generation rules for booleans?

    Production Semantic Rules
    AndExpr :
          AndExpr1 && EqExpr
     EqExpr.first = genlabel();
     AndExpr1.onTrue = EqExpr.first;
     AndExpr1.onFalse = AndExpr.onFalse;
     EqExpr.onTrue = AndExpr.onTrue;
     EqExpr.onFalse = AndExpr.onFalse;
     AndExpr.icode = AndExpr1.icode ||| gen(LABEL, EqExpr.first) ||| EqExpr.icode;
    
    OrExpr :
          OrExpr1 || AndExpr
     AndExpr.first = genlabel();
     OrExpr1.onTrue = OrExpr.onTrue;
     OrExpr1.onFalse = AndExpr.first;
     AndExpr.onTrue = OrExpr.onTrue;
     AndExpr.onFalse = OrExpr.onFalse;
     OrExpr.icode = OrExpr1.icode ||| gen(LABEL, AndExpr.first) ||| AndExpr.icode;
    
    UnaryExpr : ! UnaryExpr1
    UnaryExpr1.onTrue = UnaryExpr.onFalse
    UnaryExpr1.onFalse = UnaryExpr.onTrue
    UnaryExpr.icode = UnaryExpr1.icode
    

    Hints: parent fill's out childrens' inherited attributes...

    Intermediate Code for Relational Operators

    C syntaxgotos bool value both ?
    E-> E1 < E2 E.icode = E1.icode || E2.icode ||
          gen(BLT, E1.addr, E2.addr, E.true) ||
          gen(GOTO, E.false)
    E.addr = genlocal()
    E.icode = E1.icode || E2.icode ||
          gen(LT, E.addr, E1.addr, E2.addr)
    E.addr = genlocal()
    E.icode = E1.icode || E2.icode ||
          gen(LT, E.addr, E1.addr, E2.addr) ||
          gen(BIF, E.addr, E.true) ||
          gen(GOTO, E.false)

    Intermediate Code for Loops

    While Loops

    A while loop has semantic attributes and rules for intermediate code that are very similar to an if-statement. There is almost only one difference, the goto back to the beginning. Is there anything else missing or wrong here?

    Finishing touches: what attributes and/or labels does this plan need in order to support break and continue statements?

    For Loops

    For-loops can be trivially transformed into while loops, so they pose just about no new code generation issues. Notice that only some expressions need .true/.false: the ones used as conditionals.

    ProductionSemantic Rules
    WhileStmt : while '(' Expr ')' StmtExpr.onTrue = genlabel();
    Expr.first = genlabel();
    Expr.false = WhileStmt.follow;
    Stmt.follow = Expr.first;
    WhileStmt.icode = gen(LABEL, Expr.first) |||
       Expr.icode ||| gen(LABEL, Expr.true) |||
       Stmt.icode ||| gen(GOTO, Expr.first)
    ForStmt : for( ForInit; Expr; ForUpdate )
        Stmt

    a.k.a.

    ForInit;
    while (Expr) {
        Stmt
        ForUpdate
    }
    Expr.true = genlabel();
    Expr.first = genlabel();
    Expr.false = S.follow;
    Stmt.follow = ForUpdate.first;
    S.icode = ForInit.icode |||
       gen(LABEL, Expr.first) |||
       Expr.icode ||| gen(LABEL, Expr.true) |||
       Stmt.icode |||
       ForUpdate.icode |||
       gen(GOTO, Expr.first)

    Again: what attributes and/or labels does this plan need in order to support break and continue statements?

    lecture #24 began here

    Code generation for Switch Statements

    Consider the C switch statement
    switch(e) of {
       case v1:
          S1;
       case v2:
          S2;
       ...
       case vn-1:
          Sn-1;
       default:
          Sn;
    }
    
    The intermediate code for this might look like:
    	code for e, storing result in temp var t
    	goto Test
    L1:
    	code for S1
    L2:
    	code for S2
    	...
    Ln-1:
    	code for Sn-1
    Ln:
    	code for Sn
    	goto Next
    Test:
    	if t=v1 goto L1
    	if t=v2 goto L2
    	...
    	if t=vn-1 goto Ln-1
    	goto Ln
    Next:
    
    C "break" statements
    can be implemented in S1-Sn
    by "goto Next" instructions.









    Now, why would this code generation template for switches put the Tests last?

    This three address code is not very fast if you translate it naively. In final code, on many CPUs you can avoid the long sequence of conditional branch instructions and go to an address stored in array of addresses.

    Name Mangling

    So: prefix all globals with a prefix "_" or "k0_" or whatever.

    Intermediate Code Generation Example

    Consider the following small program. It would be fair game as input to your compiler project. In order to show blow-by-blow what the code generation process looks like, we need to construct the syntax tree and do the semantic analysis steps.

    C Kotlin
    void print(int i);
    void main()
    {
       int i;
       i = 0;
       while (i < 20)
          i = i * i + 1;
       print(i);
    }
    
    fun main() {
      var i : Int = 0;
      while (i < 20) {
        i = i * i + 1;
        }
      println("$i");
    }
        

    For intermediate code, we need to fatten up the syntax tree quite a bit. Each tree node may hold:

    Type pointer and icode are generally drawn as pointers to another rectangle, denoting a malloced block of memory. Place can be drawn as a two-field struct containing a region and an offset, as in
    LOCAL
    0
    or it can be rendered textually, e.g. LOCAL:0.

    We can do this all day long. Maybe we will. The image file here is named semtree.png. Lets see how much we can mark it up with semantic and codegen info.

    Below is a fragment with

    We proceeded with a discussion of how to build the .icode fields. One thing that got said was: .icode fields get built via a post-order traversal (synthetised attribute) but .true, .false etc. are inherited and may require a previous pass through the tree, or a pre-order traversal. If you are trying to do both in one pass it might look like the following.

    Another Code Generation Function

    void codegen(struct tree *t)
    {
       // pre-order stuff, e.g. label generation
       switch (t->prodrule) {
             ...
          case ITERATION_STMT: // inherited attributes for while loop
             // push an inherited attribute to child before visiting them
             t->child[2]->true = genlabel();
             break;
    	 ...
          }
    
       // visit children
       for(i=0; i < t->nkids; i++) codegen(t->child[0]);
    
       // post-order stuff, e.g. code generation
       switch (t->prodrule) {
             ...
          case CONDEXPR_2: // synthesized attribs for CondExpr: Expr < Expr
             t->code = concat(
    	        t->child[0]->code,
    	        t->child[2]->code,
    		gen(BLT, t->child[0]->place, t->child[2]->place, t->true),
    		gen(GOTO, t->false)
    		);
    	 break;
          case ITERATION_STMT: // synthesized attributes for while loop
    	 t->code = concat(
                        gen(LABEL, t->child[2]->first),
    		    t->child[2]->code,
                        gen(LABEL, t->child[2]->true),
       		    t->child[4]->code,
    		    gen(GOTO, t->child[2]->first));
    	 break;
    }
    
    The code for the boolean conditional expression controlling the while loop is a list of length 1, containing the instruction t0 = i < 20, or more formally
    gotos bool value
    opcodedestsrc1src2
    BLTi20E.true
    GOTOE.false
    opcodedestsrc1src2
    LTt0i20

    The actual C representation of addresses dest, src1, and src2 is a
    region

    offset
    pair, so the picture of this intermediate code instruction really looks something like this:

    gotosbool value
    opcodedestsrc1src2
    BLTlocal

    i.offset
    const

    20
    code

    (E.true's label#)
    opcodedestsrc1src2
    LTlocal

    t0.offset
    local

    i.offset
    const

    20

    Regions are expressed with a simple integer encoding like: global=1, local=2, const=3. Note that address values in all regions are offsets from the start of the region, except for region "const", which stores the actual value of a single integer as its offset.

    opcodedestsrc1src2
    MULlocal

    t1.offset
    local

    i.offset
    local

    i.offset

    The rest of class was spent elaborating on the linked list of instructions for the preceding example.

    .first and .follow for StmtList

    Suppose you have grammar rules

    FuncBody : '{' StatementList '}' ;
    StatementList : StatementList Statement ;
    StatementList :  ;
    
    What kind of .first and .follow values can we develop and pass in to children from these rules?



    Syntax Attribute Manipulations
    FuncBody : '{' StatementList '}'
    StatementList.follow = genlabel();
    FuncBody.icode = StatementList.icode ||
                     gen(LABEL,StatementList.follow) ||
    		 gen(RETURN)
    StatementList1: StatementList2 Statement
    StatementList2.follow = Statement.first;
    Statement.follow = StatementList1.follow
    StatementList1.icode = StatementList2.icode || Statement.icode
    StatementList : ;
    /* no need for a StatementList.follow */
    StatementList.first = genlabel()
    StatementList.icode = gen(LABEL, StatementList.first) || gen(NOOP)

    Final Grades and Projects; End-of-semester considerations

    HW#7

    Give me execution by any means possible

    More Intermediate Code Generation Examples

    You can implement your code generation one operator at a time, simplest expressions first. Implement and test incrementally starting from "base cases".

    Zero operators.

    if (x) S
    
    translates into
    if x != 0 goto L1
    goto L2
    label L1
    ...code for S
    label L2
    
    or if you are being fancy
    if x == 0 goto L1
    ...code for S
    label L1
    
    I may do this without comment in later examples, to keep them short.

    One relational operator.

    if (a < b) S
    
    translates into
    if a >= b goto L1
    ...code for S
    label L1
    
    One boolean operator.

    if (a < b  &&  c > d) S
    
    translates into
    if (a < b)
       if (c > d)
          ...code for S
    
    which if we expand it
    if a >= b goto L1
    if c <= d goto L2
    ...code for S
    label L2
    label L1
    
    by mechanical means, we may wind up with lots of labels for the same target (instruction), this is OK.

    Beware the following. A lazy code generator doing short-circuits might be tempted to say that

    if (a < b  ||  c > d) S
    
    translates into
    if (a < b) ...code for S
    if (c > d) ...code for S
    
    but its unacceptable to duplicate the code for S! It might be huge! Generate labels for boolean-true=yes-we-do-this-thing, not just for boolean-false=we-skip-this-thing.
    if a < b goto L1
    if c > d goto L2
    goto L3
    label L2
    label L1
    ...code for S
    label L3
    

    Object-Oriented Changes to Above Examples

    The previous examples were assuming a C-like language semantics. For an object-oriented language, the generated code for these examples is more interesting. For example, the semantics of
    if (x) S
    
    if x is an object, may be defaulted to be equivalent to
    if (x != NULL) S
    
    or more generally, the different types may have (hardwired, or overrideable) conversion rules to convert them to booleans for use in tests, such as
    tempvar := x.as_boolean()
    if (tempvar) S
    

    Old Mailbag

    Do we need to specify the RET instruction at the end of a function or does the END instruction imply that the function returns?
    I think of END in three-address code as a non-instruction (pseudo instruction) that marks the end of a procedure. So you should have a RET in front of it. But really, you are allowed to define END's semantics to also do a RET; you could call it REND.
    If we have nothing to return, can we just say RET with no parameter or must the parameter x always be there, i.e. RET x?
    I would accept a RET with no operand. You are allowed to define new opcodes in intermediate code. Native assemblers often have several variants of a given instruction -- same mnemonic, different opcodes for a related family of instructions. But in that case our final codegen would have to have a way to tell which version of RET is in use. Either a way of marking unused addresses, or a separate opcode such as RET0.
    Can you give me an example of when to use the GLOBAL and LOCAL declaration instructions?
    These are pseudo-instructions, not instructions. Globals are listed as required; at the minimum, if your program has any global variables you must have at least one GLOBAL declaration to give the size of (the sum of) the global variables. You can do one big GLOBAL and reference variables as offsets, or you can declare many GLOBAL regions, effectively defining one named region for each variable and therefore rendering the offsets moot.

    A LOCAL pseudo-instruction is listed as optional and advisory; think of it as debugging symbol information, or as an assist to the reader of your generated assembler source.

    Code Generation for Arrays

    Consider first the subscript operator for C-like arrays. Then consider how it ought to work in your compiler.

    So far, we have only said, if we passed an array as a parameter we'd have to pass its address. 3-address instructions have an "implicit dereferencing semantics" which say all addresses' values are fetched / stored by default. So when you say t1 := x + y, t1 gets values at addresses x and y, not the addresses. Once we recognize arrays are basically a pointer type, we need 3-address instructions to deal with pointers.

    now, what about arrays? reading an array value: x = a[i]. Draw the picture. Consider the machine uses byte-addressing, not word-addressing. Unless you are an array of char, you need to multiply the subscript index by the size of each array element...

    t0 := addr a
    t1 := i * 8
    t2 := plus t0 t1
    t3 := deref t2
    x  := t3
    
    What about writing an array value?

    There are similar object-oriented adaptation issues for arrays: a[i] might not be a simple array reference, it might be a call to a method, as in

    x := a.index(i)
    
    or it might be implemented like:
    x := a field i
    
    The main issue to keep straight in both the C-like example and the object-oriented discussion is: know when an instruction constructs an address and stores an address in a memory location. When you want to read or write to the address pointed to by the constructed address, you may need to do an extra level of pointer-following. Three address instructions have "implicit" pointer-following since all addresses are followed when reading or writing memory, but if what is in the address is another address, you have to be careful to keep that straight.

    Let it be noted for the record that the lecture notes have a lot of additional material on intermediate code generation, and we can spend as much time on under-described aspects and Q&A as we need, but otherwise, is it time to

    Warp to Final Code Gen?

    Supplemental Comments on Code Generation for Arrays

    In order to generalize from our example last lecture, the 3-address instructions for
    expr [ expr ]
    
    ideally should generate code that computes an address that can subsequently be read from or written to. One can certainly write a three address instruction to compute such an address. With arrays this is pointer arithmetic.

    With tables, a.k.a. dictionaries, the main wrinkle is: what to do if the key is not in the table? The behavior might be different for reading a value or writing a value:
    syntax behavior
    t[x] := y if key is not in table, insert it
    y := t[x] if key is not in table, one of:
    • produce a default value
    • raise an exception
    • ??

    Code Generation for Maps (Dictionaries, Tables)

    Consider the map type for a moment. Example presented from the point of view of Go, but Python dictionaries and Unicon tables are similar. One can generate code for maps either by extending the three-address instruction set with new instructions, or by generating function calls. How might you implement
    map construction: make(map[string]int)
    Needs to allocate one hash table (array of buckets) from the heap. For a compiler, keys were always string. In a programming language, keys may be more diverse: they can be string, or int, or object... Maybe the generated code should emit different opcodes/functions depending on key type, or at very least an argument that specifies this. When keys can be arbitrary type, it adds complexity that might get pushed off to runtime.
    Via 3-address Instructions Via function call
    MAPCREATE  dest
    
    CALL	mapcreate,?,?
    
    insert: x[s] = s2
    Needs to compute an address into which to store s2.
    Via 3-address Instructions Via Function call
    MAPINSERT  map,key,val
    
    PARAM	map
    PARAM	key
    CALL    mapinsert,?,val
    
    lookup: s = x[s2]
    Via 3-address Instructions Via Function call
    MAPLOOKUP   tmp,map,key
    ASN	    s, tmp
    
    PARAM   map
    PARAM   key
    CALL    maplookup,,tmp
    ASN     s, tmp
    

    Debugging Miscellany

    Prior experience suggests if you are having trouble debugging, check:
    makefile .h dependencies!
    if you do not list makefile dependencies for important .h files, you may get coredumps!
    traversing multiple times by accident?
    at least in my version, I found it easy to accidentally re-traverse portions of the tree. this usually had a bad effect.
    bad grammar?
    our sample grammar was adapted from good sources, but don't assume its impossible that it could have a flaw or that you might have messed it up.
    bad tree?
    its entirely possible to build a tree and forget one of your children

    A few observations from Dr. D

    I went shopping for more intermediate code examples, and while I didn't find anything as complete as I wanted, I did find updated notes from the same Jedi Master who trained me, check it:

    Dr. Debray's Intermediate Code Generation notes.

    You can consider these a recommended supplemental reading material, and we can scan through them to look and see if they add any wrinkles to our prior discussion.

    A Bit of Skeletal Assistance with Three Address Code

    Example of Generating .first and .follow Attributes

    What nodes need these?

    Call gen_firsts(root) followed by gen_follows(root) before generating code.

    .first.follow
    What?
    • synthesized attribute
    • a label (#) to precede all executable instructions for a given chunk of code
    Why?
    • loops may go back to their .first.
    • preceding statements' .follow may be inherited from your .first
    Sample code:
    void gen_firsts(nodeptr n)
    {
       if (n == NULL) return;
       for(i=0; inkids; i++)
          gen_firsts(n->kids[i]);
    
       switch (n->prodrule) {
       case LABELED_STATEMENT:
          n->first = /* ... just use the explicit label */
          break;
       case EXPRESSION_STATEMENT:
       case COMPOUND_STATEMENT:
       case SELECTION_STATEMENT:
       case ITERATION_STATEMENT:
       case JUMP_STATEMENT:
       case DECLARATION_STATEMENT:
       case TRY_BLOCK:
          n->first = genlabel();
          break;
       default:
       }
    }
    
    Why?
    • if we skip a then-part or do a then-part and have to skip an else-part
    • if we have to break out of a loop
    What?
    • inherited attribute
    • a label to go to whatever comes after the executable instructions for a given chunk of code.
    • Could try to dodge, by blindly generating labels at the end of each statement ("label sandwich" approach).
    void gen_follows(nodeptr n)
    {
       if (n == NULL) return;
    
       switch (n-<prodrule) {
       case STATEMENT_SEQ + 2: /* statement_seq : statement_seq statement */
            n->child[0]->follow = n->child[1]->first;
    	n->child[1]->follow = n->follow;
       	break;
       case COMPOUND_STATEMENT + 1: /* compstmt : '{' statement_seq_opt '}' */
            if (n->child[1] != NULL)
               n->child[1]->follow = n->follow;
       	break;
       case FUNCTION_DEFINITION + 1: /* funcdef : declarator ctor_init_opt body */
            n->child[2]->follow = genlabel();
            /* .icode must have this label and a return at the end! */
       	break;
       /* ... other cases? ... */
       default:
       }
    
       for(i=0; i<n->nkids; i++)
          gen_follows(n->kids[i]);
    }
    

    Labels for break and continue Statements

    Sample code for Option #2 is given below. Implied by the BREAK case is the notion that the .addr field for this node type will hold the label that is the target of its GOTO. How would you generalize it to handle other loop types, and the continue statement? There may be LOTS of different production rules for which you do something interesting, so you may add a lot of cases to this switch statement.

    void do_break(nodeptr n, address *loopfirst, address *loopfollow)
    {
       switch (n->prodrule) {
       case BREAK:
          if (loopfollow != NULL)
    	 n->place = *loopfollow;
          else semanticerror("break with no enclosing loop", n);
          break;
       case WHILE_LOOP:
          loopfirst = &(n->first);
          loopfollow = &(n->follow);
          break;
          ...
          }
    
       for(i=0; i<n->nkids; i++)
          do_break(nodeptr n, loopfirst, loopfollow);
    }
    

    TAC or die trying

    We need a simple example, in which you see It is easy to spend too much class time on front-end stuff before getting to a too-short and still under-explored TAC code generation phase. Our Goal: The perfect example would include a few statements, expressions, control flow constructs, and function calls. Here is an such an example. Notes for this exercise:
    C version pseudo-VGo version
    void printf(char *, int);
    int fib(int i);
    int readline(char a[]);
    int atoi(char a[]);
    int main() {
       char s[64];
       int i;
       while (readline(s)!=0 && s[0]!='\004') {
          i = atoi(s);
          if (i <= 0) break;
          printf("%d\n", fib(i));
          }
    }
    
    func fib(int i) int {
       if n <= 1 { return 1 }
       else { return fib(n-1) + fib(n-2) }
    }
    func ctoi(string s) int {
       if s == "0" {return 0}
       else if s == "1" {return 1}
       else if s == "2" {return 2}
       else if s == "3" {return 3}
       else if s == "4" {return 4}
       else if s == "5" {return 5}
       else if s == "6" {return 6}
       else if s == "7" {return 7}
       else if s == "8" {return 8}
       else if s == "9" {return 9}
    }
    func atoi(string s) int {
      var i int
      for ... /* while (#s > 0) */ {
         //?? string c = s[1]
         //?? i = i * 10 + ctoi(c)
         //?? s = s[2:0]
          }
       return i
    }
    func itoa(i int) string {
       var s string
       var div, rem int
       if i == 0 { return "0" }
       else if i == 1 { return "1" }
       else if i == 2 { return "2" }
       else if i == 3 { return "3" }
       else if i == 4 { return "4" }
       else if i == 5 { return "5" }
       else if i == 6 { return "6" }
       else if i == 7 { return "7" }
       else if i == 8 { return "8" }
       else if i == 9 { return "9" }
       else if i < 0 { return "-" + itoa(-i) }
       else {
         div = i / 10
         rem = i % 10
         return itoa(div) + itoa(rem)
       }
    }
    func main() {
      var s string
      var i int
      for ... {
          i = atoi(s)
          if i <= 0 { break }
          fmt.Println(itoa(fib(i)))
          }
    }
    

    string ftoa(double d)
    {
       if (d == 0.0) {
          return "0.0"
       }
       else if (d < 0.0) {
          return "-real"
          }
       else {
         return "real"
       }
    }
    

    Using cgram.y nonterminal names, let's focus on code generation for the main procedure.

    TAC-or-die: the First-level

    Potentially, this is a separate pass after labels have been generated.

    The first tree node that TAC code hits in its bottom up traversal is IDENTreadline (no .icode), followed by IDENTs (no .icode). Above the IDENTs, argument_expression_list is one of those non-terminals-with-only-one-child that matters and needs to be in the tree: each time it churns out an actual parameter, TAC code generates a PARAM instruction to copy the value of the parameter into the parameter region. PARAM indicates an 8-byte (word) parameter; you might also want to define PARAM4, PARAM2, PARAM1 (etc.) instructions. Q: why is the ADDR instruction here?

    	ADDR   loc:72,loc:0
    	PARAM  loc:72
    

    The postfix_expr is a function call, whose TAC codegen rule should say: allocate a temporary variable t0 (or as we called it: LOC:80) for the return value, and generate a CALL instruction

    	CALL readline,1,loc:80
    
    The next leaf (ICON0) has no .icode, which brings code generation up to the != operator. Here the code depends on the .true (L5) and .false (L2) labels. The TAC code generated is
    	BNE loc:80,const:0,lab:5
    	GOTO lab:2
    
    After that, the postfix traversal works over to IDENTs (no .icode), ICON0 (no .icode), and up to the postfix expression for the subscript operator for s[0]. It needs to generate .icode that will place into a temporary variable (its .addr, loc:88) what s[0] is.

    The basic expression for a[i] is baseaddr + index * sizeof(element). sizeof(element) is 1 in this case, so we can just add baseaddr + index. And index is 0 in this case, so an optimizer would make it all go away. But we aren't optimizing by default, we are trying to solve the general case. Calling temp = genlocal() we get a new location (loc:96) to store index * sizeof(element)

    	MUL	loc:96,const:0,const:1
    
    We want to then add that to the base address, but
    	ADD	loc:104,loc:0,loc:96
    
    would add the (word) contents of s[0-7]. Instead, we need
    	ADDR	loc:104,loc:0
    	ADD	loc:104,loc:104,loc:96
    
    After all this, loc:104 contains...the address we want to use.
    	DEREF1	loc:112,loc:104
    
    fetches (into word at loc:112) the value of s[0].

    A label L5 needs to be prefixed into the front of this:

    	LABEL	lab:5
    

    Note: an alternative to ADDR would be to define opcodes for reading and writing arrays. For example

    	SUBSC1   dest,base,index
    
    might be defined to read from base[index] and store the result in dest. Similar opcodes for ASNSUB1, SUBSC8, and ASNSUB8 could be added that assign to base[index], and to perform these operations for 8-byte elements. Even if you do this, you may need the more general ADDR instruction for arrays of arbitrary sized elements.

    CCON^D has no .icode, but the != operator has to generate code to jump to its .true (L4) or .false (L2) as in the previous case. Question: do we need to have a separate TAC instruction for char compares, or sign-extend these operands, or what? I vote: separate opcode for byte operations. BNEC is a "branch if not-equal characters" instruction.

    	BNEC loc:112,const:4,lab:4
    	GOTO lab:2
    
    The code for the entire local_and_expr is concatenated from its children:
    	ADDR   loc:72,loc:0
    	PARAM  loc:72
    	CALL   readline,1,loc:80
    	BNE    loc:80,const:0,lab:5
    	GOTO   lab:2
    	LABEL  lab:5
    	MUL    loc:96,const:0,const:1
    	ADDR   loc:104,loc:0
    	ADD    loc:104,loc:104,loc:96
    	DEREF  loc:112,loc:104
    	BNEC   loc:112,const:4,lab:4
    	GOTO   lab:2
    
    Tree traversal then moves over into the body of the while loop: its statements.

    IDENTi has no .icode. The code for atoi(s) looks almost identical to that for readline(s). The assignment to i tacks on one more instruction:

    	ADDR   loc:120,loc:0
    	PARAM  loc:120
    	CALL   atoi,1,loc:128
    	ASN    loc:64,loc:128
    
    For the second statement in the while loop, the IF statement, there is the usual conditional-followed-by-unconditional branch, the interesting part is where they go. The E.true should do the then-part (the break statement) for which we generate a .first of lab:6. The E.false should go for whatever instruction follows the if-statement, for which lab:3 has been designated.
    	BLE    loc:64,const:0,lab:6
    	GOTO   lab:3
    
    The then-part is a break statement. All then-parts will need to have a label for their .first instruction, which in this case is a trivial GOTO, but where does it go?
    	LABEL  lab:6
    	GOTO   ??
    
    The break is a major non-local goto that even the parent node (the if-statement) cannot know the target for, without obtaining it from about 7 tree-nodes higher! The iteration_statement's .follow (lab:2) is the target for break (its .first would be the target for continue).

    Dr. J has Doubts About 64-bit Ints

    Back to the TAC-or-die example

    So by one of options #1-3, we find the nearest enclosing iteration_statement's .follow field says LAB:2. Note that since we have here a label target that is itself a GOTO, an optimizer would chase back to the branch instructions that go to label 6, and have them go to label 2, allowing us to remove this instruction. By the way, if there were an else statement, the code generation for the then-part would include another GOTO (to skip over the else-part) that we'd hopefully remove in optimization.
    	LABEL  lab:6
    	GOTO   lab:2
    
    Having completed the then part, it is time to assemble the entire if-statement:
    	BLE    loc:64,const:0,lab:6
    	GOTO   lab:3
    	LABEL  lab:6
    	GOTO   lab:2
    	LABEL  lab:3
    
    The next statement is a printf statement. We need to push the parameters onto the stack and execute a call instruction. The code will be: code to evaluate the parameters (which are non-empty this time), code to push parameters (in the correct order, from their .addr values), then the call. Question: does it matter whether the evaluations all occur before the PARAM instructions, or could they (should they) be interleaved? Answer: in C++ evaluations must all occur before the PARAM instructions, all PARAM instructions for a call come after the code for evaluating those arguments, IN REVERSE ORDER, and right before the CALL instruction.

    The code for parameter 1 is empty; its string address will be pushed onto the stack when we get to that part. Here is the code for parameter 2, storing the return value in a new temporary variable.

    	PARAM  loc:64
    	CALL   fib,1,loc:136
    
    The code for the outer call is then
    	PARAM  loc:64
    	CALL   fib,1,loc:136
    	PARAM  loc:136
    	PARAM  sconst:0
    	CALL   printf,2,loc:144
    
    Given this, whole while-loop's code can finally be assembled. The while prepends a label and appends a GOTO back to the while loop's .first field. The whole function's body is just this while loop, with a procedure header and a return statement at the end:
    proc main,0,128
    	LABEL  lab:1
    	ADDR   loc:72,loc:0
    	PARAM  loc:72
    	CALL   readline,1,loc:80
    	BNE    loc:80,const:0,lab:5
    	GOTO   lab:2
    	LABEL  lab:5
    	MUL    loc:96,const:0,const:1
    	ADDR   loc:104,loc:0
    	ADD    loc:104,loc:104,loc:96
    	DEREF  loc:112,loc:104
    	BNEC   loc:112,const:4,lab:4
    	GOTO   lab:2
    	ADDR   loc:120,loc:0
    	PARAM  loc:120
    	CALL   atoi,1,loc:128
    	ASN    loc:64,loc:128
    	BLE    loc:64,const:0,lab:6
    	GOTO   lab:3
    	LABEL  lab:6
    	GOTO   lab:2
    	LABEL  lab:3
    	PARAM  loc:64
    	CALL   fib,1,loc:136
    	PARAM  loc:136
    	PARAM  sconst:0
    	CALL   printf,2,loc:144
    	GOTO   lab:1
    	LABEL  lab:2
    	RETURN
    

    Intermediate Code Generation for Structs, Classes and OO

    Consider the following simplest possible OO class example program:

    class pet {
         int happy
          pet() { happy = 50 }
          void play() {
            write("Woof!\n")
    	happy += 5
    	}
    }
    int main()
    {
        pet pet1
        pet1.play()
        return 0
    }
    
    What are the code generation issues?











    Did we get: For what its worth, one VGo test case is basically a hand-translation of this into VGo using a struct.

    Object Allocation

    memory allocation of an object is similar to other types.
    it can be in the global, local (stack-relative) or heap area
    the # of bytes (size) of the object must be computed from the class.
    each symbol table should track the size of its members
    for a global or local object, add its byte-count size requirement to its containing symbol table / region.
    effectively, no separate code generation for allocation
    translate a "new" expression into a malloc() call...
    plus for all types of object creation, a constructor function call has to happen.

    Initialization via Constructor

    Method Invocation

    Now let's discuss how to generate code for

    o.f(arg1,...,argN)

    Member variable references

    inside a member function, i.e. access member variable x.
    Handle like with arrays, by allocating a new temporary variable in which to calculate the address of this->x. Take the address in the "this" variable and add in x's offset as given in the symbol table for this's class.
    outside an object, o.x.
    Handle as above, using o's address instead of this's. You would also check o's class to make sure x is public.

    Code Generation for Dynamic OO Languages













    Your brilliant suggestions should have included: insert function pointers for all methods into the instance.

    Now let's consider a simple real-world-ish example. Class TextField, a small, simple GUI widget. A typical GUI application might have many textfields on each of many dialogs; many instances of this class will be needed.

    The source code for TextField is only 767 lines long, with 17 member variables and 47 member functions. But it is a subclass of class Component, which is a subclass of three other classes...by the time inheritance is resolved, we have 44 member variables, and 149 member functions. If we include function pointers for all methods in the instance, 77% of instance variable slots will be these function pointers, and these 77% of the slots will be identical/copied for all instances of that class.

    The logical thing to do is to share a single copy of the function pointers, either in a "class object" that is an instance of a meta-class, or more minimally, in a struct or array of function pointers that might be called (by some) a methods vector.

    Methods Vectors

    Suppose you have class A with methods f(), g(), and h(), and class B with methods e(), f(), and g(). Suppose further that you have code that calls method f(), that is designed to work either either A or B. This might happen due to polymorphism, interfaces, subclassing, virtual methods, etc. The kicker will be that in order to generate code for o.f(), a runtime lookup will be performed to obtain the function/method pointer associated with symbol f. Instead a separate structure (the "methods vector") is allocated and shared by all the instances of a given class. In this case, o.f() becomes o.__methods__.f(o)

    Old Mailbag

    What if my program has three functions, each with a local variable to declare?
    Each function's "local region" is allocated uniquely each time they are called. Each of your functions' local regions starts at offset 0 so all three local variables might say LOC:0 for local region offset zero. And yet, they never refer to the same memory, because they are always offsets relative to some base pointer register on the stack.
    When do I allocate my labels? When do I use them?
    You allocate them in one or more tree traversals prior to starting the main traversal that generates the linked lists of 3-address code. Most labels are allocated very close to where they are used. You use labels by generating pseudo-instructions in the linked list AND by filling in the target addresses used by goto instructions with LAB:#N for label number N.
    I am confused about how to access class members via the "this" pointer. I am unsure how to do the offsets from the "this" pointer in three address code without creating a new symbol table for class instances.
    An object instance is like its own little memory region. The this pointer is a parameter; offsets relative to what it points at are done via pointer arithmetic. Each class should indeed have a symbol table for its member variables' offsets.
    Do you have an example that uses each of the pseudo instructions (global, proc, local, label, and end), so we know how these should be formatted?
    No. The pseudo instructions should have opcodes and three address fields; their presence in the linked list of three address codes is the same as an instruction. Their format when you print them out is not very important since this is just intermediate code. But: instructions are on a single line that begins with a tab character, and pseudo instructions are on a single line that does not begin with a tab character.
    We have const that can hold an int/(int)char/boolean, a string region for holding a string offset, but what should we do about double const values?
    Real number constants have to be allocated space similar to other types. They could either be allocated out of a separate "real number constant region", or the constants of different types could all be allocated out of the same region, with different offsets and sizes as needed. Note that not all integer constants fit in instructions, so potentially some of them may have to be allocated as static data also.
    Do the linked lists of code really just get concatenated in order until the entire program is one big linked list? Does main() have to be at the beginning in generated code?
    Not really, and not really. It is recommended that you build one big linked list for the generated code, but I am a pragmatist; do what makes sense to you to generate all the code. In real native OS'es, the code at the beginning of the executable is not main, it is some weird startup boostrapper that sets up environment and command line arguments and then calls main(). So no, main() does not have to be at the top, unless maybe you are building an image for an embedded system that doesn't have an OS or something like that.
    How do I represent function names in addresses in 3 address code?
    One option is to totally dodge, and generate code for one function at a time, at a place where you know the function name. If you choose instead to build one big linked list, function "names" get boiled down to code region addresses. So far we have one kind of address in the code region: labels. You could literally generate label #'s for these things, but function names are more human-friendly. Unless you turn function names into labels, you should create a new region (call it PROCNAME). You could make the "offset" field in your 3 addresses a union
         struct addr {
          int region; // if PROCNAME, use u.s instead of u.i
          union {
             int offset;
    	 char *s;
             } u;
          }
    
    You could, instead, make an array of string funcnames in your compiler and have your region PROCNAME provide offsets that are subscripts into this array of funcnames.
    I am having a hard time understanding how everything will be put together in the end, will it be one linked list once all the instructions are concatenated? How should we handle assigning locations to functions like Println? Once we see import "fmt" should we go to that symbol table and assign locations to those functions then?
    Library functions like Println require that we store enough information to call them, but not that we store information to generate code for them. fmt should have an associated symbol table entry for Println which should know that Println takes a string argument. Code for a call to fmt.Println should mangle that name out to something like fmt__Println.
    can we just define a function without parameters as call, so main is equivalent to main() if not followed by parentheses?
    Do not confuse type (reference to) FUNCTION with the function's return type, which is the result of a call (postfix parentheses operator).

    Where we are at

    Schedule-wise, it says it is time to move lectures on into the area of final code generation. We have a new wrinkle there this year, and need to allow some time for it. But, intermediate code generators are not due yet, and questions about that are still totally fair game.

    One More Intermediate Code Example, with Feeling

    I glossed over the "TAC or die trying" example in the last lecture; it felt redundant for me since we did an earlier example that was similar. However, one or more of you have requested another intermediate code generation example.

    Yeah, we'll do one alright; this week. But it will take a bit more preparation, so: not today. This weekend I spent a fair bit digging into another code generation topic, namely LLVM, and we will also be talking about that.

    Mailbag

    Remind me again: What is the MINIMUM I have to do to pass this class
    That depends on exam results, but generally students need either good exam results and at least get through type checking, or less good exam results and at least some/partial intermediate code to "pass". This semester folks are behind where we are some other semesters, and that may help you, since you are graded relative to your peers. A grade extension might be an option. I am pretty darn OK with grade extension if you have been working at it but need more progress. You should check whether it would prevent you from walking in commencement, but they seem to have relaxed NMT's traditional hard-core commencement-walking rules lately.
    I have been working 12 days a week, 7 hours per day on this homework and it just isn't fair.
    Some of you are probably working more efficiently than others ("work smarter"), but the goal was not to teach you what the death marches in industry are going to be like. The goal is to cause you to increase your skill level in programming and/or software engineering.
    Where is my HW#5 feedback?
    You made it to a HW#5 submission? Congratulations! I spent quality time this weekend on regrades of HW#2-4 and did the groundwork for grading HW#5. I will get right on that.
    I'm having trouble figuring out what TAC I need to generate for a function definition. For example, given the function
    int foo(int x){
       ...somecode
    }
    I'm having trouble understanding what code needs to be generated at this level. I understand that there needs to be (at least) 1 label, at the very start (to be able to call the function).
    In final code, the procedure entry point will indeed include a label. In three address code, a function header should result in a PROC pseudo-instruction for which you create a link list element, just like everything else.
    I'm having trouble understanding what code I would create for the int return, or to define the space available for parameters.
    The "return type" at the top of a function generates no code, but it may affect what you generate when you hit a "return" statement in the function body.

    The proc pseudoinstruction includes a declaration of how many (words of parameters) it requires/assumes has been passed in to a function, from which space required may be calculated. In most native code the caller allocates this space; the called function just decides the amount of local/temp variable space on the stack that the procedure requires. So the pseudoinstructions in intermediate code that you use is something like:

    proc foo,1,nbytes_localspace
    

    This question gets more interesting if you have the ability to return multi-word values such as a struct, for which a register will not suffice! What do you think a compiler should do in that case?

    If I understand the return properly, I don't actually generate code at this (the procedure header return type) node for the return. It gets generated at the return statement in the body.
    Yes. There and at the end of any function that falls off the end. In final code the return statement will put a return value in %eax and then jump down to the end of the function to use its proper function-return assembler instruction(s).
    I guess the .addr of a parameter int x is what is really getting me. Do I really need to worry about it too much in TAC, because it is just 'local 0' (or whatever number gets generated)?
    I recommend you consider it (in TAC) to be region PARAM offset 0. That could be handled almost identically to locals in final code, unless you use the fact that parameters are passed in registers...
    Then I really end up worrying about it during final code since local 0 might actually be something like %rbp -1 or wherever the location on the stack parameters end up being located.
    If you make a separate region for parameters, by saying a variable is is PARAM region offset 0, the TAC code for parameters is distinct from locals. Parameters can be found to be at a different location relative to the %rbp (positive instead of negative offsets) or passed in registers.

    A New Fun Intermediate CodeGen Example

    This example is not burdened with redundant practice at generating code for arithmetic and assignments and such.
    package main
    import "fmt"
    func min(a,b,c int) int {
      if a<=b && a<=c {
         return a
      } else if b<=a && b<=c {
         return b
      }
      return c
    }
    
    func main() {
      fmt.Println(min(3,6,2))
    }
    

    lecture #25 began here

    Final Code Generation

    Alternatives for Final Code:
    interpret the source code
    we could build an interpreter instead of a compiler, in which the source code was kept in string or token form, and re-parsed, possibly repeatedly, during execution. Some early BASICs and operating system shell scripting languages do this, but it is Really Slow.
    interpret the parse tree
    we could write an interpreter that executes the program by walking around on the tree doing traversals of various subtrees. This is still slow, but successfully used by many "scripting languages".
    interpret the 3-address code
    we could interpret the link-list or a more compact binary representation of the intermediate code
    translate into VM instructions
    popular virtual machines such as JVM or .Net allow execution from an instruction set that is often higher level than hardware, may be independent of the underlying hardware, and may be oriented toward supporting the specific language features of the source language. For example, there are various BASIC virtual machines out there.
    translate into "native" instructions
    "native" generally means hardware instructions.
    For practical purposes, we will consider only two of these options
    1. translate into VM assembler for the LLVM IR, or
    2. translate into native x86_64 <<-- "simplest"

    Native Code Generation

    In mainstream compilers, final code generation into native code
    1. takes a linear sequence of 3-address intermediate code instructions, and
    2. translates each 3-address instruction into one or more native instructions.

    The big issues in code generation are:

    (a) instruction selection, and
    (b) register allocation and assignment.

    Registers

    Perhaps the biggest difference between intermediate code and final code are the registers. Their use is crucial to performance!

    Collecting Information Necessary for Final Code Generation

    Option #A: a top-down approach to learning your native target code.
    Study a reference work supplied by the chip manufacturer, such as the AMD64 Architecture Programmer's Manual (Vol. 2, Vol. 3).
    Option #B: a bottom-up (or reverse engineering) approach to learning your native target code.
    study an existing compiler's native code. For example, run "g++ -S" for various toy programs to learn native instructions corresponding to each expression, particularly ones equivalent to the various 3-address instructions.

    Instruction Selection

    A modern CPU usually has many different sequences of instructions that it could use to accomplish a given task. Instruction selection must choose a particular sequence. Given a choice among equivalent/alternative sequences, the decision on which sequence of instructions to use is usually based on estimates or measurements of which sequence executes the fastest.

    A good set of examples of instruction selection are to be found in the superoptimizer paper. From that paper:

    Register Allocation and Assignment

    The (register allocation) job changes as CPUs change

    Even if an instruction set does support memory-based operations, most compilers should load a value into a register while it is being used, and then spill it back out to main memory when the register is needed for another purpose. The task of minimizing memory accesses becomes the task of minimizing register loads and spills.

    Mailbag

    I'm starting to come across certain 3-address instructions that we haven't seen examples of yet, mainly the ones for branches and conditionals. Would we be able to talk about some examples in class possibly? If not, could I get an example for something like the BLT instruction in an if statement?
    BLT stands for Branch-if-less-than. It takes two operands compares them and if src1 is less than src2, it jumps the instruction pointer to the destination (the 3rd address is a label, i.e. a code region offset). It is kind of interesting to compare this with X86_64.
    IntermediateFinal
    	BLT loc:16,const:5,lab:32
    
    	cmpq $5,-24(%rbp)
    	jg .L32
    	... code for a then-part
    .L32:
    

    Native Code Generation Examples

    Reusing a Register

    Consider the statement:
       a = a+b+c+d+e+f+g+a+c+e;
    
    A naive three address code generator would generate a lot of temporary variables here, one per addition operator, when in actuality one big number is being added. How many registers does the expression need? Some variables are referenced once, some twice. GCC (32-bit) generates:

    	movl	b, %eax
    	addl	a, %eax
    	addl	c, %eax
    	addl	d, %eax
    	addl	e, %eax
    	addl	f, %eax
    	addl	g, %eax
    	addl	a, %eax
    	addl	c, %eax
    	addl	e, %eax
    	movl	%eax, a
    

    Just for fun, the corresponding .u file, with filen/line/column/synt pseudoinstructions removed, and comments added

    Now consider

       a = (a+b)*(c+d)*(e+f)*(g+a)*(c+e);
    
    How many registers are needed here?
    	movl	b, %eax
    	movl	a, %edx
    	addl	%eax, %edx
    	movl	d, %eax
    	addl	c, %eax
    	imull	%eax, %edx
    	movl	f, %eax
    	addl	e, %eax
    	imull	%eax, %edx
    	movl	a, %eax
    	addl	g, %eax
    	imull	%eax, %edx
    	movl	e, %eax
    	addl	c, %eax
    	imull	%edx, %eax
    	movl	%eax, a
    
    And now this:
       a = ((a+b)*(c+d))+((e+f)*(g+a))+(c*e);
    
    which compiles to
    	movl	b, %eax
    	movl	a, %edx
    	addl	%eax, %edx
    	movl	d, %eax
    	addl	c, %eax
    	movl	%edx, %ecx
    	imull	%eax, %ecx
    	movl	f, %eax
    	movl	e, %edx
    	addl	%eax, %edx
    	movl	a, %eax
    	addl	g, %eax
    	imull	%edx, %eax
    	leal	(%eax,%ecx), %edx
    	movl	c, %eax
    	imull	e, %eax
    	leal	(%eax,%edx), %eax
    	movl	%eax, a
    

    Mailbag

    What will HW#7 be graded on?
    A typical jeffery HW#7 rubric might look like:
    .zip Compiles w/no warnings: __/3
    Valgrind is clean: __/4
    -s: writes .s|.tac.c files: __/5
    -c: writes .o files: __/5
    fec produces executables: __/5
    Generates final code/exprs: __/7
    Generates final code/gotos: __/7
    Generates final code/funcs: __/7
    Generates final code/decls: __/7
    Total: __/50
    What constitutes "full credit" for HW#7

    Comparison of 32-bit and 64-bit x86 code

    What can be gleaned from this side-by-side of 32-bit and 64-bit assembler for a=a+b+c+d+e+f+g+a+c+e. Note that the actual variable names are in the assembler because the variables in question are globals.

    x86 32-bit x86_64
    	movl	b, %eax
    	addl	a, %eax
    	addl	c, %eax
    	addl	d, %eax
    	addl	e, %eax
    	addl	f, %eax
    	addl	g, %eax
    	addl	a, %eax
    	addl	c, %eax
    	addl	e, %eax
    	movl	%eax, a
    
    	movq	a(%rip), %rdx
    	movq	b(%rip), %rax
    	addq	%rax, %rdx
    	movq	c(%rip), %rax
    	addq	%rax, %rdx
    	movq	d(%rip), %rax
    	addq	%rax, %rdx
    	movq	e(%rip), %rax
    	addq	%rax, %rdx
    	movq	f(%rip), %rax
    	addq	%rax, %rdx
    	movq	g(%rip), %rax
    	addq	%rax, %rdx
    	movq	a(%rip), %rax
    	addq	%rax, %rdx
    	movq	c(%rip), %rax
    	addq	%rax, %rdx
    	movq	e(%rip), %rax
    	leaq	(%rdx,%rax), %rax
    	movq	%rax, a(%rip)
    

    Q: Should we be disappointed that the 64-bit code looks a lot longer?

    A: Maybe instead we should be fascinated.

    The globals are declared something like the following.

    If you allocated your globals as a region, you might have one .comm of 56 bytes named globals (or whatever) and give the addresses of your globals as numbers such as globals+32. Names are nicer but having to treat globals and locals very differently is not.

    	.comm	a,8,8
    	.comm	b,8,8
    	.comm	c,8,8
    	.comm	d,8,8
    	.comm	e,8,8
    	.comm	f,8,8
    	.comm	g,8,8
    	.text
    .globl main
    	.type	main, @function
    

    Brief Comparison of x86-64 globals vs. locals

    How does this difference inform, and affect, what we might want in our three-address code?

    x86_64 local vars x86_64 globals (as per last example)
    	movq	-48(%rbp), %rax
    	movq	-56(%rbp), %rdx
    	leaq	(%rdx,%rax), %rax
    	addq	-40(%rbp), %rax
    	addq	-32(%rbp), %rax
    	addq	-24(%rbp), %rax
    	addq	-16(%rbp), %rax
    	addq	-8(%rbp), %rax
    	addq	-56(%rbp), %rax
    	addq	-40(%rbp), %rax
    	addq	-24(%rbp), %rax
    	movq	%rax, -56(%rbp)
    
    	movq	a(%rip), %rdx
    	movq	b(%rip), %rax
    	addq	%rax, %rdx
    	movq	c(%rip), %rax
    	addq	%rax, %rdx
    	movq	d(%rip), %rax
    	addq	%rax, %rdx
    	movq	e(%rip), %rax
    	addq	%rax, %rdx
    	movq	f(%rip), %rax
    	addq	%rax, %rdx
    	movq	g(%rip), %rax
    	addq	%rax, %rdx
    	movq	a(%rip), %rax
    	addq	%rax, %rdx
    	movq	c(%rip), %rax
    	addq	%rax, %rdx
    	movq	e(%rip), %rax
    	leaq	(%rdx,%rax), %rax
    	movq	%rax, a(%rip)
    

    Parameters

    In final code, do parameters look like locals?

    Consider the following example. Note that "long" is used to more closely resemble our "everything is a 64-bit value" mind-set.

    #include <stdio.h>
    
    long f(long,long,long);
    
    int main()
    {
       long rv = f(1, 2, 3);
       printf("rv is %ld\n", rv);
    }
    
    long f(long a, long b, long c)
    {
       long d, e, f, g;
       d = 4; e = 5; f = 6; g = 7;
       a = ((a+b)*(c+d))+(((e+f)*(g+a))/(c*e));
       return a;
    }
    
    for which the generated code was
    	.file	"paramdemo.c"
    	.text
    	.section	.rodata
    .LC0:
    	.string	"rv is %ld\n"
    	.text
    	.globl	main
    	.type	main, @function
    main:
    .LFB0:
    	.cfi_startproc
    	pushq	%rbp              # save old frame pointer
    	.cfi_def_cfa_offset 16
    	.cfi_offset 6, -16
    	movq	%rsp, %rbp        # set new frame pointer
    	.cfi_def_cfa_register 6
    	subq	$16, %rsp         make 16 bytes (two words) of locals
    	movl	$3, %edx          "push" parameters back to front, first 6 in registers
    	movl	$2, %esi             note implied sign extension
    	movl	$1, %edi
    	call	f                 call normal func
    	movq	%rax, -8(%rbp)    store return in rv
    	movq	-8(%rbp), %rax    LOL
    	movq	%rax, %rsi        "push" parameter 2
    	leaq	.LC0(%rip), %rdi  "push" parameter 1
    	movl	$0, %eax          vararg # float parms
    	call	printf@PLT        call shared lib func
    	movl	$0, %eax          implicit return value is 0
    	leave                     restore frame pointer
    	.cfi_def_cfa 7, 8
    	ret                       implicit return
    	.cfi_endproc              end pseudo-instruction
    .LFE0:                            end of function label 0
    	.size	main, .-main      end pseudo-instruction
    	.globl	f           
    	.type	f, @function
    f:
    .LFB1:
    	.cfi_startproc
    	pushq	%rbp
    	.cfi_def_cfa_offset 16
    	.cfi_offset 6, -16
    	movq	%rsp, %rbp        frame pointer but no local allocation!?!
    	.cfi_def_cfa_register 6
    	movq	%rdi, -40(%rbp)   pop param 1 to "a"
    	movq	%rsi, -48(%rbp)   pop param 2 to "b"
    	movq	%rdx, -56(%rbp)   pop param 3 to "c"
    	movq	$4, -32(%rbp)     initialize d
    	movq	$5, -24(%rbp)     initialize e
    	movq	$6, -16(%rbp)     initialize f
    	movq	$7, -8(%rbp)      initialize g
    	movq	-40(%rbp), %rdx
    	movq	-48(%rbp), %rax
    	leaq	(%rdx,%rax), %rcx     a+b
    	movq	-56(%rbp), %rdx
    	movq	-32(%rbp), %rax
    	addq	%rdx, %rax            c+d
    	imulq	%rax, %rcx            (a+b)*(c+d)
    	movq	-24(%rbp), %rdx
    	movq	-16(%rbp), %rax
    	leaq	(%rdx,%rax), %rsi     e+f
    	movq	-8(%rbp), %rdx
    	movq	-40(%rbp), %rax
    	addq	%rdx, %rax            g+a
    	imulq	%rsi, %rax            (e+f)*(g+a)
    	movq	-56(%rbp), %rdx
    	movq	%rdx, %rdi
    	imulq	-24(%rbp), %rdi       c*e
    	cqto
    	idivq	%rdi
    	addq	%rcx, %rax
    	movq	%rax, -40(%rbp)       assign a
    	movq	-40(%rbp), %rax       use a as return val
    	popq	%rbp
    	.cfi_def_cfa 7, 8
    	ret
    	.cfi_endproc
    .LFE1:
    	.size	f, .-f
    	.ident	"GCC: (Ubuntu 7.5.0-3ubuntu1~18.04) 7.5.0"
    	.section	.note.GNU-stack,"",@progbits
    

    Aside on .cfi* assembler directives

    lecture #26 began here

    malloc() and a bit of pointer arithmetic

    Consider the following:
    #include 
    
    void main()
    {
      long *l = malloc(144);
      l[5] = 7;
    }
    
    it results in the following, via GCC 11.4 run in 2024. The moral of the story is, to assign into an array element, you make a pointer to that element in some register like %rax and then MOV data into (%rax):
    	.file	"malc.c"
    	.text
    	.globl	main
    	.type	main, @function
    main:
    .LFB6:
    	.cfi_startproc
    	endbr64                   no-op, ctrlflow enforcement
    	pushq	%rbp
    	.cfi_def_cfa_offset 16
    	.cfi_offset 6, -16
    	movq	%rsp, %rbp
    	.cfi_def_cfa_register 6
    	subq	$16, %rsp
    	movl	$144, %edi
    	call	malloc@PLT
    	movq	%rax, -8(%rbp)
    	movq	-8(%rbp), %rax
    	addq	$40, %rax
    	movq	$7, (%rax)
    	nop
    	leave                     restore frame pointer
    	.cfi_def_cfa 7, 8
    	ret
    	.cfi_endproc
    .LFE6:
    	.size	main, .-main
    	.ident	"GCC: (Ubuntu 11.4.0-1ubuntu1~22.04) 11.4.0"
    	.section	.note.GNU-stack,"",@progbits
    	.section	.note.gnu.property,"a"
    	.align 8
    	.long	1f - 0f
    	.long	4f - 1f
    	.long	5
    0:
    	.string	"GNU"
    1:
    	.align 8
    	.long	0xc0000002
    	.long	3f - 2f
    2:
    	.long	0x3
    3:
    	.align 8
    4:
    

    Creating an object via new

    Consider the following C++ example of final code for an object constructor. Executing the reserved word new from function main() calls two functions to create an object in the heap (via new):
    _Znwm
    similar to a malloc(); it takes an integer parameter (constant 16, the # of bytes to allocate) and returns a pointer
    _ZN1CC1Ev
    a call to a C++ constructor function, with an implicit/added first parameter for this, the object instance that the member function is working on.

    "new" in final code FYI

    class C {
      private: long x, y;
      public:  C() { x=3; y=4; }
    };
    
    int main()
    {
       C *a = new C;
    }
    
    generates
    	.file	"new.cpp"
    	.section	.text._ZN1CC2Ev,"axG",@progbits,_ZN1CC5Ev,comdat
    	.align 2
    	.weak	_ZN1CC2Ev
    	.type	_ZN1CC2Ev, @function
    _ZN1CC2Ev:
    .LFB1:
    	.cfi_startproc
    	.cfi_personality 0x3,__gxx_personality_v0
    	pushq	%rbp
    	.cfi_def_cfa_offset 16
    	.cfi_offset 6, -16
    	movq	%rsp, %rbp
    	.cfi_def_cfa_register 6
    	movq	%rdi, -8(%rbp)			// copy "self" to stack
    	movq	-8(%rbp), %rax			// load "self" 
    	movq	$3, (%rax)
    	movq	-8(%rbp), %rax
    	movq	$4, 8(%rax)
    	leave
    	.cfi_def_cfa 7, 8
    	ret
    	.cfi_endproc
    .LFE1:
    	.size	_ZN1CC2Ev, .-_ZN1CC2Ev
    	.weak	_ZN1CC1Ev
    	.set	_ZN1CC1Ev,_ZN1CC2Ev
    	.text
    .globl main
    	.type	main, @function
    main:
    .LFB3:
    	.cfi_startproc
    	.cfi_personality 0x3,__gxx_personality_v0
    	pushq	%rbp
    	.cfi_def_cfa_offset 16
    	.cfi_offset 6, -16
    	movq	%rsp, %rbp
    	.cfi_def_cfa_register 6
    	pushq	%rbx
    	subq	$24, %rsp
    	movl	$16, %edi
    	.cfi_offset 3, -24
    	call	_Znwm
    	movq	%rax, %rbx
    	movq	%rbx, %rax
    	movq	%rax, %rdi
    	call	_ZN1CC1Ev
    .L5:
    	movq	%rbx, -24(%rbp)
    	movl	$0, %eax
    	addq	$24, %rsp
    	popq	%rbx
    	leave
    	.cfi_def_cfa 7, 8
    	ret
    	.cfi_endproc
    .LFE3:
    	.size	main, .-main
    	.ident	"GCC: (GNU) 4.4.7 20120313 (Red Hat 4.4.7-3)"
    	.section	.note.GNU-stack,"",@progbits
    
    As you may observe: the final code for a new calls a memory allocator (nwm) whose return value (%rax) gets copied in as a parameter (%rdi) to the constructor (N1CC1Ev), with an interesting side trip to %rbx.

    On C/C++ Calling Convention and Order of Passed Parameters

    In compilers, the calling conventions are the set of rules by which parameters and return values are communicated between caller and callee. The calling conventions also cover things like whether the caller or the callee has to save and restore specific registers as part of the process of call/return.

    How this looks and is used inside member functions

    About name mangling in C++ vs. your compiler

    More about LEAL

    In a previous example, complicated arithmetic drove (some past version of) GCC to start "leal'ing".

    Lastly (for now) consider:

       a = ((a+b)*(c+d))+(((e+f)*(g+a))/(c*e));
    
    The division instruction adds new wrinkles. It operates on an implicit register accumulator which is twice as many bits as the number you divide by, meaning 64 bits (two registers) to divide by a 32-bit number. Note in this code that 32-bit gcc would rather spill than use %ebx. %ebx is reserved by the compiler for some (hopefully good) reason. %edi and %esi are similarly ignored/not used.
    32-bit 64-bit
    	movl	b, %eax
    	movl	a, %edx
    	addl	%eax, %edx
    	movl	d, %eax
    	addl	c, %eax
    	movl	%edx, %ecx
    	imull	%eax, %ecx
    	movl	f, %eax
    	movl	e, %edx
    	addl	%eax, %edx
    	movl	a, %eax
    	addl	g, %eax
    	imull	%eax, %edx
    	movl	c, %eax
    	imull	e, %eax
    	movl	%eax, -4(%ebp)
    	movl	%edx, %eax
    	cltd
    	idivl	-4(%ebp)
    	movl	%eax, -4(%ebp)
    	movl	-4(%ebp), %edx
    	leal	(%edx,%ecx), %eax
    	movl	%eax, a
    
    	pushq	%rbx
    	subq	$88, %rsp
    	movq	$1, -72(%rbp)
    	movq	$2, -64(%rbp)
    	movq	$3, -56(%rbp)
    	movq	$4, -48(%rbp)
    	movq	$5, -40(%rbp)
    	movq	$6, -32(%rbp)
    	movq	$7, -24(%rbp)
    	movq	-64(%rbp), %rax
    	movq	-72(%rbp), %rdx
    	leaq	(%rdx,%rax), %rcx
    	movq	-48(%rbp), %rax
    	movq	-56(%rbp), %rdx
    	leaq	(%rdx,%rax), %rax
    	imulq	%rax, %rcx
    	movq	-32(%rbp), %rax
    	movq	-40(%rbp), %rdx
    	leaq	(%rdx,%rax), %rbx
    	.cfi_offset 3, -24
    	movq	-72(%rbp), %rax
    	movq	-24(%rbp), %rdx
    	leaq	(%rdx,%rax), %rax
    	imulq	%rbx, %rax
    	movq	-56(%rbp), %rdx
    	movq	%rdx, %rbx
    	imulq	-40(%rbp), %rbx
    	movq	%rbx, -88(%rbp)
    	movq	%rax, %rdx
    	sarq	$63, %rdx
    	idivq	-88(%rbp)
    	leaq	(%rcx,%rax), %rax
    	movq	%rax, -72(%rbp)
    	movl	$.LC0, %eax
    	movq	-72(%rbp), %rdx
    	movq	%rdx, %rsi
    	movq	%rax, %rdi
    	movl	$0, %eax
    	call	printf
    	addq	$88, %rsp
    	popq	%rbx
    
    In the 32-bit version, you finally see some register spilling. In the 64-bit version, there is

    LEAVE instruction

    In our example of using new we saw a LEAVE instruction before the function returned. LEAVE restores the frame pointer to the caller's value, something like
    movq rsp, rbp ; set top of stack back to where caller had it
    popq rbp      ; set base pointer back to saved value at (%rsp)
    
    Interestingly, there is a corresponding ENTER instruction, but g++ does not tend to use it because it is slower than corresponding lower-level operations like subq $nbytes, %rsp.

    Chapter 10 of this free compiler book from Doug Thain

    For what its worth, professor Thain provides a good introduction to x86_64 assembler.

    Brief Comment on HW Resubmissions

    At various points in this course you have a choice between completing/fixing a previous homework, or working on the next homework. But sometimes you have to complete/fix an old homework for the next one to be implementable. I have been accepting resubmissions this semester, to make corrections, restoring points up to a "C" grade for a given assignment. Please test your work prior to each resubmission; I won't be able to just keep regrading it until it passes.

    More on DIV instruction

    When I looked for more, I found this Cheat Sheet, which pointed at the big books (A-MN-Z).

    Helper Function for Managing Registers

    Define a getreg() function that returns a location L to hold the value of x for the assignment
    x := y op z
    1. if y is in a register R that holds the value of no other names, AND y is not live after the execution of this instruction, THEN return register R as your location L
    2. ELSE return an empty register for L if there is one
    3. ELSE if x has a next use in the block or op is an operator that requires a register (e.g. indexing), find an occupied register R. Store R into the proper memory location(s), update the address descriptor for that location, and return R
    4. if x is not used in the block, or no suitable occupied register can be found in (3), return the memory location of x as L.

    lecture #27 began here

    Next Tuesday

    How to Compute your Pre-Final CSE 423 Grade

    1. Each Graded Item (HW, Exam, Sum of Labs) is Normalized in the range [0.0,1.0] Against the High Scores obtained by a class member for that assignment:
      Graded ItemMax
      HW#1 4
      HW#2 20
      HW#3 26
      Midterm 295
      HW#4 27
      HW#5 25
      HW#6 23
      HW#7 ??
      Final ?
      Labs 30
    2. Each normalized Graded Item is Multiplied Against a Weight. The weights are percentages of the total grade. In the end they will add to 100.
      Graded ItemWeight
      HW#1 1
      HW#2 6
      HW#3 7
      Midterm 25
      HW#4 7
      HW#5 8
      HW#6 8
      HW#7 9
      Final 25
      Labs 4
    3. Weighted scores are Summed. For each gradable item, you got some fraction of the total weight that that item was worth.
    4. Sum is divided by total weight and multiplied by 100 to give your overall percentage against an "ideal peer".
    5. Jeffery subjectively decides whether the high "A" is, and how far to draw the ABCD lines 10% apart, 12% apart, 15% apart, 20%, etc. For this semester, suppose the "A" line is at 90% of the "ideal peer", the grades are 15% apart.
      Overall PercentageGrade
      90+ A
      85+ A-
      80+ B+
      75+ B
      70+ B-
      65+ C+
      60+ C
      55+ C-
      50+ D+
      45+ D
      40+ D-
    Example: at the moment, 68% of the total weights have been graded. A student computes a weighted sum of 47, 47/68 = 69%. At the moment, this student is on the borderline between a C+ and a B-.

    Putting It All Together: A Simple Code Generator

    Example

    // make an array (12?) of these:
    struct regdescrip {
       char name[12]; // name to use in codegen, e.g. "%rbx"
       int status;    // 0=empty, 1=loaded, 2=dirty, 4=live, ...
       struct address addr;
       };
    // upgrade symbol table entry to use these instead of struct address
    struct addr_descrip {
       int status;    // 0=empty, 1=in memory, 2=in register, 3=both
       struct reg_descrip *r; // point at an elem in reg array. could use index.
       struct address a;
       };
    

    Code Generation Algorithm

    For each three-address statement of the form
    x := y op z:
    1. Use getreg() to determine location L where the result of the computation y op z should be stored.
    2. Use the address descriptor for y to determine y', a current location for y. If y is currently in a register, use the register as y'. If y is not already in L, generate the instruction MOV y',L to put a copy of y in L.
    3. Generate the instruction OP z',L where z' is a current location for z. Again, prefer a register location if z is currently in a register.

      Update the descriptor of x to indicate that it is in L. If L is a register, update its descriptor to indicate that it contains x. Remove x from all other register descriptors.

    4. If y and/or z have no next uses and are in registers, update the register descriptors to reflect that they no longer contain y and/or z respectively.

    Register Allocation

    Need to decide:

    Approaches to Register Allocation

    1. Partition the register set into groups that are use for different kinds of values. E.g. assign base addrs to one group, pointers to the stack to another, etc.

      Advantage: simple
      Disadvantage: register use may be inefficient

    2. Keep frequently used values in registers, across block boundaries. E.g. assign some fixed number of registers to hold the most active values in each inner loop.

      Advantage: simple to implement
      Disadvantage: sensitive to # of registers assigned for loop variables.

    Challenge Question we Ended with Last Time

    So, how can you know what are the frequently used variables in a function?








    x86_64 Floating Point

    Float Operations

    There is a useful set of notes from Andrew Tolmach of Portland State University. Arithmetic operations on floats have different opcodes, and results have to be stored in floating point registers, not integer registers.
    	movsd	-56(%rbp), %xmm0
    	movapd	%xmm0, %xmm1
    	addsd	-48(%rbp), %xmm1
    

    Float Constants

    Doubles are the same 64-bit size as longs. They can be loaded into memory or registers using the normal instructions like movq. A spectacular x86_64 opcode named movabsq takes an entire floating point constant as an immediate (bit pattern given as a decimal integer!) and stores it in a register. (Q: What C code (or library function) would take your double and produce the equivalent decimal integer string?)
    	movabsq	$4620355447710076109, %rax
    	movq	%rax, -8(%rbp)
    

    Simple Machine Model

    This model is probably relevant for selecting between equivalent sequences of instructions but is presented here as food for thought regarding which variables deserve to stay in registers.
    Instruction Costs
    for an instruction I, cost(I) = 1 + sum(cost(operands(I)))

    operand costs:


    Usage Counts
    In this model, each reference to a variable x accrues a savings of 1 if x is in a register.

    Cost savings flow graph example

    For the following flow graph, how much savings would be earned by leaving variables (a-f) in a register across basic blocks?

    Savings B1 B2 B3 B4 Total
    a 2 1 1 0 4
    b 2 0 2 2 6
    c 1 0 1 1 3
    d 3 1 1 1 6
    e 2 0 2 0 4
    f 1 2 1 0 4

    x86_64 Discussion

    For what its worth on Windows 64

    Warning: the Mingw64 compiler (and possibly other Windows 64-bit c compilers) do not use the same memory sizes as Linux x86_64! Beware. If you were compatible with gcc on Linux you might not be on Windows and vice versa.

    Three Kinds of Dependence

    In all three of these examples, a dependence relationship implies that in the program semantics, the second instruction depends on the first one in some way.

    a = b + c;
    ...
    d = a + e;
    
    a = b + c;
    ...
    b = d + e;
    
    a = b + c;
    ...
    a = d + e;
    

    Final Code Generation Example

    Lessons From the Final Code Generation Example

    Reverse Engineering, gcc -S, and Optimization

    I decided to fill in a missing piece of the x86_64 final code generation template page that I am providing you, and chose a real easy one: if !x goto L. I figured it would be a two-instruction analogue of if x goto L. So I constructed a simple program to try and produce the desired code.
    #include 
    int fac(long y)
    {
       long x;
       if (!y) goto L;
       printf("hello");
     L:
       return 1;
    }
    

    I was frustrated to find seemingly idiotic code as gcc's default: it was generating an extra jump and an extra label. Eventually, I tried it with -O just to see what we would get.

    The corresponding gcc -S output is as follows:

    gcc -Sgcc -O -S
    	.file	"foo.c"
    	.section	.rodata
    .LC0:
    	.string	"hello"
    	.text
    	.globl	fac
    	.type	fac, @function
    fac:
    .LFB0:
    	.cfi_startproc
    	pushq	%rbp
    	.cfi_def_cfa_offset 16
    	.cfi_offset 6, -16
    	movq	%rsp, %rbp
    	.cfi_def_cfa_register 6
    	subq	$16, %rsp
    	movq	%rdi, -8(%rbp)
    	cmpq	$0, -8(%rbp)
    	jne	.L2
    	jmp	.L3
    .L2:
    	movl	$.LC0, %edi
    	movl	$0, %eax     # num of float args, for vararg funcs
    	call	printf
    .L3:
    	movl	$1, %eax
    	leave
    	.cfi_def_cfa 7, 8
    	ret
    	.cfi_endproc
    .LFE0:
    	.size	fac, .-fac
    	.ident	"GCC: (GNU) 4.8.5 20150623 (Red Hat 4.8.5-28)"
    	.section	.note.GNU-stack,"",@progbits
    
    	.file	"foo.c"
    	.section	.rodata.str1.1,"aMS",@progbits,1
    .LC0:
    	.string	"hello"
    	.text
    	.globl	fac
    	.type	fac, @function
    fac:
    .LFB11:
    	.cfi_startproc
    	testq	%rdi, %rdi
    	je	.L4
    	subq	$8, %rsp
    	.cfi_def_cfa_offset 16
    	movl	$.LC0, %edi
    	movl	$0, %eax     # num of float args, for vararg funcs
    	call	printf
    .L2:
    	movl	$1, %eax
    	addq	$8, %rsp
    	.cfi_def_cfa_offset 8
    	ret
    .L4:
    	movl	$1, %eax
    	ret
    	.cfi_endproc
    .LFE11:
    	.size	fac, .-fac
    	.ident	"GCC: (GNU) 4.8.5 20150623 (Red Hat 4.8.5-28)"
    	.section	.note.GNU-stack,"",@progbits
    

    Flow Graphs

    In preparation for lectures discussing code optimization, a more detailed discussion of flow graphs is needed.

    Flow Graph Example

    if (x + y <= 10 && x - y >= 0) x = x + 1;
    
    Construct the flow graph from the basic blocks

    t1 := x + y
    if t1 > 10 goto L1
    
    
    t2 := x - y
    if t2 < 0 goto L1
    
    
    t3 := x + 1
    x := t3
    
    
    L1:
    
    

    Next-Use Information

    use of a name
    consider two statements
    I1: x := ... /* assigns to x */
    ...
    I2: ... := ... x ... /* has x as an operand */
    
    such that control can flow from I1 to I2 along some path that has no intervening assignments to x. Then, I2 uses the value of x computed at I1. I2 may use several assignments to x via different paths.
    live variables
    a variable x is live at a point in a flow graph if the value of x at that point is used at a later point.

    Computing Next-Use Information (within a block only)

    lecture #28 began here

    lecture #29 began here

    lecture #30 began here

    Old Mailbag

    Could we cover assembly stack allocation/management?
    Sure. I've pointed you at a lot of resources; here is another, on Eli Bendersky's site.

    When and what do we have to push and pop from the stack when we call a function?
    We (that mens you) should probably look at a bunch of examples, probably by reverse engineering them with gcc -S, to get a feel for this. A summary on which we can expand/correct is:
    Can we use the stack exclusively for all of our parameters and local variables?
    Your compiler can ignore register parameters entirely when you generate code that calls to and returns from your own functions. IFF your code needs to call C library code (such as printf, reads, etc.) you would have to use the standard calling conventions (including registers) to call those functions successfully.

    Storage for Temporaries

    Storage for Temporaries Example

    Consider the following (a dot-product code) example. This is a single basic block, subdivided using the liveness of some non-overlapping temporary variables.

    t1 live
    	prod := 0
    	i := 1
    L3:	t1 := 4 * i
    	t2 := a[t1]
    
    t3 live
    	t3 := 4 * i
    	t4 := b[t3]
    
    t5 live
    	t5 := t2 + t4
    	t6 := prod + t5
    
    t7 live
    	prod := t6
    	t7 := i + 1
    	i := t7
    
    	if i <= 20 goto L3
    

    t1, t3, t5, t7 can share the same location. What about t2, t4, and t6?

    Notes:

    DAG representation of basic blocks

    This concept is useful in code optimization. Although we are not doing a homework on optimization, you should understand it to be essential in real life and have heard and seen a bit of the terminology.

    A DAG for a basic block is one with the following labels on nodes:

    1. leaves are labelled by unique identifiers, either variable names or constants.
    2. interior nodes are labelled by operator symbols
    3. nodes are optionally given a sequence of identifiers as labels (these identifiers are deemed to have the value computed at that node).

    Example

    For the three-address code
    L:	t1 := 4 * i
    	t2 := a[t1]
    	t3 := 4 * i
    	t4 := b[t3]
    	t5 := t2 * t4
    	t6 := prod + t5
    	t7 := i + 1
    	i := t7
    	if i <= 20 goto L
    
    What should the corresponding DAG look like?







    Constructing a DAG

    Input: A basic block.

    Output: A DAG for the block, containing:

    Method: Consider an instruction x := y op z.

    1. If node(y), the node in the DAG that represents the value of y at that point, is undefined, then create a leaf labelled y. Let node(y) be this node. Similar for z.
    2. determine if there is a node labelled op with left child node(y) and right child node(z). if not, create such a node. let this node be n

    Applications of DAGs

    1. automatically detects common subexpressions
    2. can determine which identifiers have their value used in the block -- these are identifiers for which a leaf is created in step (1) at some point.
    3. Can determine which statements compute values that could be used outside the block -- these are statements s whose node n constructed in step (2) still has node(x)=n at the end of the DAG construction, where x is the identifier defined by S.
    4. Can reconstruct a simplified list of 3-addr instructions, taking advantage of common subexpressions, and not performing copyin assignments of the form x := y unless really necessary.

    Evaluating the nodes of a DAG

    The "optimized" basic block after DAG construction and common subexpression elimination equates x and z, but this behaves incorrectly when i = j.

    Code Optimization

    There are major classes of optimization that can significantly speedup a compiler's generated code. Usually you speed up code by doing the work with fewer instructions and by avoiding unnecessary memory reads and writes. You can also speed up code by rewriting it with fewer gotos.

    Constant Folding

    Constant folding is performing arithmetic at compile-time when the values are known. This includes simple expressions such as 2+3, but with more analysis some variables' values may be known constants for some of their uses.
         x = 7;
         ...
         y = x+5;
    

    Common Subexpression Elimination

    Code that redundantly computes the same value occurs fairly frequently, both explicitly because programmers wrote the code that way, and implicitly in the implementation of certain language features.

    Explicit:

        (a+b)*i + (a+b)/j;
    
    The (a+b) is a common subexpression that you should not have to compute twice.

    Implicit:

        x = a[i]; a[i] = a[j]; a[j] = x;
    
    Every array subscript requires an addition operation to compute the memory address; but do we have to compute the location for a[i] and a[j] twice in this code?

    Loop Unrolling

    Gotos are expensive (do you know why?). If you know a loop will execute at least (or exactly) 3 times, it may be faster to copy the loop body those three times than to do a goto. Removing gotos simplifies code, allowing other optimizations.

    originalunrolledafter subsequent constant folding
    for(i=0; i<3; i++) {
       x += i * i;
       y += x * x;
       }
    
       x += 0 * 0;
       y += x * x;
       x += 1 * 1;
       y += x * x;
       x += 2 * 2;
       y += x * x;
    
       y += x * x;
       x += 1;
       y += x * x;
       x += 4;
       y += x * x;
    

    Optimization Techniques, cont'd

    Algebraic Properties

    Implicit in the previous example of loop unrolling was the notion that certain computations can be simplified by basic math properties.

    name sample optimized as
    identities
    x = x * 1;
    x = x + 0;
    
    simplification
    y = (5 * x) + (7 * x);
    
    y = 12 * x;
    
    commutativity
    y = (5 * x) + (x * 7);
    
    y = (5 * x) + (7 * x);
    
    strength reduction
    x = y * 16;
    
    x = y << 4;
    
    This open-ended category might also include exploits of associativity, distributive properties, etc.

    Hoisting Loop Invariants

    This one requires knowledge, perhaps too much knowledge. I know the following optimization is safe, but does the compiler know? What would you have to know/prove in order for this example to be "safe" for a compiler to do?

    for (i=0; i<strlen(s); i++)
       s[i] = tolower(s[i]);
    
    t_0 = strlen(s);
    for (i=0; i<t_0; i++)
       s[i] = tolower(s[i]);
    

    Peephole Optimization

    Peephole optimizations look at the native code through a small, moving window for specific patterns that can be simplified. These are some of the easiest optimizations because they potentially don't require any analysis of other parts of the program in order to tell when they may be applied. Although some of these are stupid and you wouldn't think they'd come up, the simple code generation algorithm we presented earlier is quite stupid and does all sorts of obvious bad things that we can avoid.

    name sample optimized as
    redundant load or store
    MOVE R0,a
    MOVE a,R0
    
    MOVE R0,a
    
    dead code
    #define debug 0
    ...
    if (debug) printf("ugh");
    
    control flow simplification
    if a < b goto L1
    ...
    L1: goto L2
    
    if a < b goto L2
    ...
    L1: goto L2
    

    Peephole Optimization Examples

    It would be nice if we had time to develop a working demo program for peephole optimization, but let's start with the obvious.

    as generatedreplace withcomment
    	movq	%rdi, -56(%rbp)
    	cmpq	$1, -56(%rbp)
    
    	movq	%rdi, -56(%rbp)
    	cmpq	$1, %rdi
    
    reuse n that's already in a register
    	cmpq	$1, %rdi
    	setle	%al
    	movzbl	%al,%eax
    	movq	%rax, -8(%rbp)
    	cmpq	$0, -8(%rbp)
    	jne	.L0
    
    	cmpq	$1, %rdi
    	jle	.L0
    
    boolean variables are for wimps.
    setle sets a byte register (%al) to contain a boolean
    movzbl zero-extends a byte to a long (movsbl sign-extends)
    	cmpq	$1, %rdi
    	jle	.L0
    	jmp	.L1
    .L0:
    
    	cmpq	$1, %rdi
    	jg	.L1
    .L0:
    
    Use fall throughs when possible; avoid jumps.
    	movq	%rax, -16(%rbp)
    	movq	-16(%rbp), %rdi
    
    	movq	%rax, %rdi
    
    TAC code optimization might catch this sooner
    	movq	-56(%rbp), %rax
    	subq	$1, %rax
    	movq	%rax, %rdi
    
    	movq	-56(%rbp), %rdi
    	subq	$1, %rdi
    
    What was so special about %rax again?
    	movq	%rax, -40(%rbp)
    	movq	-24(%rbp), %rax
    	addq	-40(%rbp), %rax
    
    	addq	-24(%rbp), %rax
    
    Addition is commutative.

    Interprocedural Optimization

    Considering memory references across procedure call boundaries; for example, one might pass a parameter in a register if both the caller and callee generated code knows about it.

    argument culling

    when the value of a specific parameter is a constant, a custom version of a called procedure can be generated, in which the parameter is eliminated, and the constant is used directly (may allow additional constant folding).
    f(x,r,s,1);
    
    int f(int x, float y, char *z, int n)
    {
      switch (n) {
      case 1:
         do_A; break;
      case 2:
         do_B; break;
         ...
         }
    }
    
    f_1(x,r,s);
    
    int f_1(int x, float y, char *z)
    {
       do_A;
    }
    int f_2(int x, float y, char *z)
    {
       do_B;
    }
    ...
    

    Another Word on Interprocedural Optimization

    The optimization unit of this course mentions only the biggest categories of compiler optimization and gives very brief examples. That "argument culling" example of interprocedural optimization deserves at least a little more context: Warp to Final Exam Review

    Code Generation for Input/Output

    This section is on how to generate code for basic C input/output constructs.
    getchar()
    Basic appearance of a call to getchar() in final code:
    	call	getchar
    	movl	%eax, destination
    
    Of course, VGo does not have a getchar() function, it reads a line at a time. A built-in function for reading a line at a time might be built on top of this in vgo or in C, but it might be better to call a different input function.
    gets() is part of the C standard that permanently encodes a buffer overrun attack in the language for all time. However, we could use fgets(char*,int,FILE*) to implement VGo's input function.
    char *vgoread()
    {
       int i;
       char *buf = malloc(4096);
       if (buf == NULL) return NULL; // should do more
       i = fgets(buf, 4095, stdin);
       // should do more
       return buf;
    }
    
    What-all is wrong with this picture?
    printf(s...)
    First parameter is passed in %rdi. An "interesting" section in the AMD64 reference manuals explains that 32-bit operands are automatically sign-extended in 64-bit registers, but 8- and 16-bit operands are not automatically signed extended in 32-bit registers. If string s has label .LC0
    	movl	$.LC0, %eax	; load 32-bit addr
    				; magically sign-extended to 64-bits
    	movq	%rax, %rdi	; place 64-bit edition in param #1 reg.
    	call	printf		; call printf
    
    printf(s, i)
    Printf'ing an int ought to be the simplest printf. The second parameter is passed in %rsi. If you placed a 32-bit int in %esi you would still be OK.
    	movq	source, %rsi	; what we would do
    	movl	source, %esi	; "real" C int: 32, 64, same diff
    
    printf(s, c)
    Printf'ing a character involves passing that char as a parameter. Generally when passing a "char" parameter one would pass it in a (long word, aligned) slot, and it is prudent to (basically) promote it to "int" in this slot.
    	movsbl	source, %esi
    
    printf(s, s)
    Printf'ing a string involves passing that string as a parameter. For local variable string constant data, gcc does some pretty weird stuff. I'd kind of rather allocate the string constant out of the string constant region and then copy it into the local region, but perhaps calculating the contents of a string constant as a sequence of 32-bit long immediate values is an interesting exercise.

    Comments on Debugging Assembler

    The compiler writer that generates bad assembler code may need to debug the assembler code in order to understand why it is wrong.

    Dominators and Loops

    Raison d'etre: many/various Loop Optimizations require that loops be specially identified within a general flow graph context. If code is properly structured (e.g. no "goto" statements) these loop optimizations are safe to do, but in the general case for C you would have to check...
    dominator
    node d in a flow graph dominates node n (written as "d dom n") if every path from the initial node of the flow graph to n goes through d
    dominator tree
    tree formed from nodes in the flow graph whose root is the initial node, and node n is an ancestor of node m only if n dominates m. Each node in a flow graph has a unique "immediate dominator" (nearest dominator), hence a dominator tree can be formed.

    Loops in Flow Graphs

    Algorithm to construct the natural loop of a back edge

    Input: a flow graph G and a back edge n -> d
    Output: the set (named loop) consisting of all nodes in the natural loop of n -> d.

    Method: depth-first search on the reverse flow graph G'. Start with loop containing only node n and d. Consider each node m | m != d that is in loop, and insert m's predecessors in G into loop. Each node is placed once on a stack, so its predecessors will be examined. Since d is put in loop initially, its predecessors are not examined.

    procedure insert(m)
       if not member(loop, m) then {
          loop := loop ++ { m }
          push m onto stack
       }
    end
    
    main:
       stack := []
       loop := { d }
       insert(n)
       while stack not empty do {
          pop m off stack
          for each predecessor p of m do insert(p)
          }
    

    Inner Loops

    Code Generation for Virtual Machines

    A virtual machine architecture such as the JVM changes the "final" code generation somewhat. We have seen several changes, some of which simplify final code generation and some of which complicate things.
    no registers, simplified addressing
    a virtual machine may omit a register model and avoid complex addressing modes for different types of variables
    uni-size or descriptor-based values
    if all variables are "the same size", some of the details of memory management are simplified. In Java most values occupy a standard "slot" size, although some values occupy two slots. In Icon and Unicon, all values are stored using a same-size descriptor.
    runtime type system
    requiring type information at runtime may complicate the code generation task since type information must be present in generated code. For example in Java method invocation and field access instructions must encode class information.
    Just for fun, let's compare the generated code for java with that X86 native code we looked at earlier when we were talking about how to make variables spill out of registers:
    	iload_1
    	iload_2
    	iadd
    	iload_3
    	iload 4
    	iadd
    	imul
    	iload 5
    	iload 6
    	iadd
    	iload 7
    	iload_1
    	iadd
    	imul
    	iload_3
    	iload 5
    	imul
    	idiv
    	iadd
    	istore_1
    
    What do you see?





    thanks here to T. Mowry.

    Preheaders

    Loop optimizations often require code to be executed once before the loop. Example: loop hoisting. Solution: introduce a new (possibly empty) basic block for every loop. It had to have a header anyhow; give it a preheader.

    What was all that Loops/Dominators Stuff For?

    Given that you find such a natural loop, you can do:

    Loop Hoisting

    Hoisting Conditions thank you to Peter Lee Hoisting conditions For a loop-invariant definition d: t = x op y we can hoist instruction d into the loop’s pre-header if: 1. d’s block dominates all loop exits at which t is live-out, and 2. there is only one definition of t in the loop, and 3. t is not live-out of the pre-header

    Finding Loop Invariants

    OK, what can you do with this:

    Did you get:

    Exercise: run it a few billion times; see whether hoisting a couple operations out of the loop makes a measurable difference. It might not, after all... gotos are expensive. Exercise: what is wrong with this example? Another example (from Wikipedia loop-invariant code motion):

    for (int i = 0; i < n; i++) {
        x = y + z;
        a[i] = 6 * i + x * x;
    }
    
    One can hoise not just x = y + z; because it establishes x as invariant, subexpression x*x (into a temp variable) can also be hoisted.

    More on Runtime Systems

    Every compiler (including yours) needs a runtime system. A runtime system is the set of library functions and possibly global variables maintained by the language on behalf of a running program. You use one all the time; in C it functions like printf(), plus perhaps internal compiler-generated calls to do things the processor doesn't do in hardware.

    So you need a runtime system; potentially, this might be as big or bigger a job than writing the compiler. Languages vary from assembler (no runtime system) and C (small runtime system, mostly C with some assembler) on up to Java (large runtime system, mostly Java with some C) and in even higher level languages the compiler may evaporate and the runtime system become gigantic. The Unicon language has a relatively trivial compiler and gigantic virtual machine and runtime system. Other scripting languages might have no compiler at all, doing everything (even lexing and parsing) in the runtime system.

    Quick Look at the Implementation of Unicon

    The remainder of this quick look will focus on OO features, as implementing object-oriented language features is of broad interest.

    On Double Constants in Assembler

    X86 has immediate mode instructions for full 64-bit immediate constants...for at least one instruction/opcode (movabsq). Proof by contradiction. I note that the double 3.1415 was represented in output assembler by $4614256447914709615. Just for fun, I checked my earlier cast-to-long strategy:
    #include 
    using namespace std;
    int main()
    {
       double d;
       d = 3.1415;
       long l = *(long *)(&d);
       cout << "$" << l << endl;
    }
    
    outputs:
    $4614256447914709615
    
    We win.

    Tips on Invoking the Assembler and Linker from your Compiler

    The final HW calls for your compiler to produce an executable that I can run. But we have mainly discussed a compiler that writes out code suitable for input to an assembler. A student requested that I give you some tips on getting the rest of the way.

    Imports and Inheritance in Unicon

    Unicon is different from mainstream languages, and this section is not intended to tell you what you are supposed to do, it is intended to provide a basis for comparison.

    Syntax Tree Overview

    Unicon uses "iyacc", a variant of Berkeley yacc, which is a cousin of Bison. The unigram.y grammar has around 234 shift reduce conflicts. The semantic action at the import statement is illustrative of tree construction as well as what little semantic analysis Unicon does.
    import: IMPORT implist {
       $$ := node("import", $1,$2," ")
       import_class($2)
       } ;
    
    For what its worth, the tree type in Unicon is very challenging and sophisticated:
    record treenode(label, children)
    procedure node(label, kids[])
       return treenode(label, kids)
    end
    
    Actually, Unicon syntax trees are incredibly simple, except that they are actually heterogeneous trees with a mixture of treenode, string, token, and various class objects.

    Idol.icn

    Despite the generic tree, various class objects from idol.icn store all the interesting stuff in the syntax tree. It is almost really one class per non-terminal type, and those non-terminals that have symbol tables have a field in the class that contains the symbol (hash) table object.

    class Package (one of the only parts of idol.icn I didn't write) tells a real interesting story. There is both an in-memory representation of what we know about the world, and a persistent on-disk representation (in order to support separate compilation).

    #
    # a package is a virtual syntax construct; it does not appear in source
    # code, but is stored in the database.  The "fields" in a package are
    # the list of global symbols defined within that package.  The filelist
    # is the list of source files that are linked in when that package is
    # imported.
    #
    class Package : declaration(files, dir, classes)
       #
       # Add to the two global tables of imported symbols from this package's
       # set of symbols.  If sym is non-null, we are importing an individual
       # symbol (import "pack.symbol").
       #
       method add_imported(sym)
          local s, f
    
          if /dir then return
          
          f := open(dir || "/uniclass", "dr") |
    	 stop("Couldn't re-open uniclass db in " || dir)
          every s := (if \sym then sym else fields.foreach()) do {
             if member(imported, s) then
                 put(imported[s], self.name)
              else {
                 imported[s] := [self.name]
              }
    
             if fetch(f, self.name || "__" || s) then {
                if member(imported_classes, s) then
                   put(imported_classes[s], self.name)
                else {
                   imported_classes[s] := [self.name]
                }
             }
          }
          close(f)
       end
       method Read(line)
          self$declaration.Read(line)
          self.files := idTaque(":")
          self.files$parse(line[find(":",line)+1:find("(",line)] | "")
       end
       method size()
          return fields$size()
       end
       method insertfname(filename)
          /files := idTaque(":")
          if files.insert(filename) then {
             write(filename, " is added to package ", name)
             writespec()
             }
          else write(filename, " is already in Package ", name)
       end
       method insertsym(sym, filename)
          if fields.insert(sym) then {
             write(sym, " added to package ", name)
             writespec()
             }
          else write(sym, " is already in Package ", name)
       end
       method containssym(sym)
           return \fields.lookup(sym)
       end
       method String()
          s := self$declaration.String()
          fs := files.String()
          if *fs > 0 then fs := " : " || fs
          s := s[1: (*tag + *name + 2)] || fs || s[*tag+*name+2:0]
          return s
       end
       method writespec()
       if \name & (f := open(env,"d")) then {
          insert(f, name, String())
          close(f)
          return
          }
       stop("can't write package spec for ", image(name))
       end
    initially(name)
       if name[1] == name[-1] == "\"" then {
          name := name[2:-1]
          self.name := ""
          name ? {
    	 if upto('/\\') then {
    	    while self.name ||:= tab(upto('/\\')) do self.name ||:= move(1)
    	    }
    	 self.name ||:= tab(find(".")|0)
    	 }
          }
       else {
          self.name := name
          }
       if dbe := fetchspec(self.name) then {
          Read(dbe.entry)
          self.dir := dbe.dir
          }
       /tag := "package"
       /fields := classFields()
    end
    

    fetching a specification

    Given a class name, how do we find it? It must live in a GDBM database (uniclass) somewhere along the IPATH. A bunch of tedious string parsing concluding with a GDBM fetch.
    #
    # find a class specification, along the IPATH if necessary
    #
    procedure fetchspec(name)
       static white, nonwhite
       local basedir := "."
    $ifdef _MS_WINDOWS_NT
       white := ' \t;'
       nonwhite := &cset -- ' \t;'
    $else
       white := ' \t'
       nonwhite := &cset -- ' \t'
    $endif
       name ? {
          while basedir ||:= tab(upto('\\/')) do {
    	 basedir ||:= move(1)
    	 }
          name := tab(0)
          # throw away initial "." and trailing "/"
          if basedir[-1] == ("\\"|"/") then basedir := basedir[2:-1]
          }
       if f := open(basedir || "/" || env,"dr") then {
          if s := fetch(f, name) then {
    	 close(f)
    	 return db_entry(basedir, s)
    	 }
          close(f)
          }
    
       if basedir ~== "." then fail # if it gave a path, don't search IPATH
    
       ipath := ipaths()
    
       if \ipath then {
          ipath ? {
             dir := ""
    	 tab(many(white))
    	 while dir ||:= tab(many(nonwhite)) do {
    	    if *dir>0 & dir[1]=="\"" & dir[-1] ~== "\"" then {
    		dir ||:= tab(many(white)) | { fail }
    	       }
    	    else {
    		if dir[1]==dir[-1]=="\"" then dir := dir[2:-1]
    		if f := open(dir || "/" || env, "dr") then {
    		    if s := fetch(f, name) then {
    			close(f); return db_entry(dir, s) }
    		    close(f)
    		}
    		tab(many(white))
    		dir := ""
    	    }
    	}
         }
      }
    end
    

    Closure-Based Inheritance

    Unicon not only allows multiple inheritance, it is the only language that I know of that can handle cycles in the inheritance graph. It does this by having each child be completely self-centered. When they inherit, they rifle through their parents looking for spare change. This is a depth-first method that completely/perfectly inherits from the first superclass (including all its parents) and only then considers later step-parents.

    Inside class Class, supers is an object that maintains an ordered list of a class' superclasses (i.e. parents). Variable classes is effectively a global object that knows all the classes in the current package and let's you look them up by name. Variable added tracks classes already visited, and prevents repeating any classes already on the list.

      method transitive_closure()
        count := supers.size()
        while count > 0 do {
    	added := taque()
    	every sc := supers.foreach() do {
    	  if /(super := classes.lookup(sc)) then
    	    halt("class/transitive_closure: couldn't find superclass ",sc)
    	  every supersuper := super.foreachsuper() do {
    	    if / self.supers.lookup(supersuper) &
    		 /added.lookup(supersuper) then {
    	      added.insert(supersuper)
    	    }
    	  }
    	}
    	count := added.size()
    	every self.supers.insert(added$foreach())
        }
      end
    
    Given that all the superclasses have been ordered, the actual inheritance in class Class is done by a method resolve():
      method resolve()
        #
        # these are lists of [class , ident] records
        #
        self.imethods := []
        self.ifields := []
        ipublics := []
        addedfields := table()
        addedmethods := table()
        every sc := supers.foreach() do {
    	if /(superclass := classes.lookup(sc)) then
    	    halt("class/resolve: couldn't find superclass ",sc)
    	every superclassfield := superclass.foreachfield() do {
    	    if /self.fields.lookup(superclassfield) &
    	       /addedfields[superclassfield] then {
    		addedfields[superclassfield] := superclassfield
    		put ( self.ifields , classident(sc,superclassfield) )
    		if superclass.ispublic(superclassfield) then
    		    put( ipublics, classident(sc,superclassfield) )
    	    } else if \strict then {
    		warn("class/resolve: '",sc,"' field '",superclassfield,
    		     "' is redeclared in subclass ",self.name)
    	    }
    	}
    	every superclassmethod := (superclass.foreachmethod()).name() do {
    	    if /self.methods.lookup(superclassmethod) &
    	       /addedmethods[superclassmethod] then {
    		addedmethods[superclassmethod] := superclassmethod
    		put ( self.imethods, classident(sc,superclassmethod) )
    	    }
    	}
    	every public := (!ipublics) do {
    	    if public.Class == sc then
    		put (self.imethods, classident(sc,public.ident))
    	}
        }
      end
    

    Unicon Methods Vectors

    Unicon resolves each class' inheritance information at compile time, and generates a field table for runtime calculations that map field names to slot#/offsets. Methods vectors are just structs, shared by objects by means of a pointer (__m) added to class instances.

    final review began here

    Final Exam Review

    The final exam is comprehensive, but with a strong emphasis on middle and "back end" compiler issues: symbol tables, semantic analysis, and code generation. The exam is a 3 hour, open note exam. In Spring 2025 we wrote a compiler for a Kotlin dialect, so you can expect some questions about lexical, syntactic and semantic features of Kotlin that are so pervasive that we addressed them in class or in homework.

    Sample Final Exam

    This final is from a previous year and has questions specific to that year's project. But it gives you an idea of the kinds of questions that appear on the final. In Spring 2025, some of the questions may be in a different format (short answer, multiple choice, true or false).

    2023's Final Exam

    This final is from a previous year and has questions specific to that year's project. But it gives you an additional idea of the kinds of questions that appear on the final. How many of them are also on the previous example final? One thing I am very likely to do is ask you questions about semantic attributes manipulation, in a different format than shown in this example. Study available materials on semantic attributes and what you used them for in your homeworks.