Memory in LLVM

The ‘trick’ here is that while LLVM does require all register values to be in SSA form, it does not require (or permit) memory objects to be in SSA form. In the example above, note that the loads from G and H are direct accesses to G and H: they are not renamed or versioned. This differs from some other compiler systems, which do try to version memory objects. In LLVM, instead of encoding dataflow analysis of memory into the LLVM IR, it is handled with Analysis Passes which are computed on demand.

With this in mind, the high-level idea is that we want to make a stack variable (which lives in memory, because it is on the stack) for each mutable object in a function. To take advantage of this trick, we need to talk about how LLVM represents stack variables.

In LLVM, all memory accesses are explicit with load/store instructions, and it is carefully designed not to have (or need) an “address-of” operator. Notice how the type of the @G/@H global variables is actually i32* even though the variable is defined as i32. What this means is that @G defines space for an i32 in the global data area, but its name actually refers to the address for that space. Stack variables work the same way, except that instead of being declared with global variable definitions, they are declared with the LLVM alloca instruction:

  1. define i32 @example() {
  2. entry:
  3. %X = alloca i32 ; type of %X is i32*.
  4. ...
  5. %tmp = load i32* %X ; load the stack value %X from the stack.
  6. %tmp2 = add i32 %tmp, 1 ; increment it
  7. store i32 %tmp2, i32* %X ; store it back
  8. ...

This code shows an example of how we can declare and manipulate a stack variable in the LLVM IR. Stack memory allocated with the alloca instruction is fully general: we can pass the address of the stack slot to functions, we can store it in other variables, etc. In our example above, we could rewrite the example to use the alloca technique to avoid using a Phi node:

  1. @G = weak global i32 0 ; type of @G is i32*
  2. @H = weak global i32 0 ; type of @H is i32*
  3. define i32 @test(i1 %Condition) {
  4. entry:
  5. %X = alloca i32
  6. br i1 %Condition, label %cond_true, label %cond_false
  7. cond_true:
  8. %X.0 = load i32* @G
  9. store i32 %X.0, i32* %X
  10. br label %cond_next
  11. cond_false:
  12. %X.1 = load i32* @H
  13. store i32 %X.1, i32* %X
  14. br label %cond_next
  15. cond_next:
  16. %X.2 = load i32* %X
  17. ret i32 %X.2
  18. }

With this, we have discovered a way to handle arbitrary mutable variables without the need to create Phi nodes at all:

  • Each mutable variable becomes a stack allocation.
  • Each read of the variable becomes a load from the stack.
  • Each update of the variable becomes a store to the stack.
  • Taking the address of a variable just uses the stack address directly.

While this solution has solved our immediate problem, it introduced another one: we have now apparently introduced a lot of stack traffic for very simple and common operations, a major performance problem. Fortunately for us, the LLVM optimizer has a highly-tuned optimization pass named “mem2reg” that handles this case, promoting allocas like this into SSA registers, inserting Phi nodes as appropriate. If we run this example through the pass, for example, we’ll get:

  1. $ llvm-as < example.ll | opt -mem2reg | llvm-dis
  2. @G = weak global i32 0
  3. @H = weak global i32 0
  4. define i32 @test(i1 %Condition) {
  5. entry:
  6. br i1 %Condition, label %cond_true, label %cond_false
  7. cond_true:
  8. %X.0 = load i32* @G
  9. br label %cond_next
  10. cond_false:
  11. %X.1 = load i32* @H
  12. br label %cond_next
  13. cond_next:
  14. %X.01 = phi i32 [ %X.1, %cond_false ], [ %X.0, %cond_true ]
  15. ret i32 %X.01
  16. }

We say a block “A” dominates a different block “B” in the control flow graph if it’s impossible to reach “B” without passing through “A”, equivalently “A” is the dominator of “B”. The mem2reg pass implements the standard “iterated dominance frontier” algorithm for constructing SSA form and has a number of optimizations that speed up (very common) degenerate cases.

The mem2reg optimization pass is the answer to dealing with mutable variables, and we highly recommend that you depend on it. Note that mem2reg only works on variables in certain circumstances:

  • mem2reg is alloca-driven: it looks for allocas and if it can handle them, it promotes them. It does not apply to global variables or heap allocations.
  • mem2reg only looks for alloca instructions in the entry block of the function. Being in the entry block guarantees that the alloca is only executed once, which makes analysis simpler.
  • mem2reg only promotes allocas whose uses are direct loads and stores. If the address of the stack object is passed to a function, or if any funny pointer arithmetic is involved, the alloca will not be promoted.
  • mem2reg only works on allocas of first class values (such as pointers, scalars and vectors), and only if the array size of the allocation is 1 (or missing in the .ll file).
  • mem2reg is not capable of promoting structs or arrays to registers. Note that the “scalarrepl” pass is more powerful and can promote structs, “unions”, and arrays in many cases.

All of these properties are easy to satisfy for most imperative languages, and we’ll illustrate it below with Kaleidoscope. The final question you may be asking is: should I bother with this nonsense for my front-end? Wouldn’t it be better if I just did SSA construction directly, avoiding use of the mem2reg optimization pass? In short, we strongly recommend that you use this technique for building SSA form, unless there is an extremely good reason not to. Using this technique is:

  • Proven and well tested: clang uses this technique for local mutable variables. As such, the most common clients of LLVM are using this to handle a bulk of their variables. You can be sure that bugs are found fast and fixed early.
  • Extremely Fast: mem2reg has a number of special cases that make it fast in common cases as well as fully general. For example, it has fast-paths for variables that are only used in a single block, variables that only have one assignment point, good heuristics to avoid insertion of unneeded phi nodes, etc.
  • Needed for debug info generation: Debug information in LLVM relies on having the address of the variable exposed so that debug info can be attached to it. This technique dovetails very naturally with this style of debug info.

If nothing else, this makes it much easier to get our front-end up and running, and is very simple to implement. Let’s extend Kaleidoscope with mutable variables now!