Tail recursion

From Wikipedia, the free encyclopedia

In computer science, tail recursion (or tail-end recursion) is a special case of recursion in which the last operation of the function is a recursive call. Such recursions can be easily transformed to iterations. Replacing recursion with iteration, manually or automatically, can drastically decrease the amount of stack space used and improve efficiency. This technique is commonly used with functional programming languages, where the declarative approach and explicit handling of state promote the use of recursive functions that would otherwise rapidly fill the call stack.

Contents

[edit] Description

When a function is called, the computer must "remember" the place it was called from, the return address, so that it can return to that location with the result once the call is complete. Typically, this information is saved on the stack, a simple list of return locations in order of the times that the call locations they describe were reached. Sometimes, the last thing that a function does after completing all other operations is to simply call a function, possibly itself, and return its result. But in this case, there is no need to remember the place we are calling from — instead, we can leave the stack alone, and the newly called function will return its result directly to the original caller. Converting a call to a branch or jump in such a case is called a tail call optimization. Note that the tail call doesn't have to appear lexically after all other statements in the source code; it is only important that its result be immediately returned, since the calling function will never get a chance to do anything after the call if the optimization is performed.

For normal, non-recursive function calls, this is usually a micro-optimization that saves little time and space, since there are not that many different functions available to call. When dealing with recursive or mutually recursive functions, however, the stack space and the number of returns saved can grow to huge numbers, since a function can call itself, directly or indirectly, a huge number of times. In fact, it often asymptotically reduces stack space requirements from linear, or O(n), to constant, or O(1).

If several functions are mutually recursive, meaning they each call one another, and each call they make to one another in an execution sequence uses a tail call, then tail call optimization will give a properly tail recursive implementation that does not consume stack space. Proper tail recursion optimization is required by the standard definitions of some programming languages, such as Scheme.

The notion of tail position in Scheme can be defined as follows:

  1. The body of a lambda expression is in tail position.
  2. If (if E0 E1 E2) is in tail position, then both E1 and E2 are in tail position.

[edit] Examples

Take this Scheme program as an example:

(define (factorial n)
  (define (fac-times n acc)
    (if (= n 0)
        acc
        (fac-times (- n 1) (* acc n))))
  (if (< n 0)
      (display "Wrong argument!")
      (fac-times n 1)))

As you can see, the inner procedure fac-times calls itself last in the control flow. This allows an interpreter or compiler to reorganize the execution which would ordinarily look like this:

  call factorial (3)
   call fac-times (3 1)
    call fac-times (2 3)
     call fac-times (1 6)
      call fac-times (0 6)
      return 6
     return 6
    return 6
   return 6
  return 6

into the more space- (and time-) efficient variant:

  call factorial (3)
  replace arguments with (3 1), jump to "fac-times"
  replace arguments with (2 3), jump to "fac-times"
  replace arguments with (1 6), jump to "fac-times"
  replace arguments with (0 6), jump to "fac-times"
  return 6

This reorganization saves space because no state except for the calling function's address needs to be saved, either on the stack or on the heap. This also means that the programmer need not worry about running out of stack or heap space for extremely deep recursions.

Some programmers working in functional languages will rewrite recursive code to be tail-recursive so they can take advantage of this feature. This often requires addition of an "accumulator" argument (acc in the above example) to the function. In some cases (such as filtering lists) and in some languages, full tail recursion may require a function that was previously purely functional to be written such that it mutates references stored in other variables.[citation needed]

Besides space and execution efficiency, tail recursion optimization is important in the functional programming idiom known as continuation passing style (CPS), which would otherwise quickly run out of stack space.

[edit] Tail recursion modulo cons

Tail recursion modulo cons is a generalization of tail recursion introduced by David H. D. Warren. As the name suggests, the only operation needed after the recursive call is a cons, which adds a new element to the front of the list that was returned. The optimization moves this operation inside the recursive call by creating a list node with the front element, and passing a reference to this node as an argument.

For example, consider a function that duplicates a linked list, described here in C:

list *duplicate(const list *input)
{
    if (input == NULL) {
        return NULL;
    } else {
        list *head  = malloc(sizeof *head);
        head->value = input->value;
        head->next  = duplicate(input->next);
        return head;
    }
}

In this form the function is not tail-recursive, because control returns to the caller after the recursive call to set the value of head->next. But on resumption, the caller merely prepends a value to the result from the callee. So the function is tail-recursive, save for a "cons" action, that is, tail recursive modulo cons. Warren's method gives the following purely tail-recursive implementation:

list *duplicate(const list *input)
{
    list *head;
    duplicate_prime(input, &head);
    return head;
}
 
void duplicate_prime(const list *input, list **p)
{
    if (input == NULL) {
        *p = NULL;
    } else {
        *p = malloc(sizeof **p);
        (*p)->value = input->value;
        duplicate_prime(input->next, &(*p)->next);
    }
}

Note how the callee now appends to the end of the list, rather than have the caller prepend to the beginning.

The properly tail-recursive implementation can be converted to iterative form:

list *duplicate(const list *input)
{
     list *head;
     list **p = &head;
 
     while (input != NULL) {
         *p = malloc(sizeof **p);
         (*p)->value = input->value;
         input = input->next;
         p = &(*p)->next;
     }
     *p = NULL;
     return head;
}

[edit] Implementation methods

Tail recursion is important to some high-level languages, especially functional languages and members of the Lisp family. In these languages, tail recursion is the most commonly used way (and sometimes the only way available) of implementing iteration. The language specification of Scheme requires that tail-recursive operations are to be optimized so as not to grow the stack. Tail calls can also be used in Perl, with a variant of the "goto" statement that takes a function name: goto &NAME;

Since many Scheme compilers use C as an intermediate target code, the problem comes down to coding tail recursion in C without growing the stack. Many implementations achieve this by using a device known as a trampoline, a piece of code that repeatedly calls functions. All functions are entered via the trampoline. When a function has to call another, instead of calling it directly it returns the address of the function to be called, the arguments to be used, and so on, to the trampoline. This ensures that the C stack does not grow and iteration can continue indefinitely.

As this article by Samuel Jack suggests, it is possible to implement trampolining using higher-order functions in languages that support them, such as C#.

Using a trampoline for all function calls is rather more expensive than the normal C function call, so at least one Scheme compiler, Chicken, uses a technique first described by Henry Baker from an unpublished suggestion by Andrew Appel,[1] in which normal C calls are used but the stack size is checked before every call. When the stack reaches its maximum permitted size, objects on the stack are garbage-collected using the Cheney algorithm by moving all live data into a separate heap. Following this, the stack is unwound ("popped") and the program resumes from the state saved just before the garbage collection. Baker says "Appel's method avoids making a large number of small trampoline bounces by occasionally jumping off the Empire State Building."[1] The garbage collection ensures that mutual tail recursion can continue indefinitely.

[edit] See also

[edit] References

Look up tail recursion in Wiktionary, the free dictionary.
  • D. H. D. Warren, DAI Research Report 141, University of Edinburgh, 1980.
  1. ^ a b Henry Baker, "CONS Should Not CONS Its Arguments, Part II: Cheney on the M.T.A."

This article was originally based on material from the Free On-line Dictionary of Computing, which is licensed under the GFDL.