Recursion (computer science)

From Wikipedia, the free encyclopedia

Recursion in computer programming defines a function in terms of itself. One example application of recursion is in recursive descent parsers for programming languages. The great advantage of recursion is that an infinite set of possible sentences, designs, or other data can be defined, parsed, or produced by a finite computer program.

Contents

[edit] Recursive algorithms

A common method of simplification is to divide a problem into sub-problems of the same type. As a computer programming technique, this is called divide and conquer, and it is key to the design of many important algorithms, as well as being a fundamental part of dynamic programming.

Virtually all programming languages in use today allow the direct specification of recursive functions and procedures. When such a function is called, the computer (for most languages on most stack-based architectures) or the language implementation keeps track of the various instances of the function (on many architectures, by using a call stack, although other methods may be used). Conversely, every recursive function can be transformed into an iterative function by using a stack.

Any function that can be evaluated by a computer can be expressed in terms of recursive functions without the use of iteration, in continuation-passing style; and conversely any recursive function can be expressed in terms of iteration.

To make a very literal example out of this: If an unknown word is seen in a book, the reader can make a note of the current page number and put the note on a stack (which is empty so far). The reader can then look the new word up and, while reading on the subject, may find yet another unknown word. The page number of this word is also written down and put on top of the stack. At some point an article is read that does not require any explanation. The reader then returns to the previous page number and continues reading from there. This is repeated, sequentially removing the topmost note from the stack. Finally, the reader returns to the original book. This is a recursive approach.

Some languages designed for logic programming and functional programming provide recursion as the only means of repetition directly available to the programmer. Such languages generally make tail recursion as efficient as iteration, letting programmers express other repetition structures (such as Scheme's map and for) in terms of recursion.

Recursion is deeply embedded in the theory of computation, with the theoretical equivalence of mu-recursive functions and Turing machines at the foundation of ideas about the universality of the modern computer.

[edit] Recursive programming

One basic form of recursive computer program is to define one or a few base cases, and then define rules to break down other cases into the base case. This analytic method is a common design for parsers for computer languages; see Recursive descent parser.

Another, similar form is generative, or synthetic, recursion. In this scheme, the computer uses rules to assemble cases, and starts by selecting a base case. This scheme is often used when a computer must design something automatically, such as code, a machine part or some other data.

One common example (using the Pascal programming language, in this case) is the function used to calculate the factorial of an integer:

function Factorial(x: integer): integer;
begin
  if x <= 1 then
    Factorial := 1
  else
    Factorial := x * Factorial(x-1);
end

Here is the same function coded without recursion. Notice that this iterative solution requires two temporary variables; in general, recursive formulations of algorithms are often considered "cleaner" or "more elegant" than iterative formulations.

function Factorial(x: integer): integer;
var i, temp: integer;
begin
  temp := 1;
  for i := 1 to x do
    temp := temp * i
  Factorial := temp
end

Another comparison that even more clearly demonstrates the relative "elegance" of recursive functions is the Euclidean Algorithm, used to compute the greatest common factor of two integers. Below is the algorithm with recursion, coded in C:

int GCF(int x, int y)
{
  if (x % y == 0)
     return y;
  else
     return GCF(y, x % y);
}

The above can be coded to remove the extra modulo; however, it either requires a variable declaration for each level of recursion or an extra recursive call (the base case would be if y = 0, meaning if x mod y = 0 we would make one more call whereas here we simply return). All three implementations are valid, only depending on what is most efficient for the language and processor (in C++, modulo is an inline operation so the above is generally fastest, but other languages require a less-efficient Mod(x,y) function call).

Below is the same algorithm using an iterative approach:

int GCF(int x, int y)
{
  int r;
  do
  {
     r = x % y;
     x = y;
     y = r;
  } while (r != 0);
  return x;
}

The iterative algorithm requires a temporary variable, and even given knowledge of the Euclidean Algorithm it is more difficult to understand the process by simple inspection. They are though, in fact, very similar in their steps.

[edit] Recursion versus iteration

In the "factorial" example the iterative implementation is likely to be slightly faster in practice than the recursive one. This is almost definite for the Euclidean Algorithm implementation. This result is typical, because iterative functions do not pay the "function-call overhead" as many times as recursive functions, and that overhead is relatively high in many languages. (Note that an even faster implementation for the factorial function on small integers is to use a lookup table.)

There are other types of problems whose solutions are inherently recursive, because they need to keep track of prior state. One example is tree traversal; others include the Ackermann function and divide-and-conquer algorithms such as Quicksort. All of these algorithms can be implemented iteratively with the help of a stack, but the need for the stack arguably nullifies the advantages of the iterative solution.

Another possible reason for choosing an iterative rather than a recursive algorithm is that in today's programming languages, the stack space available to a thread is often much less than the space available in the heap, and recursive algorithms tend to require more stack space than iterative algorithms.

[edit] Recursive functions

Functions whose domains can be recursively defined can be given recursive definitions patterned after the recursive definition of their domain.

The canonical example of a recursively defined function is the following definition of the factorial function f(n):

f(n) =  \begin{cases}  1 & \mbox{if } n = 0 \\  n \times f(n-1) & \mbox{if } n > 0 \\  \end{cases}


Given this definition, also called a recurrence relation, we work out f(3) as follows:

f(3) = 3 * f(3 − 1)
 = 3 * f(2)
 = 3 * 2 * f(2 − 1)
 = 3 * 2 * f(1)
 = 3 * 2 * 1 * f(1 − 1)
 = 3 * 2 * 1 * f(0)
 = 3 * 2 * 1 * 1
 = 6

[edit] Tail-recursive functions

Main article: Tail recursion

Tail-recursive functions are functions ending in a recursive call. For example, the following C function to locate a value in a linked list is tail-recursive, because the last thing it does is call itself:

struct node {
    int data;
    struct node *next;
};

struct node *find_value(struct node *head, int value)
{
    if (head == NULL)
      return NULL;
    if (head->data == value)
      return head;
    return find_value(head->next, value);
}

The Euclidean Algorithm function, following a similar structure, is also tail-recursive. On the other hand, the Factorial function used as an example in the previous section is not tail-recursive, because after it receives the result of the recursive call, it must multiply that result by x before returning to its caller. That kind of function is sometimes called augmenting recursive.

The Factorial function can be turned into a tail-recursive function:

function Factorial(acc: integer, x: integer): integer;
begin
  if x <= 1 then
    Factorial := acc
  else
    Factorial := Factorial(x * acc, x - 1);
end

Function should then be called by Factorial(1, x).

Notice that a single function may be both tail-recursive and augmenting recursive, such as this function to count the odd integers in a linked list:

int count_odds(struct node *head)
{
    if (head == NULL)
      return 0;
    if (head->data % 2 == 1)
      return count_odds(head->next) + 1;  /* augmenting recursion */
    return count_odds(head->next);  /* tail recursion */
}

The significance of tail recursion is that when making a tail-recursive call, the caller's return position need not be saved on the call stack; when the recursive call returns, it will branch directly on the previously saved return position. Therefore, on compilers which support tail-recursion optimization, tail recursion saves both space and time.

[edit] See also

[edit] Reference


[edit] External Links

In other languages