Lambda calculus

From Vero - Wikipedia
Jump to navigation Jump to search

Template:Short description

The lambda abstraction decomposed. The <math>\lambda</math> indicates the start of a function. <math>x</math> is the input parameter. <math>M</math> is the body, separated by a dot separator "<math>.</math>" from the input parameter.

In mathematical logic, the lambda calculus (also written as λ-calculus) is a formal system for expressing computation based on function abstraction and application using variable binding and substitution. Untyped lambda calculus, the topic of this article, is a universal machine, a model of computation that can be used to simulate any Turing machine (and vice versa). It was introduced by the mathematician Alonzo Church in the 1930s as part of his research into the foundations of mathematics. In 1936, Church found a formulation which was logically consistent, and documented it in 1940.

Definition

Template:Main The lambda calculus consists of a language of lambda terms, that are defined by a certain formal syntax, and a set of transformation rules for manipulating the lambda terms. In BNF, the syntax is <math>e ::= x \mid \lambda x.e \mid e \, e,</math> where variables Template:Math range over an infinite set of names. Terms Template:Math range over all lambda terms. This corresponds to the following inductive definition:

A lambda term is syntactically valid if and only if it can be obtained by repeated application of these three rules. For convenience, parentheses can often be omitted when writing a lambda term—see Template:Section link for details.

In the term <math>\lambda x.M</math>, occurrences of Template:Mvar within Template:Mvar that are under the scope of this λ are termed bound; any occurrence of a variable not bound by an enclosing λ is free. Template:Math is the set of free variables of Template:Mvar. The notation <math>M[x := N]</math> denotes capture-avoiding substitution: substituting Template:Mvar for every free occurrence of Template:Mvar in Template:Mvar, while avoiding variable capture.Template:Efn This operation is defined inductively as follows:

  • <math>x[x := N] = N</math>; <math>y[x := N] = y</math> if <math>y \ne x</math>.
  • <math>(M_1 M_2)[x := N] = (M_1[x := N]) (M_2[x := N])</math>.
  • <math>(\lambda y.M)[x := N]</math> has three cases:
    • If <math>y = x</math>, becomes <math>\lambda x.M</math> (<math>x</math> is bound; no change).
    • If <math>y \notin FV(N)</math>, becomes <math>\lambda y.(M[x := N])</math>.
    • If <math>y \in FV(N)</math>, first α-rename <math>\lambda y.M</math> to <math>\lambda y'.M[y := y']</math> with <math>y'</math> fresh to avoid name collisions, then continue as above.

There are several notions of "equivalence" and "reduction" that make it possible to reduce lambda terms to equivalent lambda terms.<ref>Template:Cite journal</ref>

  • α-conversion captures the intuition that the particular choice of a bound variable, in an abstraction, does not (usually) matter. If <math>y \notin FV(M)</math>, then the terms <math>\lambda x.M</math> and <math>\lambda y.M[x := y]</math> are considered alpha-equivalent, written <math>\lambda x.M \equiv_{\alpha} \lambda y.M[x := y]</math>. The equivalence relation is the smallest congruence relation on lambda terms generated by this rule. For instance, <math>\lambda x.x</math> and <math>\lambda y.y</math> are alpha-equivalent lambda terms.
  • The β-reduction rule states that a β-redex, an application of the form <math>( \lambda x . t) s</math>, reduces to the term <math> t [ x := s]</math>.Template:Efn For example, for every <math>s</math>, <math>( \lambda x . x ) s \to x[ x := s ] = s </math>. This demonstrates that <math> \lambda x . x </math> really is the identity. Similarly, <math>( \lambda x . y ) s \to y [ x := s ] = y </math>, which demonstrates that <math> \lambda x . y </math> is a constant function.
  • η-conversion expresses extensionality and converts between <math>\lambda x . f x</math> and <math>f</math> whenever <math>x</math> does not appear free in <math>f</math>. It is often omitted in many treatments of lambda calculus.

Template:AnchorThe term redex, short for reducible expression, refers to subterms that can be reduced by one of the reduction rules. For example, (λx.M) N is a β-redex in expressing the substitution of N for x in M. The expression to which a redex reduces is called its reduct; the reduct of (λx.M) N is M[x := N].

Explanation and applications

Lambda calculus is Turing complete, that is, it is a universal model of computation that can be used to simulate any Turing machine.<ref>Template:Cite journal</ref> Its namesake, the Greek letter lambda (λ), is used in lambda expressions and lambda terms to denote binding a variable in a function.

Template:AnchorTemplate:AnchorLambda calculus may be untyped or typed. In typed lambda calculus, functions can be applied only if they are capable of accepting the given input's "type" of data. Typed lambda calculi are strictly weaker than the untyped lambda calculus, which is the primary subject of this article, in the sense that typed lambda calculi can express less than the untyped calculus can. On the other hand, more things can be proven with typed lambda calculi. For example, in simply typed lambda calculus, it is a theorem that every evaluation strategy terminates for every simply typed lambda-term, whereas evaluation of untyped lambda-terms need not terminate (see below). One reason there are many different typed lambda calculi has been the desire to do more (of what the untyped calculus can do) without giving up on being able to prove strong theorems about the calculus.

Lambda calculus has applications in many different areas in mathematics, philosophy,<ref>Template:Cite web</ref> linguistics,<ref>Template:Cite book</ref><ref>Template:Cite book</ref> and computer science.<ref>Template:Cite book.</ref><ref>Template:Cite tech report</ref> Lambda calculus has played an important role in the development of the theory of programming languages. Functional programming languages implement lambda calculus. Lambda calculus is also a current research topic in category theory.<ref>Template:Cite book</ref>

History

Lambda calculus was introduced by mathematician Alonzo Church in the 1930s as part of an investigation into the foundations of mathematics.<ref>Template:Cite journal</ref>Template:Efn The original system was shown to be logically inconsistent in 1935 when Stephen Kleene and J. B. Rosser developed the Kleene–Rosser paradox.<ref>Template:Cite journal</ref><ref>Template:Cite journal</ref>

Subsequently, in 1936 Church isolated and published just the portion relevant to computation, what is now called the untyped lambda calculus.<ref name="Church1936">Template:Cite journal</ref> In 1940, he also introduced a computationally weaker, but logically consistent system, known as the simply typed lambda calculus.<ref>Template:Cite journal</ref>

Until the 1960s when its relation to programming languages was clarified, the lambda calculus was only a formalism. Thanks to Richard Montague and other linguists' applications in the semantics of natural language, the lambda calculus has begun to enjoy a respectable place in both linguistics<ref name='mm-linguistics'>Template:Cite book</ref> and computer science.<ref>Template:Cite web</ref>

Origin of the λ symbol

Template:Anchor There is some uncertainty over the reason for Church's use of the Greek letter lambda (λ) as the notation for function-abstraction in the lambda calculus, perhaps in part due to conflicting explanations by Church himself. According to Cardone and Hindley (2006):

By the way, why did Church choose the notation "λ"? In [an unpublished 1964 letter to Harald Dickson] he stated clearly that it came from the notation "<math>\hat{x}</math>" used for class-abstraction by Whitehead and Russell, by first modifying "<math>\hat{x}</math>" to "<math>\land x</math>" to distinguish function-abstraction from class-abstraction, and then changing "<math>\land</math>" to "λ" for ease of printing.

This origin was also reported in [Rosser, 1984, p.338]. On the other hand, in his later years Church told two enquirers that the choice was more accidental: a symbol was needed and λ just happened to be chosen.

Dana Scott has also addressed this question in various public lectures.<ref>Dana Scott, "Looking Backward; Looking Forward", Invited Talk at the Workshop in honour of Dana Scott's 85th birthday and 50 years of domain theory, 7–8 July, FLoC 2018 (talk 7 July 2018). The relevant passage begins at 32:50. (See also this extract of a May 2016 talk at the University of Birmingham, UK.)</ref> Scott recounts that he once posed a question about the origin of the lambda symbol to Church's former student and son-in-law John W. Addison Jr., who then wrote his father-in-law a postcard:

Dear Professor Church,

Russell had the iota operator, Hilbert had the epsilon operator. Why did you choose lambda for your operator?

According to Scott, Church's entire response consisted of returning the postcard with the following annotation: "eeny, meeny, miny, moe".

Motivation

Computable functions are a fundamental concept within computer science and mathematics. The lambda calculus provides simple semantics for computation which are useful for formally studying properties of computation. The lambda calculus incorporates two simplifications that make its semantics simple. Template:AnchorThe first simplification is that the lambda calculus treats functions "anonymously"; it does not give them explicit names. For example, the function

<math>\operatorname{square\_sum}(x, y) = x^2 + y^2</math>

can be rewritten in anonymous form as

<math>(x, y) \mapsto x^2 + y^2</math>

(which is read as "a tuple of Template:Mvar and Template:Mvar is mapped to <math display="inline">x^2 + y^2</math>").Template:Efn Similarly, the function

<math>\operatorname{id}(x) = x</math>

can be rewritten in anonymous form as

<math>x \mapsto x</math>

where the input is simply mapped to itself.Template:Efn

The second simplification is that the lambda calculus only uses functions of a single input. An ordinary function that requires two inputs, for instance the <math display="inline">\operatorname{square\_sum}</math> function, can be reworked into an equivalent function that accepts a single input, and as output returns another function, that in turn accepts a single input. For example,

<math>(x, y) \mapsto x^2 + y^2</math>

can be reworked into

<math>x \mapsto (y \mapsto x^2 + y^2)</math>

This method, known as currying, transforms a function that takes multiple arguments into a chain of functions each with a single argument.

Function application of the <math display="inline">\operatorname{square\_sum}</math> function to the arguments (5, 2), yields at once

<math display="inline">((x, y) \mapsto x^2 + y^2)(5, 2)</math>
<math display="inline"> = 5^2 + 2^2 </math>
<math display="inline"> = 29</math>,

whereas evaluation of the curried version requires one more step

<math display="inline">\Bigl(\bigl(x \mapsto (y \mapsto x^2 + y^2)\bigr)(5)\Bigr)(2)</math>
<math display="inline"> = (y \mapsto 5^2 + y^2)(2)</math> // the definition of <math>x</math> has been used with <math>5</math> in the inner expression. This is like β-reduction.
<math display="inline"> = 5^2 + 2^2</math> // the definition of <math>y</math> has been used with <math>2</math>. Again, similar to β-reduction.
<math display="inline"> = 29 </math>

to arrive at the same result.

In lambda calculus, functions are taken to be 'first class values', so functions may be used as the inputs, or be returned as outputs from other functions. For example, the lambda term <math>\lambda x.x</math> represents the identity function, <math>x \mapsto x</math>. Further, <math>\lambda x.y</math> represents the constant function <math>x \mapsto y</math>, the function that always returns <math>y</math>, no matter the input. As an example of a function operating on functions, the function composition can be defined as <math>\lambda f. \lambda g. \lambda x. (f ( g x))</math>.

Normal forms and confluence

Template:Further It can be shown that β-reduction is confluent when working up to α-conversion (i.e. we consider two normal forms to be equal if it is possible to α-convert one into the other). If repeated application of the reduction steps eventually terminates, then by the Church–Rosser theorem it will produce a unique β-normal form. However, the untyped lambda calculus as a rewriting rule under β-reduction is neither strongly normalising nor weakly normalising; there are terms with no normal form such as Template:Mono.

Considering individual terms, both strongly normalising terms and weakly normalising terms have a unique normal form. For strongly normalising terms, any reduction strategy is guaranteed to yield the normal form, whereas for weakly normalising terms, some reduction strategies may fail to find it.

Encoding datatypes

Template:Further The basic lambda calculus may be used to model arithmetic, Booleans, data structures, and recursion, as illustrated in the following sub-sections i, ii, iii, and § iv.

Arithmetic in lambda calculus

There are several possible ways to define the natural numbers in lambda calculus, but by far the most common are the Church numerals, which can be defined as follows:

Template:Mono
Template:Mono
Template:Mono
Template:Mono

and so on. Or using an alternative syntax allowing multiple uncurried arguments to a function:

Template:Mono
Template:Mono
Template:Mono
Template:Mono

A Church numeral is a higher-order function—it takes a single-argument function Template:Mono, and returns another single-argument function. The Church numeral Template:Mono is a function that takes a function Template:Mono as argument and returns the Template:Mono-th composition of Template:Mono, i.e. the function Template:Mono composed with itself Template:Mono times. This is denoted Template:Mono and is in fact the Template:Mono-th power of Template:Mono (considered as an operator); Template:Mono is defined to be the identity function. Functional composition is associative, and so, such repeated compositions of a single function Template:Mono obey two laws of exponents, Template:Mono and Template:Mono, which is why these numerals can be used for arithmetic. (In Church's original lambda calculus, the formal parameter of a lambda expression was required to occur at least once in the function body, which made the above definition of Template:Mono impossible.)

One way of thinking about the Church numeral Template:Mono, which is often useful when analyzing programs, is as an instruction 'repeat n times'. For example, using the Template:Mono and Template:Mono functions defined below, one can define a function that constructs a (linked) list of n elements all equal to x by repeating 'prepend another x element' n times, starting from an empty list. The lambda term

Template:Mono

creates, given a Church numeral Template:Mono and some Template:Mono, a sequence of n applications

Template:Mono

By varying what is being repeated, and what argument(s) that function being repeated is applied to, a great many different effects can be achieved.

We can define a successor function, which takes a Church numeral Template:Mono and returns its successor Template:Mono by performing one additional application of the function Template:Mono it is supplied with, where Template:Mono means "n applications of f starting from x":

Template:Mono

Because the Template:Mono-th composition of Template:Mono composed with the Template:Mono-th composition of Template:Mono gives the Template:Mono-th composition of Template:Mono, Template:Mono, addition can be defined as

Template:Mono

Template:Mono can be thought of as a function taking two natural numbers as arguments and returning a natural number; it can be verified that

Template:Mono

and

Template:Mono

are beta-equivalent lambda expressions. Since adding Template:Mono to a number can be accomplished by repeating the successor operation Template:Mono times, an alternative definition is:

Template:Mono<ref>Template:Citation; A note (accessed 2017) at the original location suggests that the authors consider the work originally referenced to have been superseded by a book.</ref>

Similarly, following Template:Mono, multiplication can be defined as

Template:Mono<ref name="Selinger">

Template:Citation</ref> Thus multiplication of Church numerals is simply their composition as functions. Alternatively

Template:Mono

since multiplying Template:Mono and Template:Mono is the same as adding Template:Mono repeatedly, Template:Mono times, starting from zero.

Exponentiation, being the repeated multiplication of a number with itself, translates as a repeated composition of a Church numeral with itself, as a function. And repeated composition is what Church numerals are:

Template:Mono<ref name="BarendregtBarendsen" />

Alternatively here as well,

Template:Mono

Simplifying, it becomes

Template:Mono

but that is just an eta-expanded version of Template:Mono we already have, above.

The predecessor function, specified by two equations Template:Mono and Template:Mono, is considerably more involved. The formula

Template:Mono

can be validated by showing inductively that if T denotes Template:Mono, then Template:Mono for Template:Mono. Two other definitions of Template:Mono are given below, one using conditionals and the other using pairs. With the predecessor function, subtraction is straightforward. Defining

Template:Mono,

Template:Mono yields Template:Mono when Template:Mono and Template:Mono otherwise.

Logic and predicates

By convention, the following two definitions (known as Church Booleans) are used for the Boolean values Template:Mono and Template:Mono:

Template:Mono
Template:Mono

Then, with these two lambda terms, we can define some logic operators (these are just possible formulations; other expressions could be equally correct):

Template:Mono
Template:Mono
Template:Mono
Template:Mono

We are now able to compute some logic functions, for example:

Template:Mono
Template:Mono
Template:Mono

and we see that Template:Mono is equivalent to Template:Mono.

A predicate is a function that returns a Boolean value. The most fundamental predicate is Template:Mono, which returns Template:Mono if its argument is the Church numeral Template:Mono, but Template:Mono if its argument were any other Church numeral:

Template:Mono

The following predicate tests whether the first argument is less-than-or-equal-to the second:

Template:Mono,

and since Template:Mono, if Template:Mono and Template:Mono, it is straightforward to build a predicate for numerical equality.

The availability of predicates and the above definition of Template:Mono and Template:Mono make it convenient to write "if-then-else" expressions in lambda calculus. For example, the predecessor function can be defined as:

Template:Mono

which can be verified by showing inductively that Template:Mono is the add Template:Mono − 1 function for Template:Mono > 0.

Pairs

A pair (2-tuple) can be defined in terms of Template:Mono and Template:Mono, by using the Church encoding for pairs. For example, Template:Mono encapsulates the pair (Template:Mono,Template:Mono), Template:Mono returns the first element of the pair, and Template:Mono returns the second.

Template:Mono
Template:Mono
Template:Mono
Template:Mono
Template:Mono

A linked list can be defined as either NIL for the empty list, or the Template:Mono of an element and a smaller list. The predicate Template:Mono tests for the value Template:Mono. (Alternatively, with Template:Mono, the construct Template:Mono obviates the need for an explicit NULL test).

As an example of the use of pairs, the shift-and-increment function that maps Template:Mono to Template:Mono can be defined as

Template:Mono

which allows us to give perhaps the most transparent version of the predecessor function:

Template:Mono

Additional programming techniques

There is a considerable body of programming idioms for lambda calculus. Many of these were originally developed in the context of using lambda calculus as a foundation for programming language semantics, effectively using lambda calculus as a low-level programming language. Because several programming languages include the lambda calculus (or something very similar) as a fragment, these techniques also see use in practical programming, but may then be perceived as obscure or foreign.

Named constants

In lambda calculus, a library would take the form of a collection of previously defined functions, which as lambda-terms are merely particular constants. The pure lambda calculus does not have a concept of named constants since all atomic lambda-terms are variables, but one can emulate having named constants by setting aside a variable as the name of the constant, using abstraction to bind that variable in the main body, and apply that abstraction to the intended definition. Thus to use Template:Mono to mean N (some explicit lambda-term) in M (another lambda-term, the "main program"), one can say

Template:MonoMTemplate:Mono N

Authors often introduce syntactic sugar, such as Template:Mono,Template:Efn to permit writing the above in the more intuitive order

Template:Mono N Template:Mono M

By chaining such definitions, one can write a lambda calculus "program" as zero or more function definitions, followed by one lambda-term using those functions that constitutes the main body of the program.

A notable restriction of this Template:Mono is that the name Template:Mono may not be referenced in N, for N is outside the scope of the abstraction binding Template:Mono, which is M; this means a recursive function definition cannot be written with Template:Mono. The Template:MonoTemplate:Efn construction would allow writing recursive function definitions, where the scope of the abstraction binding Template:Mono includes N as well as M. Or self-application a-la that which leads to Template:Mono combinator could be used.

Recursion and fixed points

Template:Further Template:See also Recursion is when a function invokes itself. What would a value be which were to represent such a function? It has to refer to itself somehow inside itself, just as the definition refers to itself inside itself. If this value were to contain itself by value, it would have to be of infinite size, which is impossible. Other notations, which support recursion natively, overcome this by referring to the function by name inside its definition. Lambda calculus cannot express this, since in it there simply are no names for terms to begin with, only arguments' names, i.e. parameters in abstractions. Thus, a lambda expression can receive itself as its argument and refer to (a copy of) itself via the corresponding parameter's name. This will work fine in case it was indeed called with itself as an argument. For example, Template:Mono will express recursion when E is an abstraction which is applying its parameter to itself inside its body to express a recursive call. Since this parameter receives E as its value, its self-application will be the same Template:Mono again.

As a concrete example, consider the factorial function Template:Mono, recursively defined by

Template:Mono.

In the lambda expression which is to represent this function, a parameter (typically the first one) will be assumed to receive the lambda expression itself as its value, so that calling it with itself as its first argument will amount to the recursive call. Thus to achieve recursion, the intended-as-self-referencing argument (called Template:Mono here, reminiscent of "self", or "self-applying") must always be passed to itself within the function body at a recursive call point:

Template:Mono
with Template:Mono to hold, so Template:Mono and
Template:Mono

and we have

Template:Mono

Here Template:Mono becomes the same Template:Mono inside the result of the application Template:Mono, and using the same function for a call is the definition of what recursion is. The self-application achieves replication here, passing the function's lambda expression on to the next invocation as an argument value, making it available to be referenced there by the parameter name Template:Mono to be called via the self-application Template:Mono, again and again as needed, each time re-creating the lambda-term Template:Mono.

The application is an additional step just as the name lookup would be. It has the same delaying effect. Instead of having Template:Mono inside itself as a whole up-front, delaying its re-creation until the next call makes its existence possible by having two finite lambda-terms Template:Mono inside it re-create it on the fly later as needed.

This self-applicational approach solves it, but requires re-writing each recursive call as a self-application. We would like to have a generic solution, without the need for any re-writes:

Template:Mono
with Template:Mono to hold, so Template:Mono and
Template:Mono where Template:Mono
so that Template:Mono

Given a lambda term with first argument representing recursive call (e.g. Template:Mono here), the fixed-point combinator Template:Mono will return a self-replicating lambda expression representing the recursive function (here, Template:Mono). The function does not need to be explicitly passed to itself at any point, for the self-replication is arranged in advance, when it is created, to be done each time it is called. Thus the original lambda expression Template:Mono is re-created inside itself, at call-point, achieving self-reference.

In fact, there are many possible definitions for this Template:Mono operator, the simplest of them being:

Template:Anchor Template:Mono

In the lambda calculus, Template:Mono is a fixed-point of Template:Mono, as it expands to:

Template:Mono
Template:Mono
Template:Mono
Template:Mono
Template:Mono

Now, to perform the recursive call to the factorial function for an argument n, we would simply call Template:Mono. Given n = 4, for example, this gives: Template:Smalldiv Every recursively defined function can be seen as a fixed point of some suitably defined higher order function (also known as functional) closing over the recursive call with an extra argument. Therefore, using Template:Mono, every recursive function can be expressed as a lambda expression. In particular, we can now cleanly define the subtraction, multiplication, and comparison predicates of natural numbers, using recursion.

When Y combinator is coded directly in a strict programming language, the applicative order of evaluation used in such languages will cause an attempt to fully expand the internal self-application <math>(x x)</math> prematurely, causing stack overflow or, in case of tail call optimization, indefinite looping.<ref>Template:Cite web</ref> A delayed variant of Y, the Z combinator, can be used in such languages. It has the internal self-application hidden behind an extra abstraction through eta-expansion, as <math>(\lambda v.x x v)</math>, thus preventing its premature expansion:<ref>Template:Cite web</ref>

<math>Z = \lambda f.(\lambda x.f (\lambda v.x x v)) \ (\lambda x.f (\lambda v.x x v))\ .</math>

Standard terms

Certain terms have commonly accepted names:<ref>Template:Cite web</ref><ref>Template:Cite book</ref><ref>Template:Cite journal</ref>

Template:Anchor Template:Mono
Template:Anchor Template:Mono
Template:Anchor Template:Mono
Template:Anchor Template:Mono
Template:Anchor Template:Mono
Template:Anchor Template:Mono
Template:Anchor Template:Mono
Template:Anchor Template:Mono

Template:Mono is the identity function. Template:Mono and Template:Mono form complete combinator calculus systems that can express any lambda term - see the next section. Template:Mono is Template:Mono, the smallest term that has no normal form. Template:Mono is another such term. Template:Mono is standard and defined above, and can also be defined as Template:Mono, so that Template:Mono. Template:Mono and Template:Mono defined above are commonly abbreviated as Template:Mono and Template:Mono.

Abstraction elimination

Template:Further If N is a lambda-term without abstraction, but possibly containing named constants (combinators), then there exists a lambda-term T(Template:Mono,N) which is equivalent to Template:MonoN but lacks abstraction (except as part of the named constants, if these are considered non-atomic). This can also be viewed as anonymising variables, as T(Template:Mono,N) removes all occurrences of Template:Mono from N, while still allowing argument values to be substituted into the positions where N contains an Template:Mono. The conversion function T can be defined by:

T(Template:Mono, Template:Mono) := I
T(Template:Mono, N) := K N if Template:Mono is not free in N.
T(Template:Mono, M N) := S T(Template:Mono, M) T(Template:Mono, N)

In either case, a term of the form T(Template:Mono,N) P can reduce by having the initial combinator I, K, or S grab the argument P, just like β-reduction of Template:MonoNTemplate:Mono P would do. I returns that argument. K throws the argument away, just like Template:MonoNTemplate:Mono would do if Template:Mono has no free occurrence in N. S passes the argument on to both subterms of the application, and then applies the result of the first to the result of the second.

The combinators B and C are similar to S, but pass the argument on to only one subterm of an application (B to the "argument" subterm and C to the "function" subterm), thus saving a subsequent K if there is no occurrence of Template:Mono in one subterm. In comparison to B and C, the S combinator actually conflates two functionalities: rearranging arguments, and duplicating an argument so that it may be used in two places. The W combinator does only the latter, yielding the B, C, K, W system as an alternative to SKI combinator calculus.

Typed lambda calculus

Template:Further A typed lambda calculus is a typed formalism that uses the lambda-symbol (<math>\lambda</math>) to denote anonymous function abstraction. In this context, types are usually objects of a syntactic nature that are assigned to lambda terms; the exact nature of a type depends on the calculus considered (see Kinds of typed lambda calculi). From a certain point of view, typed lambda calculi can be seen as refinements of the untyped lambda calculus but from another point of view, they can also be considered the more fundamental theory and untyped lambda calculus a special case with only one type.<ref>Types and Programming Languages, p. 273, Benjamin C. Pierce</ref>

Typed lambda calculi are foundational programming languages and are the base of typed functional programming languages such as ML and Haskell and, more indirectly, typed imperative programming languages. Typed lambda calculi play an important role in the design of type systems for programming languages; here typability usually captures desirable properties of the program, e.g., the program will not cause a memory access violation.

Typed lambda calculi are closely related to mathematical logic and proof theory via the Curry–Howard isomorphism and they can be considered as the internal language of classes of categories, e.g., the simply typed lambda calculus is the language of a Cartesian closed category (CCC).

Reduction strategies

Template:Further Whether a term is normalising or not, and how much work needs to be done in normalising it if it is, depends to a large extent on the reduction strategy used. Common lambda calculus reduction strategies include:<ref>Template:Cite book</ref><ref>Template:Cite book</ref><ref>Template:Cite book</ref>

Normal order
The leftmost outermost redex is reduced first. That is, whenever possible, arguments are substituted into the body of an abstraction before the arguments are reduced. If a term has a beta-normal form, normal order reduction will always reach that normal form.
Applicative order
The leftmost innermost redex is reduced first. As a consequence, a function's arguments are always reduced before they are substituted into the function. Unlike normal order reduction, applicative order reduction may fail to find the beta-normal form of an expression, even if such a normal form exists. For example, the term <math>( \; \lambda x.y \;\; (\lambda z. (z z) \; \lambda z. (z z)) \; )</math> is reduced to itself by applicative order, while normal order reduces it to its beta-normal form <math>y</math>.
Full β-reductions
Any redex can be reduced at any time. This means essentially the lack of any particular reduction strategy—with regard to reducibility, "all bets are off".

Weak reduction strategies do not reduce under lambda abstractions:

Call by valueTemplate:Anchor
Like applicative order, but no reductions are performed inside abstractions. This is similar to the evaluation order of strict languages like C: the arguments to a function are evaluated before calling the function, and function bodies are not even partially evaluated until the arguments are substituted in.
Call by name
Like normal order, but no reductions are performed inside abstractions. For example, Template:Mono is in normal form according to this strategy, although it contains the redex Template:Mono.

Strategies with sharing reduce computations that are "the same" in parallel:

Optimal reduction
As normal order, but computations that have the same label are reduced simultaneously.
Call by need
As call by name (hence weak), but function applications that would duplicate terms instead name the argument. The argument may be evaluated "when needed", at which point the name binding is updated with the reduced value. This can save time compared to normal order evaluation.

Computability

There is no algorithm that takes as input any two lambda expressions and outputs Template:Mono or Template:Mono depending on whether one expression reduces to the other.<ref name="Church1936" /> More precisely, no computable function can decide the question. This was historically the first problem for which undecidability could be proven. As usual for such a proof, computable means computable by any model of computation that is Turing complete. In fact computability can itself be defined via the lambda calculus: a function F: NN of natural numbers is a computable function if and only if there exists a lambda expression f such that for every pair of x, y in N, F(x)=y if and only if f Template:Mono =β Template:Mono, where Template:Mono and Template:Mono are the Church numerals corresponding to x and y, respectively and =β meaning equivalence with β-reduction. See the Church–Turing thesis for other approaches to defining computability and their equivalence.

Church's proof of uncomputability first reduces the problem to determining whether a given lambda expression has a normal form. Then he assumes that this predicate is computable, and can hence be expressed in lambda calculus. Building on earlier work by Kleene and constructing a Gödel numbering for lambda expressions, he constructs a lambda expression Template:Mono that closely follows the proof of Gödel's first incompleteness theorem. If Template:Mono is applied to its own Gödel number, a contradiction results.

Complexity

The notion of computational complexity for the lambda calculus is a bit tricky, because the cost of a β-reduction may vary depending on how it is implemented.<ref>Template:Cite book</ref> To be precise, one must somehow find the location of all of the occurrences of the bound variable Template:Mono in the expression Template:Mono, implying a time cost, or one must keep track of the locations of free variables in some way, implying a space cost. A naïve search for the locations of Template:Mono in Template:Mono is O(n) in the length n of Template:Mono. Director strings were an early approach that traded this time cost for a quadratic space usage.<ref>Template:Cite journal</ref> More generally this has led to the study of systems that use explicit substitution.

In 2014, it was shown that the number of β-reduction steps taken by normal order reduction to reduce a term is a reasonable time cost model, that is, the reduction can be simulated on a Turing machine in time polynomially proportional to the number of steps.<ref>Template:Cite book</ref> This was a long-standing open problem, due to size explosion, the existence of lambda terms which grow exponentially in size for each β-reduction. The result gets around this by working with a compact shared representation. The result makes clear that the amount of space needed to evaluate a lambda term is not proportional to the size of the term during reduction. It is not currently known what a good measure of space complexity would be.<ref name=Reasonable>Template:Cite journal</ref>

An unreasonable model does not necessarily mean inefficient. Optimal reduction reduces all computations with the same label in one step, avoiding duplicated work, but the number of parallel β-reduction steps to reduce a given term to normal form is approximately linear in the size of the term. This is far too small to be a reasonable cost measure, as any Turing machine may be encoded in the lambda calculus in size linearly proportional to the size of the Turing machine. The true cost of reducing lambda terms is not due to β-reduction per se but rather the handling of the duplication of redexes during β-reduction.<ref name=Asperti>Template:Cite arXiv</ref> It is not known if optimal reduction implementations are reasonable when measured with respect to a reasonable cost model such as the number of leftmost-outermost steps to normal form, but it has been shown for fragments of the lambda calculus that the optimal reduction algorithm is efficient and has at most a quadratic overhead compared to leftmost-outermost.<ref name=Reasonable/> In addition the BOHM prototype implementation of optimal reduction outperformed both Caml Light and Haskell on pure lambda terms.<ref name=Asperti/>

Lambda calculus and programming languages

As pointed out by Peter Landin's 1965 paper "A Correspondence between ALGOL 60 and Church's Lambda-notation",<ref>Template:Cite journal</ref> sequential procedural programming languages can be understood in terms of the lambda calculus, which provides the basic mechanisms for procedural abstraction and procedure (subprogram) application.

Anonymous functions

Template:Further For example, in Python the "square" function can be expressed as a lambda expression as follows: <syntaxhighlight lang="python"> (lambda x: x**2) </syntaxhighlight>

The above example is an expression that evaluates to a first-class function. The symbol lambda creates an anonymous function, given a list of parameter names—just the single argument x, in this case—and an expression that is evaluated as the body of the function, x**2. Anonymous functions are sometimes called lambda expressions.

Pascal and many other imperative languages have long supported passing subprograms as arguments to other subprograms through the mechanism of function pointers. However, function pointers are an insufficient condition for functions to be first class datatypes, because a function is a first class datatype if and only if new instances of the function can be created at runtime. Such runtime creation of functions is supported in Smalltalk, JavaScript, Wolfram Language, and more recently in Scala, Eiffel (as agents), C# (as delegates) and C++11, among others.

Parallelism and concurrency

The Church–Rosser property of the lambda calculus means that evaluation (β-reduction) can be carried out in any order, even in parallel. This means that various nondeterministic evaluation strategies are relevant. However, the lambda calculus does not offer any explicit constructs for parallelism. One can add constructs such as futures to the lambda calculus. Other process calculi have been developed for describing communication and concurrency.

Semantics

The fact that lambda calculus terms act as functions on other lambda calculus terms, and even on themselves, led to questions about the semantics of the lambda calculus. Could a sensible meaning be assigned to lambda calculus terms? The natural semantics was to find a set D isomorphic to the function space DD, of functions on itself. However, no nontrivial such D can exist, by cardinality constraints because the set of all functions from D to D has greater cardinality than D, unless D is a singleton set.

In the 1970s, Dana Scott showed that if only continuous functions were considered, a set or domain D with the required property could be found, thus providing a model for the lambda calculus.<ref>Template:Cite journal Written 1969, widely circulated as an unpublished manuscript.</ref>

This work also formed the basis for the denotational semantics of programming languages.

Variations and extensions

These extensions are in the lambda cube:

These formal systems are extensions of lambda calculus that are not in the lambda cube:

These formal systems are variations of lambda calculus:

These formal systems are related to lambda calculus:

  • Combinatory logic – A notation for mathematical logic without variables
  • SKI combinator calculus – A computational system based on the S, K and I combinators, equivalent to lambda calculus, but reducible without variable substitutions

See also

Template:Portal Template:Colbegin

Template:Colend

Further reading

Monographs/textbooks for graduate students
  • Sørensen, Morten Heine and Urzyczyn, Paweł (2006), Lectures on the Curry–Howard isomorphism, Elsevier, Template:Isbn is a recent monograph that covers the main topics of lambda calculus from the type-free variety, to most typed lambda calculi, including more recent developments like pure type systems and the lambda cube. It does not cover subtyping extensions.
  • Template:Citation covers lambda calculi from a practical type system perspective; some topics like dependent types are only mentioned, but subtyping is an important topic.
Documents

Notes

Template:Notelist

References

Some parts of this article are based on material from FOLDOC, used with permission. Template:Reflist

Template:Commons category

Template:Alonzo Church Template:Mathematical logic Template:Authority control Template:Formal semantics Template:Functions navbox