User:Benmachine/Non-strict semantics

From HaskellWiki
< User:Benmachine
Revision as of 17:52, 30 April 2012 by Benmachine (talk | contribs) (New page: An expression language is said to have '''non-strict semantics''' if expressions can have a value even if some of their subexpressions do not. Haskell is one of the few modern languages to...)
(diff) ← Older revision | Latest revision (diff) | Newer revision → (diff)
Jump to navigation Jump to search

An expression language is said to have non-strict semantics if expressions can have a value even if some of their subexpressions do not. Haskell is one of the few modern languages to have non-strict semantics by default: nearly every other language has strict semantics, in which if any subexpression fails to have a value, the whole expression fails with it.

What?

Any sufficiently capable programming language is non-total, which is to say you can write expressions that do not produce a value: common examples are an infinite loop or unproductive recursion, e.g. the following definition in Haskell:

noreturn :: Integer -> Integer
noreturn x = negate (noreturn x)

or the following Python function:

def noreturn(x):
    while True:
        x = -x

    return x # not reached

both fail to produce a value when executed. We say that noreturn x is undefined, and write noreturn x = .

In Python the following expression:

{7: noreturn(5), 2: 0}[2]

also fails to have a value, because in order to construct the dictionary, the interpreter tries to work out noreturn(5), which of course doesn't return a value. This is called innermost-first evaluation: in order to call a function with some arguments, you first have to calculate what all the arguments are, starting from the innermost function call and working outwards. The result is that Python is strict, in the sense that calling any function with an undefined argument produces an undefined value, i.e. f(⊥) = ⊥.

In Haskell, an analogous expression:

lookup 2 [(7, noreturn 5), (2, 0)]

in fact has the value Just 0. The program does not try to compute noreturn 5 because it is irrelevant to the overall value of the computation: only the values that are necessary to the result are computed. This is called outermost-first evaluation because the first thing you look at is the call to lookup to decide if it needs to evaluate its arguments, and then if it does, you look at the arguments. The upshot is that Haskell has some functions which are not strict, i.e. sometimes f . This means that Haskell functions need not completely compute their arguments, which is why e.g. take 3 [1..] can produce [1,2,3] even though it is given a conceptually infinite list.

Why?

The important thing to understand about non-strict semantics is that it is not a performance feature. Non-strict semantics means that only the things that are needed for the answer are evaluated, but if you write your programs carefully, you'll only compute what is absolutely necessary anyway, so the extra time spent working out what should and shouldn't be evaluated is time wasted. For this reason, non-strict programs tend to be a little slower than strict programs.