Difference between revisions of "Monad laws"

From HaskellWiki
Jump to navigation Jump to search
m (nicer presentation of the do-notation laws)
Line 10: Line 10:
 
Let us re-write the laws in do-notation:
 
Let us re-write the laws in do-notation:
   
  +
{|
<haskell>
 
  +
|-
1. do { x' <- return x do { f x
 
  +
| '''Left identity:'''
; f x' ≡ }
 
  +
| <haskell>
}
 
 
do { x' <- return x
 
  +
; f x'
2. do { x <- m ≡ do { m
 
  +
}
; return x } }
 
 
3. do { y <- do { x <- m do { x <- m
 
; f x ; do { y <- f x
 
} ≡ ; g y
 
; g y }
 
} }
 
 
do { x <- m
 
; y <- f x
 
≡ ; g y
 
}
 
 
</haskell>
 
</haskell>
  +
| ≡
  +
| <haskell>
  +
do { f x
  +
}
 
</haskell>
  +
|-
  +
| '''Right identity:'''
  +
| <haskell>
  +
do { x <- m
  +
; return x
  +
}
  +
</haskell>
  +
| ≡
  +
| <haskell>
  +
do { m
  +
}
  +
</haskell>
  +
|-
  +
| '''Associativity:'''
  +
| <haskell>
 
do { y <- do { x <- m
  +
; f x
  +
}
  +
; g y
  +
}
  +
</haskell>
  +
| ≡
  +
| <haskell>
  +
do { x <- m
  +
; do { y <- f x
  +
; g y
 
}
  +
}
  +
</haskell>
  +
| ≡
  +
| <haskell>
  +
do { x <- m
  +
; y <- f x
  +
; g y
  +
}
  +
</haskell>
  +
|}
   
 
In this notation the laws appear as plain common sense.
 
In this notation the laws appear as plain common sense.

Revision as of 17:16, 5 April 2012

All instances of the Monad class should obey:

  1. "Left identity": return a >>= f f a
  2. "Right identity": m >>= return m
  3. "Associativity": (m >>= f) >>= g m >>= (\x -> f x >>= g)

Note: monad m as used here is a concrete type.

What is the practical meaning of the monad laws?

Let us re-write the laws in do-notation:

Left identity:
do { x' <- return x
   ; f x'
   }
do { f x
   }
Right identity:
do { x <- m
   ; return x
   }
do { m
   }
Associativity:
do { y <- do { x <- m
             ; f x
             }
   ; g y
   }
do { x <- m
   ; do { y <- f x
        ; g y
        }
   }
do { x <- m
   ; y <- f x
   ; g y
   }

In this notation the laws appear as plain common sense.

But why should monads obey these laws?

When we see a program written in a form on the LHS, we expect it to do the same thing as the corresponding RHS; and vice versa. And in practice, people do write like the lengthier LHS once in a while. First example: beginners tend to write

skip_and_get = do { unused <- getLine
                  ; line <- getLine
                  ; return line
                  }

and it would really throw off both beginners and veterans if that did not act like (by law #2)

skip_and_get = do { unused <- getLine
                  ; getLine
                  }

Second example: Next, you go ahead to use skip_and_get:

main = do { answer <- skip_and_get
          ; putStrLn answer
          }

The most popular way of comprehending this program is by inlining (whether the compiler does or not is an orthogonal issue):

main = do { answer <- do { unused <- getLine
                         ; getLine
                         }
          ; putStrLn answer
          }

and applying law #3 so you can pretend it is

main = do { unused <- getLine
          ; answer <- getLine
          ; putStrLn answer
          }

Law #3 is amazingly pervasive: you have always assumed it, and you have never noticed it.

Whether compilers exploit the laws or not, you still want the laws for your own sake, just so you can avoid pulling your hair for counter-intuitive program behaviour that brittlely depends on how many redundant "return"s you insert or how you nest your do-blocks.

But it doesn't look exactly like an "associative law"...

Not in this form, no. To see precisely why they're called "identity laws" and an "associative law", you have to change your notation slightly.

The monad composition operator (also known as the Kleisli composition operator) is defined in Control.Monad:

(>=>) :: Monad m => (a -> m b) -> (b -> m c) -> a -> m c
m >=> n = \x -> do { y <- m x; n y }

Using this operator, the three laws can be expressed like this:

  1. "Left identity": return >=> g g
  2. "Right identity": f >=> return f
  3. "Associativity": (f >=> g) >=> h f >=> (g >=> h)

It's now easy to see that monad composition is an associative operator with left and right identities.

This is a very important way to express the three monad laws, because they are precisely the laws that are required for monads to form a mathematical category. So the monad laws can be summarised in convenient Haiku form:

Monad axioms:
Kleisli composition forms
a category.