Difference between revisions of "Applications and libraries/Concurrency and parallelism"

From HaskellWiki
Jump to navigation Jump to search
(18 intermediate revisions by 2 users not shown)
Line 1: Line 1:
 
'''Concurrent and Parallel Programming'''
 
'''Concurrent and Parallel Programming'''
   
  +
Haskell offers a broad spectrum of tools for developing parallel or concurrent programs. For parallelism, Haskell libraries enable concise high-level parallel programs with results that are guaranteed to be deterministic, i.e., independent of the number of cores and the scheduling being used. Concurrency is supported with lightweight threads and high level abstractions such as [[software transactional memory]] for managing information shared across threads. Distributed programming is still mainly a research area. Some low-level tools (MPI bindings) and research prototypes are available and new approaches being developed, such as Cloud Haskell (Erlang-style actors as a Haskell library).
Haskell has been designed for parallel and concurrent programming since
 
  +
its inception. In particular, Haskell's [http://haskell.org/haskellwiki/Why_Haskell_matters purity] greatly simplifies reasoning about
 
parallel programs. This page lists libraries and extensions for programming
+
This page lists libraries and extensions for programming
concurrent and parallel applications in Haskell. See also the
+
concurrent and parallel applications in Haskell.
  +
[[Parallel|parallel portal]] for research papers, tutorials and on
 
  +
See also the
 
[[Parallel|parallel Haskell portal]] for research papers, tutorials and on
 
parallel and concurrent Haskell.
 
parallel and concurrent Haskell.
   
 
== Parallelism ==
 
== Parallelism ==
   
=== Parallel Strategies ===
+
=== Strategies and Par ===
 
Strategies provide a high-level compositional API for parallel programming.
 
 
* [http://hackage.haskell.org/package/parallel The parallel package on Hackage]
 
* [http://hackage.haskell.org/packages/archive/parallel/3.1.0.1/doc/html/Control-Parallel-Strategies.html Control.Parallel.Strategies Documentation]
 
* Latest paper: [http://research.microsoft.com/apps/pubs/default.aspx?id=138042 Seq no more: Better Strategies for Parallel Haskell]
 
* Original paper: [http://www.macs.hw.ac.uk/~dsg/gph/papers/html/Strategies/strategies.html Algorithm + Strategy = Parallelism]
 
 
=== Monad-par ===
 
   
  +
;Strategies
 
:A high-level compositional API for parallel programming.
 
:* [http://hackage.haskell.org/package/parallel The parallel package on Hackage]
 
:* [http://hackage.haskell.org/packages/archive/parallel/3.1.0.1/doc/html/Control-Parallel-Strategies.html Control.Parallel.Strategies Documentation]
 
:* Latest paper: [http://research.microsoft.com/apps/pubs/default.aspx?id=138042 Seq no more: Better Strategies for Parallel Haskell]
 
:* Original paper: [http://www.macs.hw.ac.uk/~dsg/gph/papers/html/Strategies/strategies.html Algorithm + Strategy = Parallelism]
 
;[http://hackage.haskell.org/package/monad-par Monad-par]
 
;[http://hackage.haskell.org/package/monad-par Monad-par]
 
:An alternative parallel programming API to that provided by the parallel package. The Par monad allows the simple description of parallel computations, and can be used to add parallelism to pure Haskell code. The basic API is straightforward: the monad supports forking and simple communication in terms of IVars.
 
:An alternative parallel programming API to that provided by the parallel package. The Par monad allows the simple description of parallel computations, and can be used to add parallelism to pure Haskell code. The basic API is straightforward: the monad supports forking and simple communication in terms of IVars.
   
=== Data Parallel Haskell ===
+
=== Data parallelism ===
   
  +
;[http://repa.ouroborus.net/ Repa]
  +
:REgular PArallel arrays: high performance, regular, multi-dimensional, shape polymorphic parallel arrays.
  +
;[https://github.com/AccelerateHS/accelerate Accelerate]
  +
:An embedded language of array computations for high-performance computing in Haskell. Computations on multi-dimensional, regular arrays are expressed in the form of parameterised collective operations (such as maps, reductions, and permutations).
 
;[http://haskell.org/haskellwiki/GHC/Data_Parallel_Haskell Data Parallel Haskell]
 
;[http://haskell.org/haskellwiki/GHC/Data_Parallel_Haskell Data Parallel Haskell]
 
:Implicitly parallel, high performance (nested) arrays, supporting large multicore programming.
 
:Implicitly parallel, high performance (nested) arrays, supporting large multicore programming.
   
=== Research tools ===
+
=== Research efforts ===
   
 
; Feedback-directed implicit parallelism
 
; Feedback-directed implicit parallelism
Line 39: Line 42:
   
 
== Concurrency ==
 
== Concurrency ==
  +
  +
=== Concurrent Haskell ===
   
 
;[[Concurrency|Concurrent Haskell]]
 
;[[Concurrency|Concurrent Haskell]]
 
:GHC has supported concurrency with lightweight threads for more than a decade, and it [http://shootout.alioth.debian.org/gp4/benchmark.php?test=threadring&lang=all is very fast]. Threads in Haskell are preemptively scheduled and support everything you would normally expect from threads, including blocking I/O and foreign calls.
 
:GHC has supported concurrency with lightweight threads for more than a decade, and it [http://shootout.alioth.debian.org/gp4/benchmark.php?test=threadring&lang=all is very fast]. Threads in Haskell are preemptively scheduled and support everything you would normally expect from threads, including blocking I/O and foreign calls.
* [http://www.haskell.org/ghc/docs/latest/html/libraries/base/Control-Concurrent.html Documentation]
+
:* [http://www.haskell.org/ghc/docs/latest/html/libraries/base/Control-Concurrent.html Documentation]
* [http://haskell.org/haskellwiki/Concurrency_demos Examples]
+
:* [http://haskell.org/haskellwiki/Concurrency_demos Examples]
   
 
=== Software transactional memory ===
 
=== Software transactional memory ===
Line 49: Line 54:
 
;[[Software transactional memory|Software Transactional Memory]]
 
;[[Software transactional memory|Software Transactional Memory]]
 
:GHC supports a sophisticated version of software transactional memory. Software Transactional Memory (STM) is a new way to coordinate concurrent threads.
 
:GHC supports a sophisticated version of software transactional memory. Software Transactional Memory (STM) is a new way to coordinate concurrent threads.
* [http://hackage.haskell.org/package/stm Documentation]
+
:* [http://hackage.haskell.org/package/stm Documentation]
* The paper [http://research.microsoft.com/en-us/um/people/simonpj/papers/stm/stm.pdf Composable memory transactions].
+
:* The paper [http://research.microsoft.com/en-us/um/people/simonpj/papers/stm/stm.pdf Composable memory transactions].
* The paper [http://research.microsoft.com/en-us/um/people/simonpj/papers/stm/lock-free-flops06.pdf Lock-free data structures using Software Transactional Memory in Haskell] gives further examples of concurrent programming using STM.
+
:* The paper [http://research.microsoft.com/en-us/um/people/simonpj/papers/stm/lock-free-flops06.pdf Lock-free data structures using Software Transactional Memory in Haskell] gives further examples of concurrent programming using STM.
   
=== Actors ===
+
=== Research efforts ===
   
 
;[http://hackage.haskell.org/cgi-bin/hackage-scripts/package/actor Actors with multi-headed receive clauses]
 
;[http://hackage.haskell.org/cgi-bin/hackage-scripts/package/actor Actors with multi-headed receive clauses]
Line 59: Line 64:
 
;[http://www.cs.kent.ac.uk/projects/ofa/chp/ CHP: Communicating Haskell Processes]
 
;[http://www.cs.kent.ac.uk/projects/ofa/chp/ CHP: Communicating Haskell Processes]
 
:CHP is built on the ideas of CSP (Communicating Sequential Processes), featuring encapsulated parallel processes (no shared data!) communicating over synchronous channels. This is a very composable mode that also allows choice on communications, so that a process may offer to either read on one channel or write on another, but will only take the first that is available.
 
:CHP is built on the ideas of CSP (Communicating Sequential Processes), featuring encapsulated parallel processes (no shared data!) communicating over synchronous channels. This is a very composable mode that also allows choice on communications, so that a process may offer to either read on one channel or write on another, but will only take the first that is available.
 
=== Helper tools ===
 
 
;[[WrapConc | Wrapped Concurrency]]
 
:A wrapper around Control.Concurrency and Control.Exception that provides versions of forkIO that have more guarantees.
 
   
 
----
 
----
Line 85: Line 85:
 
:Eden extends Haskell with a small set of syntactic constructs for explicit process specification and creation. While providing enough control to implement parallel algorithms efficiently, it frees the programmer from the tedious task of managing low-level details by introducing automatic communication (via head-strict lazy lists), synchronisation, and process handling.
 
:Eden extends Haskell with a small set of syntactic constructs for explicit process specification and creation. While providing enough control to implement parallel algorithms efficiently, it frees the programmer from the tedious task of managing low-level details by introducing automatic communication (via head-strict lazy lists), synchronisation, and process handling.
   
  +
;[[Cloud Haskell]]
=== Research tools ===
 
  +
:Cloud Haskell is a domain-specific language for developing programs for a distributed computing environment. Implemented as a shallow embedding in Haskell, it provides a message passing communication model, inspired by Erlang, without introducing incompatibility with Haskell’s established shared-memory concurrency.
  +
 
=== Research efforts ===
   
 
;[http://www.cs.kent.ac.uk/~cr3/HCPN/ HCPN: Haskell-Coloured Petri Nets]
 
;[http://www.cs.kent.ac.uk/~cr3/HCPN/ HCPN: Haskell-Coloured Petri Nets]
 
:Haskell-Coloured Petri Nets (HCPN) are an instance of high-level Petri Nets, in which anonymous tokens are replaced by Haskell data objects (and transitions can operate on that data, in addition to moving it around). This gives us a hybrid graphical/textual modelling formalism for Haskell, especially suited for modelling concurrent and distributed systems.
 
:Haskell-Coloured Petri Nets (HCPN) are an instance of high-level Petri Nets, in which anonymous tokens are replaced by Haskell data objects (and transitions can operate on that data, in addition to moving it around). This gives us a hybrid graphical/textual modelling formalism for Haskell, especially suited for modelling concurrent and distributed systems.
  +
  +
  +
[[Category:Parallel]]

Revision as of 09:57, 17 August 2012

Concurrent and Parallel Programming

Haskell offers a broad spectrum of tools for developing parallel or concurrent programs. For parallelism, Haskell libraries enable concise high-level parallel programs with results that are guaranteed to be deterministic, i.e., independent of the number of cores and the scheduling being used. Concurrency is supported with lightweight threads and high level abstractions such as software transactional memory for managing information shared across threads. Distributed programming is still mainly a research area. Some low-level tools (MPI bindings) and research prototypes are available and new approaches being developed, such as Cloud Haskell (Erlang-style actors as a Haskell library).

This page lists libraries and extensions for programming concurrent and parallel applications in Haskell.

See also the parallel Haskell portal for research papers, tutorials and on parallel and concurrent Haskell.

Parallelism

Strategies and Par

Strategies
A high-level compositional API for parallel programming.
Monad-par
An alternative parallel programming API to that provided by the parallel package. The Par monad allows the simple description of parallel computations, and can be used to add parallelism to pure Haskell code. The basic API is straightforward: the monad supports forking and simple communication in terms of IVars.

Data parallelism

Repa
REgular PArallel arrays: high performance, regular, multi-dimensional, shape polymorphic parallel arrays.
Accelerate
An embedded language of array computations for high-performance computing in Haskell. Computations on multi-dimensional, regular arrays are expressed in the form of parameterised collective operations (such as maps, reductions, and permutations).
Data Parallel Haskell
Implicitly parallel, high performance (nested) arrays, supporting large multicore programming.

Research efforts

Feedback-directed implicit parallelism
Implicit parallelism in Haskell, and a feedback-directed mechanism to increase its granularity (FDIP paper)
GpH: Glasgow Parallel Haskell
A complete, GHC-based implementation of the parallel Haskell extension GpH and of evaluation strategies is available. Extensions of the runtime-system and language to improve performance and support new platforms are under development.

Concurrency

Concurrent Haskell

Concurrent Haskell
GHC has supported concurrency with lightweight threads for more than a decade, and it is very fast. Threads in Haskell are preemptively scheduled and support everything you would normally expect from threads, including blocking I/O and foreign calls.

Software transactional memory

Software Transactional Memory
GHC supports a sophisticated version of software transactional memory. Software Transactional Memory (STM) is a new way to coordinate concurrent threads.

Research efforts

Actors with multi-headed receive clauses
Actor-based concurrency for Haskell
CHP: Communicating Haskell Processes
CHP is built on the ideas of CSP (Communicating Sequential Processes), featuring encapsulated parallel processes (no shared data!) communicating over synchronous channels. This is a very composable mode that also allows choice on communications, so that a process may offer to either read on one channel or write on another, but will only take the first that is available.

Distributed programming

MPI

hMPI
hMPI is an acronym for HaskellMPI. It is a Haskell binding conforming to MPI (Message Passing Interface) standard 1.1/1.2. The programmer is in full control over the communication between the nodes of a cluster.
Haskell-MPI
Haskell-MPI provides a Haskell interface to MPI, built on top of the foreign function interface. It is notionally a descendant of hMPI, but is mostly a rewrite.

Distributed Haskell

GdH: Glasgow Distributed Haskell
GdH supports distributed stateful interactions on multiple locations. It is a conservative extension of both Concurrent Haskell and GpH, enabling the distribution of the stateful IO threads of the former on the multiple locations of the latter. The programming model includes forking stateful threads on remote locations, explicit communication over channels, and distributed exception handling.
Eden
Eden extends Haskell with a small set of syntactic constructs for explicit process specification and creation. While providing enough control to implement parallel algorithms efficiently, it frees the programmer from the tedious task of managing low-level details by introducing automatic communication (via head-strict lazy lists), synchronisation, and process handling.
Cloud Haskell
Cloud Haskell is a domain-specific language for developing programs for a distributed computing environment. Implemented as a shallow embedding in Haskell, it provides a message passing communication model, inspired by Erlang, without introducing incompatibility with Haskell’s established shared-memory concurrency.

Research efforts

HCPN: Haskell-Coloured Petri Nets
Haskell-Coloured Petri Nets (HCPN) are an instance of high-level Petri Nets, in which anonymous tokens are replaced by Haskell data objects (and transitions can operate on that data, in addition to moving it around). This gives us a hybrid graphical/textual modelling formalism for Haskell, especially suited for modelling concurrent and distributed systems.