Skip to main content

Communicating Operators - The Heart of PM


 
Communicating operators lie at the heart of the PM parallelisation model. They are designed to provide a compromise between the direct access to global data structures (particularly arrays) offered by approaches such as Partitioned Global Address Space and the straightforward synchronisation provided by Communicating Sequential Processes. In common with most data-parallel languages, PM contains a parallel version of the for statement which runs all invocations of its enclosed statement list concurrently:

      for element1 in array1, element2 in array2 do  
           element2=process_element(element1)  
      endfor  

Most real models will require some interaction between adjacent array elements. In PM this is achieved by using either a local or a global communicating operator. The global operator @v returns an array whose elements comprise the values of loop-local variable v in each invocation of the enclosing for statement. The neighbourhood operator v@{nbd} provides a more localised array view of v, containing only the local value and those of its neighbours, as defined by the array domain topology.

A very simple example is the implementation of a mean filter over a rectangular array. A 3x3 mean filter simply replaces each element of a grid by the mean of itself and its eight immediate neighbours. In PM this could be coded:

  for element in array do  
    element = sum(element@{-1..1,-1..1})/count(element@{-1..1,-1..1})  
  endfor  

Here sum and count are straightforward array operations. It is necessary to use count rather than simply dividing by nine since off-edge array elements are returned by @ as missing values. The communicating operators both request and supply values for the given variable, thus providing logical synchronisation between concurrent invocations.

There are also reduction versions of the communicating operators, which apply an operation globally or over a neighbourhood, rather than returning an array. The above example may be more efficiently coded as:


  for element in array do  
    element = sum::element@{-1..1,-1..1}/count::element@{-1..1,-1..1}  
  endfor  


The :: @ pair is the reduction version of the neighbourhood @ operator. The global reduction operator just uses ::. To normalise an array so that the sum of its elements sum to one:

 for element in array do  
  element = element/sum::element  
 endfor  


There is more to communicating operators, including their interaction with nested conditional/looping statements and their packaging into ‘loop procedures’. I will come back to this topic in later posts.

 

Comments

Popular posts from this blog

Data in, data out

When a program is running over multiple processors or processing cores then it is important to be able to map where data are flowing. One way in which data flows can be translated into programming language semantics is though argument passing to subroutines. In the absence of global variables, data transfer to and from a subroutine straightforwardly can be directly related to communication to and from another processor. In PM, the emphasis is on data parallelism through the for statement. for index in array do index=process(index) endfor      So what are the rules governing data flow into and out of this statement? These are primarily based around variable locality. x:= 1 for index in array do y:=func(index) index=process(x,y) endfor Here x is relatively global to the for statement and y is local to it (a variables scope only extends to the end of the statement block in which it is defined).  There are two ways t...

Compile time, run time, coffee time

[ Please note that in the following discussion I will be using PM 0.5 notation, which is slightly different to PM 0.4 and also in some flux as PM 0.5 is still in development. ]   Many programming languages (and most newly developed ones) include some form of compile-time programming, ranging from simple preprocessors ( #define , #if in C) to fully blown macro systems capable of re-writing the abstract syntax tree during compilation (Rust, Julia, etc .). In line with its philosophy of keeping things as simple as possible, but not simpler, PM takes a middle road with compile-time programming supported primarily through the type system. There is nothing too radical here – this is not an area where the language aims to break new ground.  The PM approach centres around the use of compile-time types. Many languages use special types to denote literal values and PM follows this trend. Literal integers, reals, strings and Booleans each have their own types: literal(int) , litera...

The PM Type System

In common with other aspects of PM design, the type system is designed to facilitate the use of flexible programming constructs, such as polymorphism and dynamic dispatch, while emphasising the generation of fast, static, code.  In common with many other modern languages, PM also attempts to combine the safety of static typing with the expressivity of dynamically typed languages such as Python . The type of any PM expression may usually be determined by a static examination of the code. Variables take their types from an initialising expression using the ‘:=’ declaration syntax popularised by Go . a:= 100 ! Declare ‘a’ as an integer (int) variable Variables do not change type during the course of their lifetime – polymorphic programming requires the use of special values, described later. Composite values such as structures are generated using specialised expressions rather than through the invocation of type-specific creator functions: b:= struct var_descriptor...