From Seo Wiki - Search Engine Optimization and Programming Languages
|Designed by||John Hartmann|
|Stable release||110B0020 (2009-01-09)|
|Influenced by||Pipeline (Unix), APL|
A Hartmann pipeline is an extension of the Unix pipeline concept, providing for more complex paths, multiple input/output streams, and other features. It is an example and extension of Pipeline programming.
A Hartmann pipe is a non-procedural representation of a solution of a data processing problem as a dataflow. The error-prone step of translating the dataflow to a traditional procedural programming language is eliminated. Hartmann pipelines may thus be considered as an executable specification language.
The concept was developed by John Hartmann, a Danish engineer with IBM. It is available as a software product CMS Pipelines for a number of IBM platforms. A somewhat backlevel version is included with every level of VM/ESA and z/VM.
A pipeline consists of a collection of stages, joined together by stage separators and connectors. Stages can be written in a variety of languages, and are either filters that process data records or device drivers (sources and sinks) that read data into or out of the pipeline. Unlike other implementations of pipeline programming, Hartmann's design has multiple streams in and out of each stage and can interconnect them non-sequentially. Unlike many programming languages, pipelines have a very small amount of notation, limited to stage separators (typically "
|"), pipeline separators (typically "
;" or "
?"), and label separators ("
:"). Due to common usage, the
diskread stage is also known as
>, however all stages have names that are words in or make some sense in English.
A simple example that reads a disk file, separates records containing the string "Hello" from those that do not, and writes both sets of records to different disk files can be written as:
< input.txt | A: locate /Hello/ | > found.txt ; A: | > notfound.txt
< stage reads the input disk file, the two
> stages write the output disk files, and the
locate stage separates the input stream into two output streams.
locate's primary output (records containing Hello) is passed to the first
> stage, and its secondary output (records not containing Hello) is passed through the
A: connector to the second
> stage. The ; divides the specification into 2 pipelines. The collection of pipelines is called a pipeline set.
Some of the salient characteristics that distinguish Hartmann Pipeline from ordinary Unix pipes are:
- Filters may have multiple inputs and multiple outputs. For example, a selection filter can send the found records down one output pipe and the not found records down another.
- A linear notation for representing pipeline networks.
- An interface that allows REXX programs to act as stages.
- A pacing strategy in the Pipeline supervisor that allows, for example, a stream to be split, say by a selection filter, and the records on the output legs to be processed by other filters, then merged by a join filter and have the record order preserved in result stream.
- As implied by the previous item, data streams are (generally) not simply buffered and passed along to the next filter. The filters operate in parallel with input and output records handled by the Pipeline supervisor.
Similarity to APL
Programmers familiar with the APL programming language will see some similarities in Hartmann pipelines. It is obvious that the author was influenced by APL; some of the filters have names and functions similar to specific APL primitive functions. Examples include the TAKE filter, which passes a specified number of records, and the DEAL filter, which spreads its input records out across its output streams, in imitation of the APL deal operator.
As with APL, programmers adept in the use of pipelines will have their view of data processing problems and how they may be best solved fundamentally and permanently changed.
- ↑ Melinda Varian (November 1995) (PDF). Plunging Into Pipes. http://vm.marist.edu/~pipeline/plunge.pdf. Retrieved 2006-11-08.