|| This section needs additional citations for verification.
Please help improve this article by adding reliable references. Unsourced material may be challenged and removed. (August 2009) </td>
The first incarnation of what was later to be the APL programming language was A Programming Language, a book describing a notation invented in 1957 by Kenneth E. Iverson while at Harvard University. Published in 1962, the notation described in the book was recognizable yet distinct from APL.
IBM was chiefly responsible for the introduction of APL to the marketplace. In 1965, a portion of the notation was reworked and implemented as a programming language known as IVSYS. APL was first available in 1967 for the IBM 1130 as APL\1130. It would run in as little as 8k 16 bit words of memory, and used a dedicated 1 megabyte hard disk. APL gained its foothold on mainframe timesharing systems from the late 1960s through the 1980s. Later, when suitably performing hardware was finally available starting in the early to mid-1980s, many users migrated their applications to the personal computer environment.
Early IBM APL interpreters for IBM 360 and IBM 370 hardware implemented their own multi-user management instead of relying on the host services, thus they were timesharing systems in their own right. First introduced in 1966, the APL\360 system was a multi-user interpreter. In 1973, IBM released APL.SV which was a continuation of the same product, but which offered shared variables as a means to access facilities outside of the APL system, such as operating system files. In the mid 1970s, the IBM mainframe interpreter was even adapted for use on the IBM 5100 desktop computer, which had a small CRT and an APL keyboard, when most other small computers of the time only offered BASIC. In the 1980s, the VSAPL program product enjoyed widespread usage with CMS, TSO, VSPC, and CICS users.
Several timesharing firms sprang up in the 1960s and 1970s which sold APL services using modified versions of the IBM APL\360 interpreter. In North America, the better-known ones were I. P. Sharp Associates, Scientific Time Sharing Corporation, and The Computer Company (TCC). With the advent first of less expensive mainframes such as the IBM 4331 and later the personal computer, the timesharing industry had all but disappeared by the mid 1980s.
Sharp APL was available from I. P. Sharp Associates, first on a timesharing basis in the 1960s, and later as a program product starting around 1979. Sharp APL was an advanced APL implementation with many language extensions, such as packages (the ability to put one or more objects into a single variable), file system, nested arrays, and shared variables.
APL interpreters were available from other mainframe and mini-computer manufacturers as well, notably Burroughs, CDC, Data General, DEC, Harris, Hewlett-Packard, Siemens, Xerox, and others.
In 1979, Iverson received the Turing Award for his work on APL.
Starting in the early 1980s, IBM APL development, under the leadership of Dr Jim Brown, implemented a new version of the APL language which contained as its primary enhancement the concept of nested arrays where an array may contain other arrays, plus new language features which facilitated the integration of nested arrays into program workflow. Ken Iverson, no longer in control of the development of the APL language, left IBM and joined I. P. Sharp Associates where he, among other things, directed the evolution of Sharp APL to be more in accordance with his vision.
As other vendors were busy developing APL interpreters for new hardware, notably Unix-based microcomputers, APL2 was almost always the standard chosen for new APL interpreter developments. Even today, most APL vendors cite APL2 compatibility, which only approaches 100%, as a selling point for their products.
APL2 for IBM mainframe computers is still available today, and was first available for CMS and TSO in 1984. The APL2 Workstation edition (Windows, OS/2, AIX, Linux, and Solaris) followed much later in the early 1990s.
The first microcomputer implementation of APL was on the 8008-based MCM/70, the first general purpose personal computer, in 1973. A Small APL for the Intel 8080 called EMPL was released in 1977, and Softronics APL, with most of the functions of full APL, for 8080-based CP/M systems was released in 1979.
In 1977, the Canadian firm Telecompute Integrated Systems, Inc. released a business-oriented APL interpreter known as TIS APL, for Z80-based systems. It featured the full set of file functions for APL, plus a full screen input and switching of right and left arguments for most dyadic operators by introducing ~. prefix to all single character dyadic functions such as - or /.
Vanguard APL was available for Z80 CP/M-based processors in the late 1970s. TCC released APL.68000 in the early 1980s for Motorola 68000-based processors, this system being the basis for MicroAPL Limited's APLX product. I. P. Sharp Associates released a version of their APL interpreter for the IBM PC and PC/370 - for the IBM PC, an emulator was written which facilitated reusing much of the IBM 370 mainframe code. Arguably, the best known APL interpreter for the IBM Personal Computer was STSC's APL*Plus/PC.
The Commodore SuperPET, introduced in 1981, included an APL interpreter developed by the University of Waterloo.
In the early 1980s, the Analogic Corporation developed The APL Machine, which was an array processing computer designed to be programmed only in APL. There were actually three processing units, the user's workstation, an IBM PC, where programs were entered and edited, a Motorola 6800 processor which ran the APL interpreter, and the Analogic array processor which executed the primitives. At the time of its introduction The APL Machine was likely the fastest APL system available. Although a technological success, The APL Machine was a marketing failure. The initial version supported a single process at a time. At the time the project was discontinued, the design had been completed to allow multiple users. As an aside, an unusual aspect of The APL Machine was that the library of workspaces was organized such that a single function or variable which was shared by many workspaces existed only once in the library. Several of the members of The APL Machine project had previously spent a number of years with Burroughs implementing APL\700.
At one stage, Microsoft Corporation planned to release a version of APL, but these plans never materialized.
An early 1978 publication of Rodnay Zaks from Sybex was A microprogrammed APL implementation ISBN 0895880059 which is the complete, total source listing for the microcode for a PDP / LSI-11 processor implementing APL. This may have been the substance of his PhD thesis.
Recent extensions to APL include:
Over a very wide set of problem domains (math, science, engineering, computer design, robotics, data visualization, actuarial science, traditional DP, etc.) APL is an extremely powerful, expressive and concise programming language, typically set in an interactive environment. It was originally created, among other things, as a way to describe computers, by expressing mathematical notation in a rigorous way that could be interpreted by a computer. It is easy to learn but some APL programs can take some time to understand, especially for a newcomer. Few other programming languages offer the comprehensive array functionality of APL.
Unlike traditionally structured programming languages, code in APL is typically structured as chains of monadic or dyadic functions and operators acting on arrays. As APL has many nonstandard primitives (functions and operators, indicated by a single symbol or a combination of a few symbols), it does not have function or operator precedence. Early APL implementations did not have control structures (do or while loops, if-then-else), but by using array operations, usage of structured programming constructs was just not necessary. For example, the iota function (which yields a one-dimensional array, or vector, from 1 to N) can replace for-loop iteration. More recent implementations of APL generally include comprehensive control structures, thus data structure and program control flow can be clearly and cleanly separated.
The APL environment is called a workspace. In a workspace the user can define programs and data, i.e. the data values exist also outside the programs, and the user can manipulate the data without the necessity to define a program. For example,
- <math>N \leftarrow 4\ 5\ 6\ 7</math>
assigns the vector values 4 5 6 7 to N;
adds 4 to all values (giving 8 9 10 11) and prints them (a return value not assigned at the end of a statement to a variable using the assignment arrow <math>\leftarrow </math> is displayed by the APL interpreter);
prints the sum of N, i.e. 22.
The user can save the workspace with all values, programs and execution status.
APL is well-known for its use of a set of non-ASCII symbols that are an extension of traditional arithmetic and algebraic notation. Having single character names for SIMD vector functions is one way that APL enables compact formulation of algorithms for data transformation such as computing Conway's Game of Life in one line of code (example). In nearly all versions of APL, it is theoretically possible to express any computable function in one expression, that is, in one line of code.
Because of its condensed nature and non-standard characters, APL has sometimes been termed a "write-only language", and reading an APL program can at first feel like decoding Egyptian hieroglyphics. Because of the unusual character set, many programmers use special keyboards with APL keytops for authoring APL code. Although there are various ways to write APL code using only ASCII characters, in practice, it is almost never done. (This may be thought to support Iverson’s thesis about notation as a tool of thought.) Most if not all modern implementations use standard keyboard layouts, with special mappings or input method editors to access non-ASCII characters. Historically, the APL font has been distinctive, with uppercase italic alphabetic characters and upright numerals and symbols. Most vendors continue to display the APL character set in a custom font.
Advocates of APL claim that the examples of so-called write-only code are almost invariably examples of poor programming practice or novice mistakes, which can occur in any language. Advocates of APL also claim that they are far more productive with APL than with more conventional computer languages, and that working software can be implemented in far less time and with far fewer programmers than using other technology. APL lets an individual solve harder problems faster. Also, being compact and terse, APL lends itself well to larger scale software development as complexity arising from a large number of lines of code can be dramatically reduced. Many APL advocates and practitioners view programming in standard programming languages, such as COBOL and Java, as comparatively tedious. APL is often found where time-to-market is important, such as with trading systems.
Iverson later designed the J programming language which uses ASCII with digraphs instead of special symbols.
Today, most APL language activity takes place under the Microsoft Windows operating system, with some activity under Linux, Unix, and Mac OS. Comparatively little APL activity takes place today on mainframe computers.
APLNow (formerly APL2000) offers an advanced APL interpreter which operates under Linux, Unix, and Windows. It supports Windows automation, supports calls to operating system and user defined DLLs, has an advanced APL File System, and represents the current level of APL language development. APL2000's product is an advanced continuation of STSC's successful APL*Plus/PC and APL*Plus/386 product line.
Dyalog APL is an advanced APL interpreter which operates under Linux, Unix, and Windows. Dyalog has innovative extensions to the APL language which include new object oriented features, numerous language enhancements, plus a consistent namespace model used for both its Microsoft Automation interface, as well as native namespaces. For the Windows platform, Dyalog APL offers tight integration with Microsoft .Net, plus limited integration with the Microsoft Visual Studio development platform.
IBM offers a version of IBM APL2 for IBM AIX, Linux, Sun Solaris and Windows systems. This product is a continuation of APL2 offered for IBM mainframes. IBM APL2 was arguably the most influential APL system, which provided a solid implementation standard for the next set of extensions to the language, focusing on nested arrays.
NARS2000 is an open source APL interpreter written by Bob Smith, a well-known APL developer and implementor from STSC in the 1970s and 1980s. NARS2000 contains advanced features and new datatypes, runs natively under Windows (32- and 64-bit versions), and runs under Linux and Apple Mac OS with Wine.
MicroAPL Limited offers APLX, a full-featured 64 bit interpreter for Linux, Windows, and Apple Mac OS systems.
Soliton Associates offers the SAX interpreter (Sharp APL for Unix) for Unix and Linux systems, which is a further development of I. P. Sharp Associates' Sharp APL product. Unlike most other APL interpreters, Kenneth E. Iverson had some influence in the way nested arrays were implemented in Sharp APL and SAX. Nearly all other APL implementations followed the course set by IBM with APL2, thus some important details in Sharp APL differ from other implementations.
APL programs are normally interpreted and less often compiled. In reality, most APL compilers translated source APL to a lower level language such as C, leaving the machine-specific details to the lower level compiler. Compilation of APL programs was a frequently discussed topic in conferences. Although some of the newer enhancements to the APL language such as nested arrays have rendered the language increasingly difficult to compile, the idea of APL compilation is still under development today.
In the past, APL compilation was regarded as a means to achieve execution speed comparable to other mainstream languages, especially on mainframe computers.
Several APL compilers achieved some levels of success, though comparatively little of the development effort spent on APL over the years went to perfecting compilation into machine code.
As is the case when moving APL programs from one vendor's APL interpreter to another, APL programs invariably will require changes to their content. Depending on the compiler, variable declarations might be needed, certain language features would need to be removed or avoided, or the APL programs would need to be cleaned up in some way. Some features of the language, such as the execute function (an expression evaluator) and the various reflection and introspection functions from APL, such as the ability to return a function's text or to materialize a new function from text, are simply not practical to implement in machine code compilation.
A commercial compiler was brought to market by STSC in the mid 1980s as an add-on to IBM's VSAPL Program Product. Unlike more modern APL compilers, this product produced machine code which would execute only in the interpreter environment, it was not possible to eliminate the interpreter component. The compiler could compile many scalar and vector operations to machine code, but it would rely on the APL interpreter's services to perform some more advanced functions, rather than attempt to compile them. However, dramatic speedups did occur, especially for heavily iterative APL code.
Around the same time, the book An APL Compiler by Timothy Budd appeared in print. This book detailed the construction of an APL translator, written in C, which performed certain optimizations such as loop fusion specific to the needs of an array language. The source language was APL-like in that a few rules of the APL language were changed or relaxed to permit more efficient compilation. The translator would emit C code which could then be compiled and run well outside of the APL workspace.
Today, execution speed is less critical and many popular languages are implemented using virtual machines - instructions that are interpreted at runtime. The Burroughs/Unisys APLB interpreter (1982) was the first to use dynamic incremental compilation to produce code for an APL-specific virtual machine. It recompiled on-the-fly as identifiers changed their functional meanings. In addition to removing parsing and some error checking from the main execution path, such compilation also streamlines the repeated entry and exit of user-defined functional operands. This avoids the stack setup and take-down for function calls made by APL's built-in operators such as Reduce and Each.
APEX, a research APL compiler, is available from Snake Island Research Inc. APEX compiles flat APL (a subset of ISO N8485) into SAC, a functional array language with parallel semantics, and currently runs under Linux. APEX-generated code uses loop fusion and array contraction, special-case algorithms not generally available to interpreters (e.g., upgrade of permutation vector), to achieve a level of performance comparable to that of Fortran.
The APLNext VisualAPL system is a departure from a conventional APL system in that VisualAPL is a true .Net language which is fully inter-operable with other .Microsoft .Net languages such as VB.Net and C#. VisualAPL is inherently object oriented and Unicode-based. While VisualAPL incorporates most of the features of standard APL implementations, the VisualAPL language extends standard APL to be .Net-compliant. VisualAPL is hosted in the standard Microsoft Visual Studio IDE and as such, invokes compilation in a manner identical to that of other .Net languages. By producing .Net common language runtime (CLR) code, it utilizes the Microsoft just-in-time compiler (JIT) to support 32-bit or 64-bit hardware. Substantial performance speed-ups over standard APL have been reported, especially when (optional) strong typing of function arguments is used.
An APL to C# translator is available from Causeway Graphical Systems. This product was designed to allow the APL code, translated to equivalent C#, to run completely outside of the APL environment. The Causeway compiler requires a run-time library of array functions. Some speedup, sometimes dramatic, is visible, but happens on account of the optimisations inherent in Microsoft's .Net framework.
A source of links to existing compilers is at APL2C.
 Matrix optimizations
APL was unique in the speed with which it could perform complicated matrix operations. For example, a very large matrix multiplication would take only a few seconds on a machine which was much less powerful than those today. There were both technical and economic reasons for this advantage:
- Commercial interpreters delivered highly-tuned linear algebra library routines.
- Very low interpretive overhead was incurred per-array—not per-element.
- APL response time compared favorably to the runtimes of early optimizing compilers.
- IBM provided microcode assist for APL on a number of IBM/370 mainframes.
A widely cited paper "An APL Machine" (authored by Phil Abrams) perpetuated the myth that APL made pervasive use of lazy evaluation where calculations would not actually be performed until the results were needed and then only those calculations strictly required. An obvious (and easy to implement) lazy evaluation is the J-vector : when a monadic iota is encountered in the code, it is kept as a representation instead of being calculated at once, thus saving some time as well as memory.
Although this technique was not generalized, it embodies the language's best survival mechanism: not specifying the order of scalar operations. Even as eventually standardized by ANSI working group X3J10, APL is so highly data-parallel, it gives language implementors immense freedom to schedule operations as efficiently as possible. As computer innovations such as cache memory, and SIMD execution became commercially available, APL programs ported with little extra effort spent re-optimizing low-level details.
APL makes a clear distinction between functions and operators. Functions take arrays (variables or constants or expressions) as arguments, and return arrays as results. Operators (similar to higher-order functions) take functions or arrays as arguments, and derive related functions. For example the "sum" function is derived by applying the "reduction" operator to the "addition" function. Applying the same reduction operator to the "maximum" function (which returns the larger of two numbers) derives a function which returns the largest of a group (vector) of numbers. In the J language, Iverson substituted the terms 'verb' & ('adverb' or 'conjunction') for 'function' and 'operator'.
APL also identifies those features built into the language, and represented by a symbol, or a fixed combination of symbols, as primitives. Most primitives are either functions or operators. Coding APL is largely a process of writing non-primitive functions and (in some versions of APL) operators. However a few primitives are considered to be neither functions nor operators, most noticeably assignment.
This displays "Hello, world":
The following expression sorts a word list stored in matrix X according to word length:
This following immediate-mode expression generates a typical set of Pick 6 lottery numbers: six pseudo-random integers ranging from 1 through 40, guaranteed non-repeating, and displays them sorted in ascending order:
This combines the following APL functions:
- The first to be executed (APL executes from right to left) is the dyadic function “?” (named “Deal” when dyadic) that returns a vector consisting of a select number (left argument: 6 in this case) of random integers ranging from 1 through a specified maximum (right argument: 40 in this case), which, if said maximum ≥ vector length, is guaranteed to be non-repeating.
- This vector is then assigned to the variable x, because it is needed later.
- This vector is then sorted in ascending order by the monadic “⍋” function, which has as its right argument everything to the right of it up to the next unbalanced close-bracket or close-parenthesis. The result of ⍋ is the indices which will put its argument into ascending order.
- Then the output of ⍋ is applied to the variable x which we saved earlier, and it puts the items of x into ascending sequence.
Since there is no function to the left of the first x to tell APL what to do with the result, it simply outputs it to the display (on a single line, separated by spaces) without needing any explicit instruction to do that.
(Note that “?” also has a monadic equivalent called “Roll” which simply returns a single random integer between 1 and its sole operand [to the right of it], inclusive. Thus, a Role-Playing Game program might use the expression “?20” to roll a twenty-sided die.)
The equivalent of the above in a traditional imperative/procedural programming language might follow an algorithm similar to this:
from random import shuffle # import randomization library
one_thru_40 = range(1, 41) # generate list of numbers 1 thru 40 inclusive
shuffle(one_thru_40) # randomly reorder the list
numbers = one_thru_40[:6] # pick the first 6 elements from the list
numbers.sort() # sort the numbers in the list
print(" ".join(str(num) for num in numbers)) # convert numbers to strings and output with spaces between them
The following function "life", written in Dyalog APL, takes a boolean matrix and calculates the new generation according to Conway's Game of Life:
In the following example, also Dyalog, the first line assigns some HTML code to a variable "txt" and then uses an APL expression to remove all the HTML tags, returning the text only as shown in the last line.
The following expression finds all prime numbers from 1 to R. In both time and space, the calculation complexity is O(R²) (in Big O notation).
Executed from right to left, this means:
- ιR creates a vector containing integers from 1 to R (if R = 6 at the beginning of the program, ιR is 1 2 3 4 5 6)
- Drop first element of this vector (↓ function), i.e. 1. So 1↓ιR is 2 3 4 5 6
- Set R to the new vector (←, assignment primitive), i.e. 2 3 4 5 6
- Generate outer product of R multiplied by R, i.e. a matrix which is the multiplication table of R by R (°.× function), i.e.
- Build a vector the same length as R with 1 in each place where the corresponding number in R is in the outer product matrix (∈, set inclusion function), i.e. 0 0 1 0 1
- Logically negate the values in the vector (change zeros to ones and ones to zeros) (∼, negation function), i.e. 1 1 0 1 0
- Select the items in R for which the corresponding element is 1 (/ function), i.e. 2 3 5
(Note, this assumes the APL origin is 1, ie, indices start with 1. APL can be set to use 0 as the origin, which is convenient for some calculations.)
 Character set
Main article: APL (codepage)
APL has always been criticized for its choice of a unique, non-standard character set. The observation that some who learn it usually become ardent adherents shows that there is some weight behind Iverson's idea that the notation used does make a difference. In the beginning, there were few terminal devices which could reproduce the APL character set—the most popular ones employing the IBM Selectric print mechanism along with a special APL type element. Over time, with the universal use of high-quality graphic display, printing devices and Unicode support, the APL character font problem has largely been eliminated; however, the problem of entering APL characters requires the use of input method editors or special keyboard mappings, which may frustrate beginners accustomed to other programming languages.
APL has long had a small and fervent user base. It was and still is popular in financial and insurance applications, in simulations, and in mathematical applications. APL has been used in a wide variety of contexts and for many and varied purposes. A newsletter titled "Quote-Quad" dedicated to APL has been published since the 1970s by the SIGAPL section of the Association for Computing Machinery (Quote-Quad is the name of the APL character used for text input and output).
Before the advent of full-screen systems and until as late as the mid-1980s systems were written such that the user entered instructions in his own business specific vocabulary. APL timesharing vendors delivered applications in this form. On the I. P. Sharp timesharing system, a workspace called 39 MAGIC offered access to financial and airline data plus sophisticated (for the time) graphing and reporting. Another example is the GRAPHPAK workspace supplied with IBM's APL2.
Because of its matrix operations, APL was for some time quite popular for computer graphics programming, where graphic transformations could be encoded as matrix multiplications. One of the first commercial computer graphics houses, Digital Effects, based in New York City, produced an APL graphics product known as "Visions," which was used to create television commercials and film animation for the 1982 film Tron (Digital Effects use of APL was informally described at a number of SIGAPL conferences in the late 1980s, examples discussed included the early UK Channel 4 TV logo/ident. What is not clear is the extent to which APL was directly involved in the making of "Tron", and at this point in time the reference is more of an urban legend or historic curio than much else).
Interest in APL has steadily declined since the mid 1980s. This was partially due to the lack of a smooth migration path from higher performing mainframe implementations to low-cost personal computer alternatives, as APL implementations for computers before the Intel 80386 were suitable for small applications. The growth of end-user computing tools such as Microsoft Excel and Microsoft Access also eroded into potential APL usage. These are appropriate platforms for what may have been mainframe APL applications in the 1970s and 1980s. Some APL users migrated to the J programming language, which offers more advanced features. Lastly, the decline was also due in part to the growth of MATLAB, GNU Octave, and Scilab. These scientific computing array-oriented platforms provide an interactive computing experience similar to APL, but more resemble conventional programming languages such as Fortran, and use standard ASCII.
Notwithstanding this decline, APL finds continued use in certain fields, such as accounting research.
APL has been standardized by the ANSI working group X3J10 and ISO/IEC Joint Technical Committee 1 Subcommittee 22 Working Group 3. The Core APL language is specified in ISO 8485:1989, and the Extended APL language is specified in ISO/IEC 13751:2001.
Some words used in APL literature have meanings that differ from those in both mathematics and the generality of computer science.
||operation or mapping that takes zero, one (right) or two (left & right) array valued arguments and may return an array valued result. A function may be:
|Primitive - built-in and represented by a single glyph;
|Defined - as a named and ordered collection of program statements;
|Derived - as a combination of an operator with it arguments.
||data valued object of zero or more orthogonal dimensions in row-major order in which each item is a primitive scalar datum or another array.
||not taking or requiring any arguments,
||requiring only one argument; on the right for a function, on the left for an operator, unary
||requiring both a left and a right argument, binary
|ambivalent or nomadic
||capable of use in a monadic or dyadic context, permitting its left argument to be elided
||operation or mapping that takes one (left) or two (left & right) function or array valued arguments (operands) and derives a function. An operator may be:
|Primitive - built-in and represented by a single glyph;
|Defined - as a named and ordered collection of program statements.
 See also
- ↑ 1.0 1.1 "A Bibliography of APL and J". Jsoftware.com. http://www.jsoftware.com/jwiki/Essays/Bibliography. Retrieved 2010-02-03.
- ↑ 2.0 2.1 "Kx Systems - An Interview with Arthur Whitney - Jan 2004". Kx.com. 2004-01-04. http://kx.com/Company/press-releases/arthur-interview.php. Retrieved 2010-02-03.
- ↑ 3.0 3.1 "About Q'Nial". Nial.com. http://www.nial.com/AboutQNial.html. Retrieved 2010-02-03.
- ↑ 4.0 4.1 4.2 "The Growth of MatLab - Cleve Moler" (PDF). http://www.mathworks.com/company/newsletters/news_notes/clevescorner/jan06.pdf. Retrieved 2010-02-03.
- ↑ 5.0 5.1 Iverson, Kenneth E. (1962). A Programming Language. Wiley. ISBN 0471430145.
- ↑ "an experimental APL interpreter". NARS2000. http://www.nars2000.org/. Retrieved 2010-02-03.
- ↑ "Dyalog V12 Platforms". Dyalog.com. http://www.dyalog.com/version12.html. Retrieved 2010-02-03.
- ↑ 8.0 8.1 "The Future of APL in the Insurance World. Gary A. Bergquist". Portal.acm.org. http://portal.acm.org/ft_gateway.cfm?id=347203&type=pdf&coll=GUIDE&dl=GUIDE&CFID=46163917&CFTOKEN=91975100. Retrieved 2010-02-03.
- ↑ "APLX version 4 – from the viewpoint of an experimental physicist. Vector 23.3". Vector.org.uk. 2008-05-20. http://www.vector.org.uk/archive/v233/webber.htm. Retrieved 2010-02-03.
- ↑ OOSTATS – A New Approach to Statistics via APL
- ↑ "ACM Award Citation – John Backus. 1977". Awards.acm.org. 1924-12-03. http://awards.acm.org/citation.cfm?id=0703524&srt=all&aw=140&ao=AMTURING. Retrieved 2010-02-03.
- ↑ "Agile Approach". Agile Approach. http://www.agileapproach.co.uk/. Retrieved 2010-02-03.
- ↑ Larry Breed (August 2006). "How We Got to APL\1130". Vector (British APL Association) 22 (3). ISSN 0955-1433. http://www.vector.org.uk/archive/v223/APL_1130.htm.
- ↑ 14.0 14.1 Pakin, Sandra (1968). APL\360 Reference Manual. Science Research Associates, Inc.. ISBN 0-574-16135-X.
- ↑ "Turing Award Citation 1979". Awards.acm.org. http://awards.acm.org/citation.cfm?id=9147499&srt=all&aw=140&ao=AMTURING. Retrieved 2010-02-03.
- ↑ Falkoff, Adin D. (1991). "The IBM family of APL systems" (PDF). IBM Systems Journal (IBM) 30 (4): 416–432. Archived from the original on (unknown). http://184.108.40.206/search?q=cache:WGQh6w1-YHAJ:www.research.ibm.com/journal/sj/304/ibmsj3004C.pdf. Retrieved June 13, 2009.
- ↑ 17.0 17.1 17.2 "APLX : New features in Version 5". Microapl.co.uk. 2009-04-01. http://www.microapl.co.uk/apl/aplxv5.html. Retrieved 2010-02-03.
- ↑ "APL2000 NetAccess". Apl2000.com. http://www.apl2000.com/netaccess.php. Retrieved 2010-02-03.
- ↑ "Introduction to Visual APL". Visualapl.com. http://www.visualapl.com/library/aplnext/Visual_APLNext.htm. Retrieved 2010-02-03.
- ↑ "Dyalog Built-in Charting". Dyalog.com. http://www.dyalog.com/version12.html. Retrieved 2010-02-03.
- ↑ "XML-to-APL Conversion tool in Dyalog APL and APLX". Dyalog.com. http://www.dyalog.com/help/12.1/html/xml%20convert.htm. Retrieved 2010-02-03.
- ↑ "Dynamic Functions in Dyalog APL - John Scholes. 1997" (PDF). http://www.dyalog.com/download/dfns.pdf. Retrieved 2010-02-03.
- ↑ Dickey, Lee, A list of APL Transliteration Schemes, 1993
- ↑ Iverson K.E.,"Notation as a tool of thought", Communications of the ACM, 23: 444-465 (August 1980).
- ↑ "Stanford Accounting PhD requirements". Gsb.stanford.edu. http://www.gsb.stanford.edu/phd/fields/accounting/index.html. Retrieved 2010-02-03.
- ↑ 26.0 26.1 26.2 26.3 26.4 26.5 "APL concepts". Microapl.co.uk. http://www.microapl.co.uk/APL/apl_concepts_chapter6.html. Retrieved 2010-02-03.
- ↑ "Nested array theory". Nial.com. http://www.nial.com/ArrayTheory.html. Retrieved 2010-02-03.
- ↑ 28.0 28.1 28.2 "Programmera i APL", Bohman, Fröberg, Studentlitteratur, ISBN-91-44-13162-3
 Further reading
 External links
ko:APL (프로그래밍 언어)
ms:Bahasa pengaturcaraan APL
pl:APL (język programowania)
ru:АПЛ (язык программирования)
sk:APL (programovací jazyk)
tg:АПЛ (забони барноманависӣ)