This articleneeds additional citations forverification. Please helpimprove this article byadding citations to reliable sources. Unsourced material may be challenged and removed. Find sources: "Implicit parallelism" – news ·newspapers ·books ·scholar ·JSTOR(February 2024) (Learn how and when to remove this message) |
Incomputer science,implicit parallelism is a characteristic of a programming language that allows acompiler orinterpreter to automatically exploit theparallelism inherent to the computations expressed by some of the language's constructs. A pure implicitly parallel language does not need special directives, operators or functions to enable parallel execution, as opposed toexplicit parallelism.
Programming languages with implicit parallelism includeAxum,BMDFM,HPF,Id,LabVIEW,MATLAB M-code,NESL,SaC,SISAL,ZPL, and pH.[1]
If a particular problem involves performing the same operation on a group of numbers (such as taking thesine orlogarithm of each in turn), a language that provides implicit parallelism might allow the programmer to write the instruction thus:
numbers=[01234567];result=sin(numbers);
The compiler or interpreter can calculate the sine of each element independently, spreading the effort across multiple processors if available.
A programmer that writes implicitly parallel code does not need to worry about task division or process communication, focusing instead on the problem that their program is intended to solve. Implicit parallelism generally facilitates the design of parallel programs and therefore results in a substantial improvement of programmer productivity.
Many of the constructs necessary to support this also add simplicity or clarity even in the absence of actual parallelism. The example above, oflist comprehension in thesin() function, is a useful feature in of itself. By using implicit parallelism, languages effectively have to provide such useful constructs to users simply to support required functionality (a language without a decentfor loop, for example, is one few programmers will use).
Languages with implicit parallelism reduce the control that the programmer has over the parallel execution of the program, resulting sometimes in less-than-optimalparallel efficiency. The makers of theOz programming language also note that their early experiments with implicit parallelism showed that implicit parallelism made debugging difficult and object models unnecessarily awkward.[2]
A larger issue is that every program has some parallel and some serial logic. Binary I/O, for example, requires support for such serial operations asWrite() andSeek(). If implicit parallelism is desired, this creates a new requirement for constructs and keywords to support code that cannot be threaded or distributed.
Thiscomputer science article is astub. You can help Wikipedia byadding missing information. |