Parallel Computing

1
7/17/2019 Parallel Computing http://slidepdf.com/reader/full/parallel-computing-5690d7b46d7c7 1/1 Parallel computing is a form of computation in which many calculations are carried out simultaneously,[1] operating on the principle that large problems can often be divided into smaller ones, which are then solved at the same time. There are several different forms of parallel computing: bit-level, instruction level, data, and tas parallelism. Parallelism has been employed for many years, mainly in high-performance computing, but interest in it has grown lately due to the physical constraints preventing fre!uency scaling.["] #s power consumption $and conse!uently heat generation% by computers has become a concern in recent years,[&] parallel computing has become the dominant paradigm in computer architecture, mainly in the form of multi-core processors.['] Parallel computing is closely related to concurrent computing ( they are fre!uently used together, and often conflated, though the two are distinct: it is possible to have parallelism without concurrency $such as bit-level parallelism%, and concurrency without parallelism $such as multitasing by time-sharing on a single-core )P*%.[+][] n parallel computing, a computational tas is typically broen down in several, often many, very similar subtass that can be processed independently and whose results are combined afterwards, upon completion. n contrast, in concurrent computing, the various processes often do not address related tass when they do, as is typical in distributed computing, the separate tass may have a varied nature and often re!uire some inter-process communication during e/ecution. Parallel computers can be roughly classified according to the level at which the hardware supports parallelism, with multi-core and multi-processor computers having multiple processing elements within a single machine, while clusters, 0PPs, and grids use multiple computers to wor on the same tas. peciali2ed parallel computer architectures are sometimes used alongside traditional processors, for accelerating specific tass.

description

Parallel Computing

Transcript of Parallel Computing

Page 1: Parallel Computing

7/17/2019 Parallel Computing

http://slidepdf.com/reader/full/parallel-computing-5690d7b46d7c7 1/1

Parallel computing is a form of computation in which many calculations are carried outsimultaneously,[1] operating on the principle that large problems can often be dividedinto smaller ones, which are then solved at the same time. There are several differentforms of parallel computing: bit-level, instruction level, data, and tas parallelism.Parallelism has been employed for many years, mainly in high-performancecomputing, but interest in it has grown lately due to the physical constraints preventing

fre!uency scaling.["] #s power consumption $and conse!uently heat generation% bycomputers has become a concern in recent years,[&] parallel computing has becomethe dominant paradigm in computer architecture, mainly in the form of multi-coreprocessors.[']

Parallel computing is closely related to concurrent computing ( they are fre!uentlyused together, and often conflated, though the two are distinct: it is possible to haveparallelism without concurrency $such as bit-level parallelism%, and concurrencywithout parallelism $such as multitasing by time-sharing on a single-core )P*%.[+][] nparallel computing, a computational tas is typically broen down in several, oftenmany, very similar subtass that can be processed independently and whose resultsare combined afterwards, upon completion. n contrast, in concurrent computing, the

various processes often do not address related tass when they do, as is typical indistributed computing, the separate tass may have a varied nature and often re!uiresome inter-process communication during e/ecution.

Parallel computers can be roughly classified according to the level at which thehardware supports parallelism, with multi-core and multi-processor computers havingmultiple processing elements within a single machine, while clusters, 0PPs, and gridsuse multiple computers to wor on the same tas. peciali2ed parallel computerarchitectures are sometimes used alongside traditional processors, for acceleratingspecific tass.