The Massively Parallel Processing System JUMP-1Hidehiko Tanaka, Yōichi Muraoka Ohmsha, 1996 - 235 lappuses The work features the development of the fundamental technologies for massively parallel processing, covering research on the applications, the language, the operating system and the hardware architecture. Also the present status and future plans are addressed. The following topics are discussed in the section on applications: the MGCG Method; Parallelization of FEM; Modeling of Group Behaviors; Parallel Visualization; Functional Memory Type Parallel Processing; a Parallel Reduction Algorithm and Combination Algorithm. As for the programming languages, the SIMD-Based Language NCX, the Dataflow-based Language V and the Parallel Object-Oriented Language A-NETL are discussed. In the chapter on operating systems, the subjects Design Philosophy and Objectives; COS Software Architecture and Elements of the Operating System are - amongst others - addressed. Finally, the part on hardware architecture covers an Overview of the JUMP-1 System; Memory Architecture; Network Architecture; I/O Architecture and Implementation Issues. Massively parallel processing is expected to play a crucial role in the development of almost all advanced technologies for the 21st century. This book is intended to serve a large variety of researchers in the area of parallel computing. |
No grāmatas satura
1.–3. rezultāts no 34.
203. lappuse
... transfer is further reduced by exchanging the vector's order so as to make the best use of the following properties : ( a ) each node can use its own upper torus without message transfer , ( b ) no additional routing is required when ...
... transfer is further reduced by exchanging the vector's order so as to make the best use of the following properties : ( a ) each node can use its own upper torus without message transfer , ( b ) no additional routing is required when ...
204. lappuse
... transfer on the PRDT ( 2 , R ) shown in Figure 5.18 . Since these transfers are performed without any conflict , the hyper- cube emulation can be performed with 2 steps . As described above , the PRDT ( 2 , R ) ( R≤ 4 ) can be emulated ...
... transfer on the PRDT ( 2 , R ) shown in Figure 5.18 . Since these transfers are performed without any conflict , the hyper- cube emulation can be performed with 2 steps . As described above , the PRDT ( 2 , R ) ( R≤ 4 ) can be emulated ...
216. lappuse
... transfer data at a rate proportional to the performance of the system . In parallel processing systems , processed data from each processing element or cluster must be gathered to a display device ( frame memory ) . However , if a large ...
... transfer data at a rate proportional to the performance of the system . In parallel processing systems , processed data from each processing element or cluster must be gathered to a display device ( frame memory ) . However , if a large ...
Citi izdevumi - Skatīt visu
The Massively Parallel Processing System JUMP-1 Hidehiko Tanaka,Yōichi Muraoka Priekšskatījums nav pieejams - 1996 |
Bieži izmantoti vārdi un frāzes
A-NETL address space algorithm allocated architecture array basic block buffer calculated cluster convergence data structures data-parallel Dirichlet boundary condition disk distributed shared memory double-word efficient equation example execution expression Fibonacci tree Fibonet field FIFO fine-grained FMPP function grid hardware highly parallel implementation inner points input interface iteration L2 cache language latency mapping massively parallel computer massively parallel processing memory access memory objects mesh MFlops MGCG method Micro Kernel MIMD mono multigrid method multiple Neumann boundary condition nodes number of processors operating system overhead parallel machines parallel stream partition performance problem Proc processing elements protocol prototype PUX server queue ray tracing reduction computation shown in Figure SIMD simulation speedup STAFF-Link substructure method synchronization T-algorithm Table task topology torus toruses TraceRay update variable vector virtual processors voxel