Partitioned global address space

Partitioned global address space

In parallel computing of computer science, the partitioned global address space (PGAS) is a parallel programming model. It assumes a global memory address space that is logically partitioned and a portion of it is local to each processor. The novelty of PGAS is that the portions of the shared memory space may have an affinity for a particular thread, thereby exploiting locality of reference. The PGAS model is the basis of Unified Parallel C, Co-array Fortran, Titanium, Fortress, Chapel and X10.

External links

* [http://crd.lbl.gov/UPC/images/b/b5/PGAS_Tutorial_sc2003.pdf Programming in the Partitioned Global Address Space Model] (2003)
* [http://www.sdsc.edu/pmac/workshops/geo2006/pubs/Yelick.pdf Performance and Productivity Opportunities using Global Address Space Programming Models] (2006)
* [http://www.ahpcrc.org/conferences/PGAS2006/ Conference on Partitioned Global Address Space Programming Models]
* [http://gasnet.cs.berkeley.edu/ GASNet Communication System] - provides a software infrastructure for PGAS languages over high-performance networks
* [http://projectfortress.sun.com/Projects/Community Sun's project page on Fortress]
* [http://chapel.cs.washington.edu Chapel Programming Language Homepage]


Wikimedia Foundation. 2010.

Игры ⚽ Нужна курсовая?

Look at other dictionaries:

  • Unified Parallel C — (UPC) is an extension of the C programming language designed for high performance computing on large scale parallel machines, including those with a common global address space (SMP and NUMA) and those with distributed memory (eg. clusters). The… …   Wikipedia

  • PGAS — Dieser Artikel wurde aufgrund von inhaltlichen Mängeln auf der Qualitätssicherungsseite der Redaktion Informatik eingetragen. Dies geschieht, um die Qualität der Artikel aus dem Themengebiet Informatik auf ein akzeptables Niveau zu bringen. Hilf… …   Deutsch Wikipedia

  • Multi-core processor — Diagram of a generic dual core processor, with CPU local level 1 caches, and a shared, on die level 2 cache …   Wikipedia

  • Message Passing Interface — MPI, the Message Passing Interface, is standardized and portable message passing system designed by a group of researchers from academia and industry to function on a wide variety of parallel computers. The standard defines the syntax and… …   Wikipedia

  • Co-array Fortran — Coarray Fortran Paradigm(s) multi paradigm: parallel, message passing, imperative (procedural, object oriented), structured Designed by Robert Numrich and John Reid Developer PL22.3 Fortran Committee Stable release …   Wikipedia

  • Locality of reference — In computer science, locality of reference, also known as the principle of locality, is the phenomenon of the same value or related storage locations being frequently accessed. There are two basic types of reference locality. Temporal locality… …   Wikipedia

  • Parallel programming model — A parallel programming model is a set of software technologies to express parallel algorithms and match applications with the underlying parallel systems. It encloses the areas of applications, programming languages, compilers, libraries,… …   Wikipedia

  • OpenMP — Original author(s) OpenMP Architecture Review Board[1] Developer(s) OpenMP Architecture Review Board …   Wikipedia

  • Dataflow programming — Programming paradigms Agent oriented Automata based Component based Flow based Pipelined Concatenative Concurrent computing …   Wikipedia

  • Concurrency (computer science) — The Dining Philosophers , a classic problem involving concurrency and shared resources In computer science, concurrency is a property of systems in which several computations are executing simultaneously, and potentially interacting with each… …   Wikipedia

Share the article and excerpts

Direct link
Do a right-click on the link above
and select “Copy Link”