Asynchronous array of simple processors

Asynchronous array of simple processors

The asynchronous array of simple processors (AsAP) architecture comprises a 2-D array of reduced complexity programmable processors with small memories interconnected by a reconfigurable mesh network. AsAP was developed by researchers in the VLSI Computation Laboratory (VCL) at the University of California, Davis and achieves high performance and high energy-efficiency, while requiring a relatively small circuit area.

AsAP processors are well suited for implementation in future fabrication technologies, and are clocked in aGlobally Asynchronous Locally Synchronous (GALS) fashion. Individual oscillators fully halt (leakage only) in 9 cycles when there is no work to do, and restart at full speed in less than one cycle after work is available. The chip requires no crystal oscillators, PLLs, DLLs, or any global frequency or phase-related signals whatsoever.

The multi-processor architecture efficiently makes use of task-level parallelism in many complex DSP applications, and also efficiently computes many large tasks utilizing fine-grain parallelism.

Key features

AsAP utilizes a number of novel key features of which four are listed here.
* Chip multi-processor (CMP) architecture designed to achieve high performance and low power for many DSP applications.
* Small memories and a simple architecture in each processor to achieve high energy efficiency.
* Globally Asynchronous Locally Synchronous (GALS) clocking simplifies the clock design, greatly increases ease of scalability, and can be used to further reduce power dissipation.
* Inter-processor communication is accomplished using a nearest neighbor network to avoid long global wires and increase scalability to large arrays and in advanced fabrication technologies. Each processor can receive data from any two neighbors and send data to any combination of its four neighbors.

AsAP 1.0 Chip

A chip containing 36 (6x6) programmable processors was taped-out in May 2005 in 0.18μm CMOSusing a synthesized standard cell technology and is fully functional. Processors on the chip operate at clock rates from 520MHz to 540MHz at 1.8V and each processor dissipates 32mW on average while executing applications at 475MHz. Most processors run at clock rates over 600MHz at 2.0V, which makes AsAP among the highest clock rate fabricated processors (programmable or non-programmable) ever designed in a university--it may be the highest ever published.At 0.9V, the average application power is 2.4mW at 116MHz. Each processor occupies only 0.66mm².

Applications

The coding of many DSP and general tasks for AsAP has been completed. Mapped tasks include:filters, convolutional coders, interleavers, sorting, square root, CORDIC sin/cos/arcsin/arccos, matrix multiplication, pseudo random number generators, Fast Fourier Transforms (FFT) of lengths 32-1024, a complete k=7 viterbi decoder, a JPEG encoder, and a complete fully compliant baseband processor for an IEEE 802.11g/11a wireless LAN transmitter.Blocks plug directly together with no required modifications whatsoever. Power, throughput, and area results are typically many times better compared to existing programmable DSP processors.

The architecture enables a clean separation between programming and inter-processor timing handled entirely by hardware. A recently completed C compiler and automatic mapping tool further simplify programming.

References

cite journal
last = Baas | first = Bevan
coauthors = Yu, Zhiyi; Meeuwsen, Michael; Sattari, Omar; Apperson, Ryan; Work, Eric; Webb, Jeremy; Lai, Michael; Mohsenin, Tinoosh; Truong, Dean; Cheung, Jason
title = As

journal = IEEE Micro
volume = 27
issue = 2
date = March/April 2007

cite conference
last = Baas | first = Bevan
coauthors = Yu, Zhiyi; Meeuwsen, Michael; Sattari, Omar; Apperson, Ryan; Work, Eric; Webb, Jeremy; Lai, Michael; Gurman, Daniel; Chen, Chi; Cheung, Jason; Truong, Dean; Mohsenin, Tinoosh
title = Hardware and Applications of As

booktitle = In Proceedings of the IEEE HotChips Symposium on High-Performance Chips, (HotChips 2006)
place = Stanford
date = August 2006
url = http://www.hotchips.org/archives/hc18

cite conference
last = Yu | first = Zhiyi
coauthors = Meeuwsen, Michael; Apperson, Ryan; Sattari, Omar; Lai, Michael; Webb, Jeremy; Work, Eric; Mohsenin, Tinoosh; Singh, Mandeep; Baas, Bevan M.
title = An Asynchronous Array of Simple Processors for DSP Applications
booktitle = In Proceedings of the IEEE International Solid-State Circuits Conference, (ISSCC '06)
place = San Francisco, CA
date = February 2006
pages = 428-429, 663

External links

* [http://www.ece.ucdavis.edu/vcl VLSI Computation Lab, UC Davis]
* [http://www.ece.ucdavis.edu/vcl/asap Asynchronous Array of Simple Processors (AsAP) project]
* [http://www.eet.com/news/latest/showArticle.jhtml?articleID=179102587 EETimes article describing AsAP]


Wikimedia Foundation. 2010.

Игры ⚽ Нужно решить контрольную?

Look at other dictionaries:

  • Multi-core processor — Diagram of a generic dual core processor, with CPU local level 1 caches, and a shared, on die level 2 cache …   Wikipedia

  • ASAP — as an acronym refers to: * As Soon As Possible. It is pronounced letter by letter, or as a single word: A s a p. * As Southern As Possible *Any Surprise, Any Place, a novel Indian company that specializes in offering surprises to one s family… …   Wikipedia

  • ASAP — (As Soon As Possible, en español: Tan pronto como sea posible), es un acrónimo en inglés que se puede referir a: Música y Televisión ASAP (banda), banda de Hard rock creada por el guitarrista y vocalista Adrian Smith. ASAP (canción de T.I.) As… …   Wikipedia Español

  • Oxygene (programming language) — Oxygene Developer RemObjects Software Stable release 3.0.21 (August 29, 2009; 2 years ago (2009 08 29)) Influenced by Object Pas …   Wikipedia

  • Central processing unit — CPU redirects here. For other uses, see CPU (disambiguation). An Intel 80486DX2 CPU from above An Intel 80486DX2 from below …   Wikipedia

  • Algorithmic efficiency — In computer science, efficiency is used to describe properties of an algorithm relating to how much of various types of resources it consumes. Algorithmic efficiency can be thought of as analogous to engineering productivity for a repeating or… …   Wikipedia

  • Message Passing Interface — MPI, the Message Passing Interface, is standardized and portable message passing system designed by a group of researchers from academia and industry to function on a wide variety of parallel computers. The standard defines the syntax and… …   Wikipedia

  • Cellular neural network — Cellular neural networks (CNN) are a parallel computing paradigm similar to neural networks, with the difference that communication is allowed between neighbouring units only. Typical applications include image processing, analyzing 3D surfaces,… …   Wikipedia

  • computer programming language — Introduction       any of various languages for expressing a set of detailed instructions for a digital computer. Such instructions can be executed directly when they are in the computer manufacturer specific numerical form known as machine… …   Universalium

  • Burroughs large systems — The Burroughs large systems were the largest of three series of Burroughs Corporation mainframe computers. Founded in the 1880s, Burroughs was the oldest continuously operating entity in computing, but by the late 1950s its computing equipment… …   Wikipedia

Share the article and excerpts

Direct link
Do a right-click on the link above
and select “Copy Link”