Cache

Cache

In computer engineering, a cache (play /ˈkæʃ/ kash[1]) is a component that transparently stores data so that future requests for that data can be served faster. The data that is stored within a cache might be values that have been computed earlier or duplicates of original values that are stored elsewhere. If requested data is contained in the cache (cache hit), this request can be served by simply reading the cache, which is comparatively faster. Otherwise (cache miss), the data has to be recomputed or fetched from its original storage location, which is comparatively slower. Hence, the greater the number of requests that can be served from the cache, the faster the overall system performance becomes.

To be cost efficient and to enable an efficient use of data, caches are relatively small. Nevertheless, caches have proven themselves in many areas of computing because access patterns in typical computer applications have locality of reference. References exhibit temporal locality if data is requested again that has been recently requested already. References exhibit spatial locality if data is requested that is physically stored close to data that has been requested already.

Diagram of a CPU memory cache

Contents

Operation

Hardware implements cache as a block of memory for temporary storage of data likely to be used again. CPUs and hard drives frequently use a cache, as do web browsers and web servers.

A cache is made up of a pool of entries. Each entry has a datum (a nugget (piece) of data) - a copy of the same datum in some backing store. Each entry also has a tag, which specifies the identity of the datum in the backing store of which the entry is a copy.

When the cache client (a CPU, web browser, operating system) needs to access a datum presumed to exist in the backing store, it first checks the cache. If an entry can be found with a tag matching that of the desired datum, the datum in the entry is used instead. This situation is known as a cache hit. So, for example, a web browser program might check its local cache on disk to see if it has a local copy of the contents of a web page at a particular URL. In this example, the URL is the tag, and the contents of the web page is the datum. The percentage of accesses that result in cache hits is known as the hit rate or hit ratio of the cache.

The alternative situation, when the cache is consulted and found not to contain a datum with the desired tag, has become known as a cache miss. The previously uncached datum fetched from the backing store during miss handling is usually copied into the cache, ready for the next access.

During a cache miss, the CPU usually ejects some other entry in order to make room for the previously uncached datum. The heuristic used to select the entry to eject is known as the replacement policy. One popular replacement policy, "least recently used" (LRU), replaces the least recently used entry (see cache algorithm). More efficient caches compute use frequency against the size of the stored contents, as well as the latencies and throughputs for both the cache and the backing store. This works well for larger amounts of data, longer latencies and slower throughputs, such as experienced with a hard drive and the Internet, but is not efficient for use with a CPU cache.[citation needed]

Writing Policies

A Write-Through cache with No-Write Allocation
A Write-Back cache with Write Allocation

When a system writes a datum to cache, it must at some point write that datum to backing store as well. The timing of this write is controlled by what is known as the write policy.


There are two basic writing approaches:

  • Write-through - Write is done synchronously both to the cache and to the backing store.
  • Write-back (or Write-behind) - Writing is done only to the cache. A modified cache block is written back to the store, just before it is replaced.

Write-back cache is more complex to implement, since it needs to track which of its locations have been written over, and mark them as dirty for later writing to the backing store. The data in these locations are written back to the backing store only when they are evicted from the cache, an effect referred to as a lazy write. For this reason, a read miss in a write-back cache (which requires a block to be replaced by another) will often require two memory accesses to service: one to write the replaced data from the cache back to the store, and then one to retrieve the needed datum.

Other policies may also trigger data write-back. The client may make many changes to a datum in the cache, and then explicitly notify the cache to write back the datum.


Since that on write operations, no actual data are needed back, there are two approaches for situations of write-misses:

  • Write allocate (aka Fetch on write) - Datum at the missed-write location is loaded to cache, followed by a write-hit operation. In this approach, write misses are similar to read-misses.
  • No-write allocate (aka Write-no-allocate, Write around) - Datum at the missed-write location is not loaded to cache, and is written directly to the backing store. In this approach, actually only system reads are being cached.

Both write-through and write-back policies can use either of these write-miss policies, but usually they are paired in this way:

  • A write-back cache uses write allocate, hoping for a subsequent writes (or even reads) to the same location, which is now cached.
  • A write-through cache uses no-write allocate. Here, subsequent writes have no advantage, since they still need to be written directly to the backing store.


Entities other than the cache may change the data in the backing store, in which case the copy in the cache may become out-of-date or stale. Alternatively, when the client updates the data in the cache, copies of those data in other caches will become stale. Communication protocols between the cache managers which keep the data consistent are known as coherency protocols.

Applications

CPU cache

Small memories on or close to the CPU can operate faster than the much larger main memory. Most CPUs since the 1980s have used one or more caches, and modern high-end embedded, desktop and server microprocessors may have as many as half a dozen, each specialized for a specific function. Examples of caches with a specific function are the D-cache and I-cache (data cache and instruction cache).

Disk cache

While CPU caches are generally managed entirely by hardware, a variety of software manages other caches. The page cache in main memory, which is an example of disk cache, is managed by the operating system kernel.

While the hard drive's hardware disk buffer is sometimes misleadingly referred to as "disk cache", its main functions are write sequencing and read prefetching. Repeated cache hits are relatively rare, due to the small size of the buffer in comparison to the drive's capacity. However, high-end disk controllers often have their own on-board cache of hard disk data blocks.

Finally, fast local hard disk can also cache information held on even slower data storage devices, such as remote servers (web cache) or local tape drives or optical jukeboxes. Such a scheme is the main concept of hierarchical storage management.

Web cache

Web browsers and web proxy servers employ web caches to store previous responses from web servers, such as web pages. Web caches reduce the amount of information that needs to be transmitted across the network, as information previously stored in the cache can often be re-used. This reduces bandwidth and processing requirements of the web server, and helps to improve responsiveness for users of the web.

Web browsers employ a built-in web cache, but some internet service providers or organizations also use a caching proxy server, which is a web cache that is shared among all users of that network.

Another form of cache is P2P caching, where the files most sought for by peer-to-peer applications are stored in an ISP cache to accelerate P2P transfers. Similarly, decentralised equivalents exist, which allow communities to perform the same task for P2P traffic, for example, Corelli. [2]

Other caches

The BIND DNS daemon caches a mapping of domain names to IP addresses, as does a resolver library.

Write-through operation is common when operating over unreliable networks (like an Ethernet LAN), because of the enormous complexity of the coherency protocol required between multiple write-back caches when communication is unreliable. For instance, web page caches and client-side network file system caches (like those in NFS or SMB) are typically read-only or write-through specifically to keep the network protocol simple and reliable.

Search engines also frequently make web pages they have indexed available from their cache. For example, Google provides a "Cached" link next to each search result. This can prove useful when web pages from a web server are temporarily or permanently inaccessible.

Another type of caching is storing computed results that will likely be needed again, or memoization. ccache, a program that caches the output of the compilation to speed up the second-time compilation, exemplifies this type.

Database caching can substantially improve the throughput of database applications, for example in the processing of indexes, data dictionaries, and frequently used subsets of data.

Distributed caching[3] uses caches spread across different networked hosts, for example, Corelli

The difference between buffer and cache

The terms "buffer" and "cache" are not mutually exclusive and the functions are frequently combined; however, there is a difference in intent.

A buffer is a temporary memory location that is traditionally used because CPU instructions cannot directly address data stored in peripheral devices. Thus, addressable memory is used as an intermediate stage. Additionally, such a buffer may be feasible when a large block of data is assembled or disassembled (as required by a storage device), or when data may be delivered in a different order than that in which it is produced. Also, a whole buffer of data is usually transferred sequentially (for example to hard disk), so buffering itself sometimes increases transfer performance or reduces the variation or jitter of the transfer's latency as opposed to caching where the intent is to reduce the latency. These benefits are present even if the buffered data are written to the buffer once and read from the buffer once.

A cache also increases transfer performance. A part of the increase similarly comes from the possibility that multiple small transfers will combine into one large block. But the main performance-gain occurs because there is a good chance that the same datum will be read from cache multiple times, or that written data will soon be read. A cache's sole purpose is to reduce accesses to the underlying slower storage. Cache is also usually an abstraction layer that is designed to be invisible from the perspective of neighboring layers.

See also

Further reading

References

  1. ^ "Cache - Definition". Merriam-Webster Online Dictionary. Merriam-Webster, Incorporated. http://www.merriam-webster.com/dictionary/cache. Retrieved 2 May 2011. 
  2. ^ Gareth Tyson, Andreas Mauthe, Sebastian Kaune, Mu Mu and Thomas Plagemann. Corelli: A Dynamic Replication Service for Supporting Latency-Dependent Content in Community Networks. In Proc. MMCN'09 [1]
  3. ^ Paul, S; Z Fei (2001-02-01). "Distributed caching with centralized control". Computer Communications 24 (2): 256–268. doi:10.1016/S0140-3664(00)00322-4. 

Wikimedia Foundation. 2010.

Игры ⚽ Нужно решить контрольную?

Look at other dictionaries:

  • caché — caché …   Dictionnaire des rimes

  • Cache — [kæʃ] bezeichnet in der EDV einen schnellen Puffer Speicher, der Zugriffe auf ein langsames Hintergrundmedium oder zeitaufwendige Neuberechnungen nach Möglichkeit vermeidet. Meist werden hierzu Inhalte/Daten gepuffert, die bereits einmal… …   Deutsch Wikipedia

  • Cache — Saltar a navegación, búsqueda Diagrama de una memoria cache de CPU. En informática, una cache o caché (esta última única forma reconocida por la RAE[1] …   Wikipedia Español

  • cache- — ⇒CACHE , élément préf. Élément de compos., déverbal de cacher, servant à la formation de subst. masc., appartenant notamment au vocab. des jeux, de la mode ou de la technol. et désignant des objets ou instruments dont on se sert pour cacher. A.… …   Encyclopédie Universelle

  • Caché — Caché, también escrito cache puede referirse a: Informática Caché (informática), es un conjunto de datos duplicados de otros originales. Caché web, es la que almacena documentos web. Caché Robson Caché de disco DNS cache poisoning Coherencia de… …   Wikipedia Español

  • cache — 1. (ka ch ) s. f. Lieu propre à cacher ou à se cacher. •   On n est pas peu embarrassé à inventer dans toute une maison une cache fidèle, MOL. l Av. I, 4. •   Il dit au roi : Je sais, sire, une cache, Et ne crois pas qu autre que moi la sache, LA …   Dictionnaire de la Langue Française d'Émile Littré

  • caché — caché, ée (ka ché, chée) part. passé. 1°   Dérobé à la vue. Serpent caché sous l herbe, sous terre. Caché aux regards du peuple. J ai tenu ce proscrit caché chez moi. Écueils cachés. Sentiers cachés. Il y a quelque piége caché. •   On m élevait… …   Dictionnaire de la Langue Française d'Émile Littré

  • cache-cœur — [ kaʃkɶr ] n. m. • 1952; de 1. cacher et cœur ♦ Gilet court, à col en V très échancré, croisé sur la poitrine, utilisé dans l habillement ou en layette. Des cache cœurs. ● cache cœur nom masculin invariable Corsage court dont les devants en… …   Encyclopédie Universelle

  • caché — 1. Adaptación gráfica de la voz francesa cachet, usada en español con los sentidos de ‘distinción o elegancia’: «Hay quien nace [...] con estilo y caché» (Vergés Cenizas [R. Dom. 1980]); y ‘cotización o remuneración de un artista’: «¿Qué méritos… …   Diccionario panhispánico de dudas

  • Caché — Données clés Réalisation Michael Haneke Scénario Michael Haneke Acteurs principaux Daniel Auteuil Juliette Binoche Maurice Bénichou Sociétés de production Les Films du Losange …   Wikipédia en Français

  • Cache — Cache, OK U.S. city in Oklahoma Population (2000): 2371 Housing Units (2000): 952 Land area (2000): 3.388615 sq. miles (8.776472 sq. km) Water area (2000): 0.014556 sq. miles (0.037701 sq. km) Total area (2000): 3.403171 sq. miles (8.814173 sq.… …   StarDict's U.S. Gazetteer Places

Share the article and excerpts

Direct link
Do a right-click on the link above
and select “Copy Link”