answersLogoWhite

0


Best Answer

Cache memory necessarily has a smaller latency than RAM. The entire purpose of a memory cache is to reduce memory-access overhead by exploiting the principle of 'program locality' (the idea that any datum that is used at a particular point in a program is highly likely to be used again within a short period).

There are several levels of cache used in modern architectures, both on and off-chip. A modern processor will have at least a level-1 and level-2 cache on-die, with chips like the AMD Phenom II and the Intel Core i7 having L3 caches also on-chip.

The speed difference between an L1 cache and RAM latency can be anywhere between 5x to 100x faster.

User Avatar

Wiki User

14y ago
This answer is:
User Avatar
More answers
User Avatar

Wiki User

10y ago

Yep. Sandy Bridge's onboard cache operates around 50 GBps versus my RAM runs at 5 GBps read speed and 8 GBps write speed.

This answer is:
User Avatar

Add your answer:

Earn +20 pts
Q: Cache memory is faster to access than RAM?
Write your answer...
Submit
Still have questions?
magnify glass
imp
Related questions

Is Cache memory generally slower than RAM?

Cache memory is random access memory (RAM) that a computer microprocessor can access more quickly than it can access regular RAM. As the microprocessor processes data, it looks first in the cache memory and if it finds the data there (from a previous reading of data), it does not have to do the more time-consuming reading of data from larger memory. Cache Memory generally comes in smaller size (3MB, 6MB etc) than the RAM (512 MB, 1GB,2GB.....)


Cache memory and register memory?

Register memory are smaller in size than cache memory and registers are faster than cache..Cache memory store the frequently used data from main memory..


What is the function of CPU cache?

Many CPUs have what is known as a CPU cache. The function of this CPU cache is to speed up access to data.


A memory cache is small amount of RAM that is much faster than the rest of RAM?

A memory cache is small amount of RAM that is much faster than the rest of RAM?


What is a cache?

Answer:- The Cache memory is placed between the CPU and the main memory. It is a fast speed memory and is expensive and faster than the main memory. Cache memory is used to store the frequently accessed data of main memory. The instructions that are frequently used by the CPU are stored in the cache memory. It is used to reduce the average access time for address, instructions or data, which are normally stored in the main memory. Cache memory increases the operating speed of the system. But is much costlier than main memory. From economic considerations, the capacity of the cache memory is much less as compared to main memory..


Is cache memory a removable memory?

No, a cache memory is often used to store data that has been needed recently on grounds that it will be faster to access when/if it is needed again. When data that is requested is contained in the cache you have a cache hit, and when you have to retrieve it from the hard drive (or where ever its original storage was) again it is called a cache miss. Retrieving data from the hard drive is slower than retrieving it from the cache.


Why is cache memory more expensive than hard disk?

as the cache memory is mory faster than RAM or hard disk ,but the main reason is that the cache memory has comparator and storage medium at the same time, the comparator checks if the address of the value being accessed is in the associative memory(a part of cache)or not, and the ordinary memory(another part of cache)has the data....


What are Limitations of cache Memory?

size. cache memory is situated on the CPU, and stores recently used data, in case the CPU needs again soon. cache memory is many, many times faster than RAM, just RAM is much faster than the HD.


What is main memory in computer architecture?

main memory is RAM.The words Memory,Buffer,Cache are all refers Ram. Which is Nearly 11,000 times faster than secondary memory (Hard Disk) in Random Access.


What is role of cache memory in microprocessor what are the typical cache capacities of p4 processor?

Cache memory is random access memory (RAM) that a computer microprocessor can access more quickly than it can access regular RAM. As the microprocessor processes data, it looks first in the cache memory and if it finds the data there (from a previous reading of data), it does not have to do the more time-consuming reading of data from larger memory.


What are Limitations of cache?

size. cache memory is situated on the CPU, and stores recently used data, in case the CPU needs again soon. cache memory is many, many times faster than RAM, just RAM is much faster than the HD.


What type of memory stores frequently used instructions or data needed by the CPU?

the memory that directly communicates with the CPU is called Cache Memory this is; When the processor(CPU) needs to read from or write to a location in main memory, it first checks whether a copy of that data is in the cache. If so, the processor immediately reads from or writes to the cache, which is much faster than reading from or writing to main memory. A CPU cache is a cache used by the central processing unit of a computer to reduce the average time to access memory. The cache is a smaller, faster memory which stores copies of the data from the most frequently used main memory locations. As long as most memory accesses are cached memory locations, the average latency of memory accesses will be closer to the cache latency than to the latency of main memory. [by Deogratius Ambrose IAA-Tanzania] the memory that directly communicates with the CPU is called Cache Memorythis is; When the processor(CPU) needs to read from or write to a location in main memory, it first checks whether a copy of that data is in the cache. If so, the processor immediately reads from or writes to the cache, which is much faster than reading from or writing to main memory. A CPU cache is a cache used by the central processing unit of a computer to reduce the average time to access memory. The cache is a smaller, faster memory which stores copies of the data from the most frequently used main memory locations. As long as most memory accesses are cached memory locations, the average latency of memory accesses will be closer to the cache latency than to the latency of main memory. [by Deogratius Ambrose IAA-Tanzania]