Basically for the same reason that cache itself helps increase the speed. If there is more cache memory, the computer will have more information readily available, and there will be less cases where the information CAN'T be found in cache, and the CPU needs to access actual RAM, which is slower (or whatever other source of data the cache is supposed to speed up).
cache memory
A
Getting data from memory, or the hard drive is slow. If you store a part of the memory you think you will need soon, or often in cache, it will speed up processing by reducing wait time. Cache is much smaller, but much faster than memory and sits on the processor die.
RAM (Random Access Memory) is used to store information for the programs running now while the computer is powered. Cache memory is RAM that a computer microprocessor can access more quickly than it can access regular RAM.
Yes ,i think cache memory is very fast memory in conputer system.
Cache memory is used to hold a copy of the data from the most frequently accessed memory locations to minimise calculation time. For example, a CPU's cache will hold this data so that it doesn't have to call data from the RAM for every calculation it makes, and merely draws from its local cache.
It takes a long time, in computer time, for the CPU to retrieve stuff out of RAM while running a program. So to speed things up cache was built. Cache holds a small part of the program you are working with and it is easy and fast for the CPU to retrieve it from cache. Cache is memory and it holds data for the CPU.
A cache miss penalty occurs when the CPU needs data that is not in the cache memory, causing a delay as it fetches the data from the slower main memory. This delay can significantly impact the performance of a computer system by slowing down processing speed and increasing latency in executing tasks.
When a computer gets ready to execute the next instruction, it pulls it out of memory of some sort or another. first it tries it's local high speed cache RAM, usually a part of the CPU chip. If it's not there, then it looks in the slower speed RAM. If it finds it there, the memory controller pulls a block of memory from RAM to cache then executes it there. If it doesn't see it in RAM, it looks for it in virtual memory, which is actually a part of the hard disk drive. When it finds it there, it pulls a block into RAM, then into Cache memory, where it is executed. Actually, the move from virtual memory to RAM is done way ahead of time, as the controllers see that the computer might need that block of memory in the near future. So you can see, all the instructions are executed in the small high speed cache RAM. This is done for speed. If all the instructions were executed in RAM, as computers once did, they would be 10 times slower. A lot of computer design is optimizing the memory controllers so that almost all of the instructions are executed out of high speed cache, and the processor rarely has to wait for the cache to fill up. If the computer executed out of hard disk space, it would be thousands of time slower.
Answer:- The Cache memory is placed between the CPU and the main memory. It is a fast speed memory and is expensive and faster than the main memory. Cache memory is used to store the frequently accessed data of main memory. The instructions that are frequently used by the CPU are stored in the cache memory. It is used to reduce the average access time for address, instructions or data, which are normally stored in the main memory. Cache memory increases the operating speed of the system. But is much costlier than main memory. From economic considerations, the capacity of the cache memory is much less as compared to main memory..
Register. The memory hierarchy in the decreasing order of speeds: 1. Register 2. Cache 3. Main memory 4. Disks 5. Auxiliary memory
Cache.