Basically for the same reason that cache itself helps increase the speed. If there is more cache memory, the computer will have more information readily available, and there will be less cases where the information CAN'T be found in cache, and the CPU needs to access actual RAM, which is slower (or whatever other source of data the cache is supposed to speed up).
The computer uses the cache memory as extra high-speed space to supplement the internal RAM. Windows automatically creates a pre-set cache, but you can increase the size yourself.
cache memory
A
Getting data from memory, or the hard drive is slow. If you store a part of the memory you think you will need soon, or often in cache, it will speed up processing by reducing wait time. Cache is much smaller, but much faster than memory and sits on the processor die.
Cache memory is a small, high-speed memory located on the CPU chip that temporarily stores frequently accessed data and instructions to speed up processing. RAM (Random Access Memory) is larger, slower memory that provides temporary storage for data and programs currently in use by the computer. Cache memory is faster but smaller in size compared to RAM.
Yes ,i think cache memory is very fast memory in conputer system.
Cache memory is used to hold a copy of the data from the most frequently accessed memory locations to minimise calculation time. For example, a CPU's cache will hold this data so that it doesn't have to call data from the RAM for every calculation it makes, and merely draws from its local cache.
It takes a long time, in computer time, for the CPU to retrieve stuff out of RAM while running a program. So to speed things up cache was built. Cache holds a small part of the program you are working with and it is easy and fast for the CPU to retrieve it from cache. Cache is memory and it holds data for the CPU.
When a computer gets ready to execute the next instruction, it pulls it out of memory of some sort or another. first it tries it's local high speed cache RAM, usually a part of the CPU chip. If it's not there, then it looks in the slower speed RAM. If it finds it there, the memory controller pulls a block of memory from RAM to cache then executes it there. If it doesn't see it in RAM, it looks for it in virtual memory, which is actually a part of the hard disk drive. When it finds it there, it pulls a block into RAM, then into Cache memory, where it is executed. Actually, the move from virtual memory to RAM is done way ahead of time, as the controllers see that the computer might need that block of memory in the near future. So you can see, all the instructions are executed in the small high speed cache RAM. This is done for speed. If all the instructions were executed in RAM, as computers once did, they would be 10 times slower. A lot of computer design is optimizing the memory controllers so that almost all of the instructions are executed out of high speed cache, and the processor rarely has to wait for the cache to fill up. If the computer executed out of hard disk space, it would be thousands of time slower.
Answer:- The Cache memory is placed between the CPU and the main memory. It is a fast speed memory and is expensive and faster than the main memory. Cache memory is used to store the frequently accessed data of main memory. The instructions that are frequently used by the CPU are stored in the cache memory. It is used to reduce the average access time for address, instructions or data, which are normally stored in the main memory. Cache memory increases the operating speed of the system. But is much costlier than main memory. From economic considerations, the capacity of the cache memory is much less as compared to main memory..
Register. The memory hierarchy in the decreasing order of speeds: 1. Register 2. Cache 3. Main memory 4. Disks 5. Auxiliary memory
Cache is a high speed memory which is basically used for the following reason: As the speed of the main memory is not as much as the speed of the CPU.so just to compensate the speed mistmatch between the CPU and main memory the cache is used in between the two.so whenever the CPU asks for any data its being checked with the cache memory and if present then "cache hit" occurs or else "cache miss" occurs wher the CPU takes the data form the main memory and that data's cpoy is being send to the cache for any further operation where the CPU can request for the same data. Anand bhat(mca@kiit-870024)
Cache.