because cache memory is costlier than main memory and physical size of cache memory also matters.ignoring the cost , if we use large cache memory, it will take larger physical space.so mother board won't be able to accomodate it
Register memory are smaller in size than cache memory and registers are faster than cache..Cache memory store the frequently used data from main memory..
NO...The cache is a smaller, faster memory which stores copies of the data from the most frequently used main memory locations. As long as most memory accesses are cached memory locations, the average latency of memory accesses will be closer to the cache latency than to the latency of main memory.Thus Cache memory is not main memory of computer system.--- form Wiki
Answer:- The Cache memory is placed between the CPU and the main memory. It is a fast speed memory and is expensive and faster than the main memory. Cache memory is used to store the frequently accessed data of main memory. The instructions that are frequently used by the CPU are stored in the cache memory. It is used to reduce the average access time for address, instructions or data, which are normally stored in the main memory. Cache memory increases the operating speed of the system. But is much costlier than main memory. From economic considerations, the capacity of the cache memory is much less as compared to main memory..
the memory that directly communicates with the CPU is called Cache Memory this is; When the processor(CPU) needs to read from or write to a location in main memory, it first checks whether a copy of that data is in the cache. If so, the processor immediately reads from or writes to the cache, which is much faster than reading from or writing to main memory. A CPU cache is a cache used by the central processing unit of a computer to reduce the average time to access memory. The cache is a smaller, faster memory which stores copies of the data from the most frequently used main memory locations. As long as most memory accesses are cached memory locations, the average latency of memory accesses will be closer to the cache latency than to the latency of main memory. [by Deogratius Ambrose IAA-Tanzania] the memory that directly communicates with the CPU is called Cache Memorythis is; When the processor(CPU) needs to read from or write to a location in main memory, it first checks whether a copy of that data is in the cache. If so, the processor immediately reads from or writes to the cache, which is much faster than reading from or writing to main memory. A CPU cache is a cache used by the central processing unit of a computer to reduce the average time to access memory. The cache is a smaller, faster memory which stores copies of the data from the most frequently used main memory locations. As long as most memory accesses are cached memory locations, the average latency of memory accesses will be closer to the cache latency than to the latency of main memory. [by Deogratius Ambrose IAA-Tanzania]
Level 2 (L2) cache is built onto the processor. On Older slot-mounted processors, L2 cache was external to the processor die, and ran at slower speeds than the processor. on socketed processors, L2 cache is built onto the processor. If the processor does not find the desired memory locations in L1 cache, it checks L2 cache next. However Processors with larger L2 caches perform most tasks much more quickly than processors that have smaller L2 caches for two reasons. Cache memory is faster than main memory and the processor checks cache memory for needed information before checking main memory.
Level 2 (L2) cache is built onto the processor. On Older slot-mounted processors, L2 cache was external to the processor die, and ran at slower speeds than the processor. on socketed processors, L2 cache is built onto the processor. If the processor does not find the desired memory locations in L1 cache, it checks L2 cache next. However Processors with larger L2 caches perform most tasks much more quickly than processors that have smaller L2 caches for two reasons. Cache memory is faster than main memory and the processor checks cache memory for needed information before checking main memory.
Level 2 (L2) cache is built onto the processor. On Older slot-mounted processors, L2 cache was external to the processor die, and ran at slower speeds than the processor. on socketed processors, L2 cache is built onto the processor. If the processor does not find the desired memory locations in L1 cache, it checks L2 cache next. However Processors with larger L2 caches perform most tasks much more quickly than processors that have smaller L2 caches for two reasons. Cache memory is faster than main memory and the processor checks cache memory for needed information before checking main memory.
I think...the above Question you asked in which case ..are you focessing over Graphics or Perfomace based..., Actually Cache is the Excellent Feautre in Processor....,HIgh Availablity of Cache Improves the Perfomance... The cache is a smaller, faster memory which stores copies of the data from the most frequently used main memory locations. As long as most memory accesses are cached memory locations, the average latency of memory accesses will be closer to the cache latency than to the latency of main memory.
Cache memory is random access memory (RAM) that a computer microprocessor can access more quickly than it can access regular RAM. As the microprocessor processes data, it looks first in the cache memory and if it finds the data there (from a previous reading of data), it does not have to do the more time-consuming reading of data from larger memory. Cache Memory generally comes in smaller size (3MB, 6MB etc) than the RAM (512 MB, 1GB,2GB.....)
as the cache memory is mory faster than RAM or hard disk ,but the main reason is that the cache memory has comparator and storage medium at the same time, the comparator checks if the address of the value being accessed is in the associative memory(a part of cache)or not, and the ordinary memory(another part of cache)has the data....
Getting data from memory, or the hard drive is slow. If you store a part of the memory you think you will need soon, or often in cache, it will speed up processing by reducing wait time. Cache is much smaller, but much faster than memory and sits on the processor die.
A cache miss is where the processor requests a memory transfer, and that data is not in cache. This requires the bus interface unit to perform a slow access to memory, as opposed to a fast access to cache, or it requires the cache manager to make disk accesses, which can be millions of times slower than main memory. Depending on the cache level, a consistently high percentage of cache misses can impact performance significantly. This is most often seen in low physical memory machines, where the swap file hit-miss ratio is poor. The working set is the memory that is most recently used. Ideally, you want short-term working set to always be smaller than physical memory. Since working set is hard to measure, you can use commit charge, though that is not as accurate. You want commit charge for currently active applications plus kernel memory to be less than physical memory.