answersLogoWhite

0


Best Answer

cache is more expensive and it will increase the cost of the system terribly.

processing of more than one cache will complicate the design of CPU and increase the burden on CPU.

User Avatar

Wiki User

11y ago
This answer is:
User Avatar

Add your answer:

Earn +20 pts
Q: Why do systems not use more or large cache memory if they are so useful?
Write your answer...
Submit
Still have questions?
magnify glass
imp
Continue Learning about Engineering

Which is preferred for writing a Merge Sort routine contiguous or linked memory?

Linked memory because its very useful primarily when the lists to be sorted are very large and the size of data to be moved is small.


Which systems support decision making by enabling users to extract useful information that was previously buried in large quantities of data?

DSS


What database types is useful for storing java applets as well as processing large numbers of transactions?

OODBMS, object oriented database management systems


If a cache can be made as large as the device for which it is caching for instance a cache as large as a disk why not make it that large and eliminate the device?

Though the size of a cache has increased over time, so too has the size of hard disk. An economical comparison of cache versus hard disk space in a cost per MB analsysis will show that a cache would be significantly more expensive. Furthermore, cache in general is considered "temporary" or volatile storage which means that the contents of the storage device is lost when the system is powered off. A hard disk, on the other hand, is "long term" or non-volatile storage; when the system is powered off, the hard disk still safely holds the data stored on it.


What is memory allocation?

Memory allocation is the act of reserving a chunk of memory for some set of data. In programming terms, this is normally done by declaring a variable. Large arrays of data will require large blocks of contiguous memory, which the programmer must request from the operating system.

Related questions

Why is cache memory smaller than main memory?

because cache memory is costlier than main memory and physical size of cache memory also matters.ignoring the cost , if we use large cache memory, it will take larger physical space.so mother board won't be able to accomodate it


What if Cache size is equal to main memory size?

If we took size of cache as large as size of main memory then the main purpose of cache(take less time) would come to end, as larger the cache slower would be it's processing speed.


Which of the memory has shortest access time?

The processor registers! Relatively, they are very small (a register on an Intel 32-bit processor is only 4 bytes large!) but they are very fast. Programs use them to store the part of data they are working on and some memory addresses.


What is meant by cache memory?

In a computer, cache memory is a special type of fast access memory that is between the CPU and the main memory. If the CPU always had to access main memory, it would spend most of its time idle waiting for that memory to respond. But because memory accesses statistically tend to cluster around each other in real programs instead of completely randomly scattering across memory, a single CPU memory access can cause the cache memory controller to perform a fast burst access of main memory including that address to load an entire "line" of cache memory. If a following CPU memory access is in this same "line" of cache memory that has already been loaded, it will not have to wait for the main memory to respond, instead the cache responds first providing the copy it has of that address' contents.Cache memory was originally invented in the late 1950s by IBM for their 7030 Stretch supercomputer (a machine built entirely using discrete germanium transistors, no integrated circuits at all). However all the 7030 documents use the term "virtual memory" for what is now universally called cache memory, and "virtual memory" means something entirely different now.Each cache memory unit is composed of three sections:cache controllercontent addressable memory (CAM)fast access static random access memory (SRAM)The cache controller section contains all the logic circuits that coordinate the operation between the CPU(s), cache, and main memory. The CAM is a special type of memory used to store the memory address of each active cache "line" along with some status bits that is addressed by it contents (not an address number as in ordinary RAM and ROM). The SRAM contains the copies of address contents previously read from main memory, and is addressed through the CAM. Cache memory is organized into levels (L1, L2, L3, etc.) with the L1 Cache closest to the CPU and each additional level further away until the final Cache level connects directly to the memory. The L1 Cache uses the fastest speed SRAM but has the smallest amount of SRAM and each additional level uses slower speed SRAM but has more SRAM than the previous level. The L1 Cache is usually divided into two independent Caches (L1 Instruction Cache and L1 Data Cache) and is dedicated to supporting only one CPU but each additional level is usually a single Cache (shared by both instructions and data) and in systems having multiple CPUs may be shared by two or more CPUs. In systems having multiple CPUs the cache controllers must implement special bus protocols to coordinate line invalidations and updates to prevent some of the CPUs from accessing obsolete Cache contents from its Caches when a different CPU that does not use those Caches has modified memory that is mapped to both its local Caches and the other CPUs local caches.Proper design of a Cache memory system for a computer requires extensive simulation of typical real code expected to be used on the system. Any problems found must be corrected, so that the Cache usage will remain even and balanced, and anticipated performance verified.


What is an L1 cache?

Cache is a special type of memory which is quicker to access. stuff you have recently used will go to your cache so when you perform a function you have just performed it will happen quicker. 4MB is a large cache. Level two Cache is a special memory type that is faster than Lv1 Cache, it is like RAM but it is used by the CPU as a fuction Storage to make that process faster.


Difference of cache memory and normal memory?

A cache memory is a component that stores data so that future requests for that data can be served faster. The data that is stored within a cache might be values that have been computed earlier or duplicates of original values that are stored elsewhere. Primary memory (often referred to simply as memory) is the only one directly accessible to the CPU. The CPU continuously reads instructions stored there and executes them as required. Any data actively operated on is also stored there in uniform manner. RAM is one commonly used primary memory.


Would it be better to have on your computer a large cache or large MHZ?

cache should small for super computer.


What kind of program uses a large system cache?

What kind of computer program uses a large system cache?


What are the difference between main memory and secondary storage?

Main has the following characteristics that differentiates it from physical or secondary memory. It is volatile, expensive, small in size and fast while physical memory is permanent, large in capacity and is cheaper than main memory.


What is memory sharing?

In computer hardware, shared memory refers to a (typically) large block of random access memory that can be accessed by several different central processing units (CPUs) in a multiple-processor computer system. The issue with shared memory systems is that the many CPUs need fast access to memory and will likely cache memory. Whenever one cache is updated with information that may be used by other processors, the change needs to be reflected to the other processors, otherwise the different processors will be working with incoherent data (see cache coherence and memory coherence). Such coherence protocols can when they work well provide extremely high performance access to shared information between multiple processors. On the other hand they can sometimes become overloaded and become a bottleneck to performance. The alternatives to shared memory are distributed memory and distributed shared memory, with another, similar set of issues. See also Non-Uniform Memory Access. In software the term shared memory refers to memory that is accessible by more than one process, where a process is a running instance of a program. In this context, shared memory is used to facilitate inter-process communication.


Should I tick programs or system cache under memory usage from control panel system options?

Tick Program option if you are using the machine as a workstation, rather than as a server. Your programs will work faster and your system cache size will be the default size that came with Windows XP. Tick System Cache option if you are using the machine as a server, or if you use programs that require a large system cache.


Use the word cache in a sentence?

He kept a large cache of papers in his locker.yes i can, here it is, Can u use the french word cache in a sentence?