Misplaced Pages

Front-side bus

Article snapshot taken from Wikipedia with creative commons attribution-sharealike license. Give it a read and then ask your questions in the chat. We can research this topic together.

The front-side bus ( FSB ) is a computer communication interface ( bus ) that was often used in Intel -chip-based computers during the 1990s and 2000s. The EV6 bus served the same function for competing AMD CPUs. Both typically carry data between the central processing unit (CPU) and a memory controller hub, known as the northbridge .

#966033

68-650: Depending on the implementation, some computers may also have a back-side bus that connects the CPU to the cache . This bus and the cache connected to it are faster than accessing the system memory (or RAM) via the front-side bus. The speed of the front side bus is often used as an important measure of the performance of a computer. The original front-side bus architecture was replaced by HyperTransport , Intel QuickPath Interconnect , and Direct Media Interface , followed by Intel Ultra Path Interconnect and AMD's Infinity Fabric . The term came into use by Intel Corporation about

136-526: A back-side bus ( BSB ), or backside bus , was a computer bus used on early Intel platforms to connect the CPU to CPU cache memory , usually off-die L2. If a design utilizes a back-side bus along with a front-side bus (FSB), the design is said to use a dual-bus architecture , or in Intel 's terminology Dual Independent Bus (DIB) architecture. The back-side bus architecture evolved when newer processors like

204-456: A bandwidth of 3200 megabytes per second (MB/s): The number of transfers per clock cycle depends on the technology used. For example, GTL+ performs 1 transfer/cycle, EV6 2 transfers/cycle, and AGTL+ 4 transfers/cycle. Intel calls the technique of four transfers per cycle Quad Pumping . Many manufacturers publish the frequency of the front-side bus in MHz, but marketing materials often list

272-563: A few dozen or few hundred bits of such memory could be provided. The first practical form of random-access memory was the Williams tube . It stored data as electrically charged spots on the face of a cathode-ray tube . Since the electron beam of the CRT could read and write the spots on the tube in any order, memory was random access. The capacity of the Williams tube was a few hundred to around

340-470: A hard drive. This entire pool of memory may be referred to as "RAM" by many developers, even though the various subsystems can have very different access times , violating the original concept behind the random access term in RAM. Even within a hierarchy level such as DRAM, the specific row, column, bank, rank , channel, or interleave organization of the components make the access time variable, although not to

408-424: A memory capacity that is a power of two. Usually several memory cells share the same address. For example, a 4 bit "wide" RAM chip has four memory cells for each address. Often the width of the memory and that of the microprocessor are different, for a 32 bit microprocessor, eight 4 bit RAM chips would be needed. Often more addresses are needed than can be provided by a device. In that case, external multiplexors to

476-404: A portion of a computer's RAM, allowing it to act as a much faster hard drive that is called a RAM disk . A RAM disk loses the stored data when the computer is shut down, unless memory is arranged to have a standby battery source, or changes to the RAM disk are written out to a nonvolatile disk. The RAM disk is reloaded from the physical disk upon RAM disk initialization. Sometimes, the contents of

544-420: A preset multiplier setting into the chip. It is possible to unlock some locked CPUs; for instance, some AMD Athlon processors can be unlocked by connecting electrical contacts across points on the CPU's surface. Some other processors from AMD and Intel are unlocked from the factory and labeled as an "enthusiast-grade" processors by end users and retailers because of this feature. For all processors, increasing

612-498: A processor running at 3200 MHz might be using a 400 MHz FSB. This means there is an internal clock multiplier setting (also called bus/core ratio) of 8. That is, the CPU is set to run at 8 times the frequency of the front-side bus: 400 MHz × 8 = 3200 MHz. Different CPU speeds are achieved by varying either the FSB frequency or the CPU multiplier, this is referred to as overclocking or underclocking . Setting an FSB speed

680-555: A relatively slow ROM chip are copied to read/write memory to allow for shorter access times. The ROM chip is then disabled while the initialized memory locations are switched in on the same block of addresses (often write-protected). This process, sometimes called shadowing , is fairly common in both computers and embedded systems . As a common example, the BIOS in typical personal computers often has an option called "use shadow BIOS" or similar. When enabled, functions that rely on data from

748-663: A shared FSB, though performance could not scale linearly due to bandwidth bottlenecks . The front-side bus was used in all Intel Atom , Celeron , Pentium , Core 2 , and Xeon processor models through about 2008 and was eliminated in 2009. Originally, this bus was a central connecting point for all system devices and the CPU. The potential of a faster CPU is wasted if it cannot fetch instructions and data as quickly as it can execute them. The CPU may spend significant time idle while waiting to read or write data in main memory, and high-performance processors therefore require high bandwidth and low latency access to memory. The front-side bus

SECTION 10

#1732798161967

816-542: A single MOS transistor per capacitor. The first commercial DRAM IC chip, the 1K Intel 1103 , was introduced in October 1970. Synchronous dynamic random-access memory (SDRAM) was reintroduced with the Samsung KM48SL2000 chip in 1992. Early computers used relays , mechanical counters or delay lines for main memory functions. Ultrasonic delay lines were serial devices which could only reproduce data in

884-413: A small amount of work on each element of a large data set , FSB speed becomes a major performance issue. A slow FSB will cause the CPU to spend significant amounts of time waiting for data to arrive from system memory . However, if the computations involving each element are more complex, the processor will spend longer performing these; therefore, the FSB will be able to keep pace because the rate at which

952-491: A switch that lets the control circuitry on the chip read the capacitor's state of charge or change it. As this form of memory is less expensive to produce than static RAM, it is the predominant form of computer memory used in modern computers. Both static and dynamic RAM are considered volatile , as their state is lost or reset when power is removed from the system. By contrast, read-only memory (ROM) stores data by permanently enabling or disabling selected transistors, such that

1020-611: A thousand bits, but it was much smaller, faster, and more power-efficient than using individual vacuum tube latches. Developed at the University of Manchester in England, the Williams tube provided the medium on which the first electronically stored program was implemented in the Manchester Baby computer, which first successfully ran a program on 21 June, 1948. In fact, rather than the Williams tube memory being designed for

1088-466: A traditional architecture, the front-side bus served as the immediate data link between the CPU and all other devices in the system, including main memory. In HyperTransport- and QPI-based systems, system memory is accessed independently by means of a memory controller integrated into the CPU, leaving the bandwidth on the HyperTransport or QPI link for other uses. This increases the complexity of

1156-615: Is a type of flip-flop circuit, usually implemented using FETs . This means that SRAM requires very low power when not being accessed, but it is expensive and has low storage density. A second type, DRAM, is based around a capacitor. Charging and discharging this capacitor can store a "1" or a "0" in the cell. However, the charge in this capacitor slowly leaks away, and must be refreshed periodically. Because of this refresh process, DRAM uses more power, but it can achieve greater storage densities and lower unit costs compared to SRAM. To be useful, memory cells must be readable and writable. Within

1224-403: Is far more expensive than the dynamic RAM used for larger memories. Static RAM also consumes far more power. CPU speed improvements slowed significantly partly due to major physical barriers and partly because current CPU designs have already hit the memory wall in some sense. Intel summarized these causes in a 2005 document. First of all, as chip geometries shrink and clock frequencies rise,

1292-419: Is more expensive to produce, but is generally faster and requires less dynamic power than DRAM. In modern computers, SRAM is often used as cache memory for the CPU . DRAM stores a bit of data using a transistor and capacitor pair (typically a MOSFET and MOS capacitor , respectively), which together comprise a DRAM cell. The capacitor holds a high or low charge (1 or 0, respectively), and the transistor acts as

1360-489: Is reduced by the size of the shadowed ROMs. The ' memory wall is the growing disparity of speed between CPU and the response time of memory (known as memory latency ) outside the CPU chip. An important reason for this disparity is the limited communication bandwidth beyond chip boundaries, which is also referred to as bandwidth wall . From 1986 to 2000, CPU speed improved at an annual rate of 55% while off-chip memory response time only improved at 10%. Given these trends, it

1428-412: Is related directly to the speed grade of memory a system must use. The memory bus connects the northbridge and RAM, just as the front-side bus connects the CPU and northbridge. Often, these two buses must operate at the same frequency. Increasing the front-side bus to 450 MHz in most cases also means running the memory at 450 MHz. In newer systems, it is possible to see memory ratios of "4:5" and

SECTION 20

#1732798161967

1496-423: Is the processor-memory performance gap, which can be addressed by 3D integrated circuits that reduce the distance between the logic and memory aspects that are further apart in a 2D chip. Memory subsystem design requires a focus on the gap, which is widening over time. The main method of bridging the gap is the use of caches ; small amounts of high-speed memory that houses recent operations and instructions nearby

1564-565: Is usually on-die. In the latter case, the BSB clock frequency is typically equal to the processor's, and the back-side bus can also be made much wider (256-bit, 512-bit) than either off-chip or on-chip FSB. The dual-bus architecture was used in a number of designs, including the IBM and Freescale (formerly the semiconductor division of Motorola ) PowerPC processors (certain PowerPC 604 models,

1632-578: The Atanasoff–Berry Computer , the Williams tube and the Selectron tube . In 1966, Robert Dennard invented modern DRAM architecture for which there is a single MOS transistor per capacitor. While examining the characteristics of MOS technology, he found it was capable of building capacitors , and that storing a charge or no charge on the MOS capacitor could represent the 1 and 0 of a bit, while

1700-831: The PowerPC 7xx family, and the Freescale 7xxx line), as well as the Intel Pentium Pro , Pentium II and early Pentium III processors, which used it to access their L2 cache (earlier Intel processors accessed the L2 cache over the FSB, while later processors moved it on-chip). Random-access memory Random-access memory ( RAM ; / r æ m / ) is a form of electronic computer memory that can be read and changed in any order, typically used to store working data and machine code . A random-access memory device allows data items to be read or written in almost

1768-432: The 1960s with bipolar memory, which used bipolar transistors . Although it was faster, it could not compete with the lower price of magnetic core memory. In 1957, Frosch and Derick manufactured the first silicon dioxide field-effect transistors at Bell Labs, the first transistors in which drain and source were adjacent at the surface. Subsequently, in 1960, a team demonstrated a working MOSFET at Bell Labs. This led to

1836-505: The BIOS's ROM instead use DRAM locations (most can also toggle shadowing of video card ROM or other ROM sections). Depending on the system, this may not result in increased performance, and may cause incompatibilities. For example, some hardware may be inaccessible to the operating system if shadow RAM is used. On some systems the benefit may be hypothetical because the BIOS is not used after booting in favor of direct hardware access. Free memory

1904-544: The Baby, the Baby was a testbed to demonstrate the reliability of the memory. Magnetic-core memory was invented in 1947 and developed up until the mid-1970s. It became a widespread form of random-access memory, relying on an array of magnetized rings. By changing the sense of each ring's magnetization, data could be stored with one bit stored per ring. Since every ring had a combination of address wires to select and read or write it, access to any memory location in any sequence

1972-462: The CPU design but offers greater throughput as well as superior scaling in multiprocessor systems. The bandwidth or maximum theoretical throughput of the front-side bus is determined by the product of the width of its data path, its clock frequency (cycles per second) and the number of data transfers it performs per clock cycle. For example, a 64- bit (8- byte ) wide FSB operating at a frequency of 100 MHz that performs 4 transfers per cycle has

2040-424: The FSB speed can be done to boost processing speed by reducing latency between CPU and the northbridge. This practice pushes components beyond their specifications and may cause erratic behavior, overheating or premature failure. Even if the computer appears to run normally, problems may appear under a heavy load. Most PCs purchased from retailers or manufacturers, such as Hewlett-Packard or Dell , do not allow

2108-411: The FSB, only allowing Intel processors in the CPU socket. The first example was field-programmable gate array (FPGA) co-processors, a result of collaboration between Intel- Xilinx - Nallatech and Intel- Altera -XtremeData (which shipped in 2008). The frequency at which a processor (CPU) operates is determined by applying a clock multiplier to the front-side bus (FSB) speed in some cases. For example,

Front-side bus - Misplaced Pages Continue

2176-402: The MOS transistor could control writing the charge to the capacitor. This led to his development of a single-transistor DRAM memory cell. In 1967, Dennard filed a patent under IBM for a single-transistor DRAM memory cell, based on MOS technology. The first commercial DRAM IC chip was the Intel 1103 , which was manufactured on an 8   μm MOS process with a capacity of 1   kbit , and

2244-441: The RAM comes in an easily upgraded form of modules called memory modules or DRAM modules about the size of a few sticks of chewing gum. These can be quickly replaced should they become damaged or when changing needs demand more storage capacity. As suggested above, smaller amounts of RAM (mostly SRAM) are also integrated in the CPU and other ICs on the motherboard , as well as in hard-drives, CD-ROMs , and several other parts of

2312-444: The RAM device, multiplexing and demultiplexing circuitry is used to select memory cells. Typically, a RAM device has a set of address lines A 0 , A 1 , . . . A n {\displaystyle A_{0},A_{1},...A_{n}} , and for each combination of bits that may be applied to these lines, a set of memory cells are activated. Due to this addressing, RAM devices virtually always have

2380-587: The SP95 memory chip for the System/360 Model 95 . Dynamic random-access memory (DRAM) allowed replacement of a 4 or 6-transistor latch circuit by a single transistor for each memory bit, greatly increasing memory density at the cost of volatility. Data was stored in the tiny capacitance of each transistor, and had to be periodically refreshed every few milliseconds before the charge could leak away. Toshiba 's Toscal BC-1411 electronic calculator , which

2448-412: The back-side bus can be optimized for communication with cache, thus eliminating protocol overheads and additional signals that are required on a general-purpose bus. Furthermore, since a BSB operates over a shorter distance, it can typically operate at higher clock speeds, increasing the computer's overall performance. Cache connected with a BSB was initially external to the microprocessor die , but now

2516-552: The chipset in order for data to flow between the connected devices. These secondary system buses usually run at speeds derived from the front-side bus clock, but are not necessarily synchronized to it. In response to AMD 's Torrenza initiative, Intel opened its FSB CPU socket to third party devices. Prior to this announcement, made in Spring 2007 at Intel Developer Forum in Beijing, Intel had very closely guarded who had access to

2584-405: The computer system. In addition to serving as temporary storage and working space for the operating system and applications, RAM is used in numerous other ways. Most modern operating systems employ a method of extending RAM capacity, known as "virtual memory". A portion of the computer's hard drive is set aside for a paging file or a scratch partition , and the combination of physical RAM and

2652-448: The development of metal–oxide–semiconductor (MOS) memory by John Schmidt at Fairchild Semiconductor in 1964. In addition to higher speeds, MOS semiconductor memory was cheaper and consumed less power than magnetic core memory. The development of silicon-gate MOS integrated circuit (MOS IC) technology by Federico Faggin at Fairchild in 1968 enabled the production of MOS memory chips . MOS memory overtook magnetic core memory as

2720-408: The device are used to activate the correct device that is being accessed. RAM is often byte addressable, although it is also possible to make RAM that is word-addressable. One can read and over-write data in RAM. Many computer systems have a memory hierarchy consisting of processor registers , on- die SRAM caches, external caches , DRAM , paging systems and virtual memory or swap space on

2788-437: The dominant memory technology in the early 1970s. Integrated bipolar static random-access memory (SRAM) was invented by Robert H. Norman at Fairchild Semiconductor in 1963. It was followed by the development of MOS SRAM by John Schmidt at Fairchild in 1964. SRAM became an alternative to magnetic-core memory, but required six MOS transistors for each bit of data. Commercial use of SRAM began in 1965, when IBM introduced

Front-side bus - Misplaced Pages Continue

2856-410: The extent that access time to rotating storage media or a tape is variable. The overall goal of using a memory hierarchy is to obtain the fastest possible average access time while minimizing the total cost of the entire memory system (generally, the memory hierarchy follows the access time with the fast CPU registers at the top and the slow hard drive at the bottom). In many modern personal computers,

2924-696: The form of integrated circuit (IC) chips with MOS (metal–oxide–semiconductor) memory cells . RAM is normally associated with volatile types of memory where stored information is lost if power is removed. The two main types of volatile random-access semiconductor memory are static random-access memory (SRAM) and dynamic random-access memory (DRAM). Non-volatile RAM has also been developed and other types of non-volatile memories allow random access for read operations, but either do not allow write operations or have other kinds of limitations. These include most types of ROM and NOR flash memory . The use of semiconductor RAM dates back to 1965 when IBM introduced

2992-515: The front-side bus for timing. Overclocking is the practice of making computer components operate beyond their stock performance levels by manipulating the frequencies at which the component is set to run, and, when necessary, modifying the voltage sent to the component to allow it to operate at these higher frequencies with more stability. Many motherboards allow the user to manually set the clock multiplier and FSB settings by changing jumpers or BIOS settings. Almost all CPU manufacturers now "lock"

3060-412: The fundamental building block of computer memory . The memory cell is an electronic circuit that stores one bit of binary information and it must be set to store a logic 1 (high voltage level) and reset to store a logic 0 (low voltage level). Its value is maintained/stored until it is changed by the set/reset process. The value in the memory cell can be accessed by reading it. In SRAM, the memory cell

3128-435: The like. The memory will run 5/4 times as fast as the FSB in this situation, meaning a 400 MHz bus can run with the memory at 500 MHz. This is often referred to as an 'asynchronous' system. Due to differences in CPU and system architecture, overall system performance can vary in unexpected ways with different FSB-to-memory ratios. In image, audio, video, gaming, FPGA synthesis and scientific applications that perform

3196-545: The means of producing inductance within solid state devices, resistance-capacitance (RC) delays in signal transmission are growing as feature sizes shrink, imposing an additional bottleneck that frequency increases don't address. The RC delays in signal transmission were also noted in "Clock Rate versus IPC: The End of the Road for Conventional Microarchitectures" which projected a maximum of 12.5% average annual CPU performance improvement between 2000 and 2014. A different concept

3264-436: The memory cannot be altered. Writable variants of ROM (such as EEPROM and NOR flash ) share properties of both ROM and RAM, enabling data to persist without power and to be updated without requiring special equipment. ECC memory (which can be either SRAM or DRAM) includes special circuitry to detect and/or correct random faults (memory errors) in the stored data, using parity bits or error correction codes . In general,

3332-466: The memory is accessed is reduced. Similar to the memory bus, the PCI and AGP buses can also be run asynchronously from the front-side bus. In older systems, these buses are operated at a set fraction of the front-side bus frequency. This fraction was set by the BIOS . In newer systems, the PCI, AGP, and PCI Express peripheral buses often receive their own clock signals , which eliminates their dependence on

3400-467: The monolithic (single-chip) 16-bit SP95 SRAM chip for their System/360 Model 95 computer, and Toshiba used bipolar DRAM memory cells for its 180-bit Toscal BC-1411 electronic calculator , both based on bipolar transistors . While it offered higher speeds than magnetic-core memory , bipolar DRAM could not compete with the lower price of the then-dominant magnetic-core memory. In 1966, Dr. Robert Dennard invented modern DRAM architecture in which there's

3468-443: The order it was written. Drum memory could be expanded at relatively low cost but efficient retrieval of memory items requires knowledge of the physical layout of the drum to optimize speed. Latches built out of triode vacuum tubes , and later, out of discrete transistors , were used for smaller and faster memories such as registers . Such registers were relatively large and too costly to use for large amounts of data; generally only

SECTION 50

#1732798161967

3536-571: The paging file form the system's total memory. (For example, if a computer has 2 GB (1024 B) of RAM and a 1 GB page file, the operating system has 3 GB total memory available to it.) When the system runs low on physical memory, it can " swap " portions of RAM to the paging file to make room for new data, as well as to read previously swapped information back into RAM. Excessive use of this mechanism results in thrashing and generally hampers overall system performance, mainly because hard drives are far slower than RAM. Software can "partition"

3604-577: The processor, speeding up the execution of those operations or instructions in cases where they are called upon frequently. Multiple levels of caching have been developed to deal with the widening gap, and the performance of high-speed modern computers relies on evolving caching techniques. There can be up to a 53% difference between the growth in speed of processor and the lagging speed of main memory access. Solid-state hard drives have continued to increase in speed, from ~400 Mbit/s via SATA3 in 2012 up to ~7 GB/s via NVMe / PCIe in 2024, closing

3672-455: The same amount of time irrespective of the physical location of data inside the memory, in contrast with other direct-access data storage media (such as hard disks and magnetic tape ), where the time required to read and write data items varies significantly depending on their physical locations on the recording medium, due to mechanical limitations such as media rotation speeds and arm movement. In today's technology, random-access memory takes

3740-435: The same type, simply because it takes longer for signals to traverse a larger circuit. Constructing a memory unit of many gibibytes with a response time of one clock cycle is difficult or impossible. Today's CPUs often still have a mebibyte of 0 wait state cache memory, but it resides on the same chip as the CPU cores due to the bandwidth limitations of chip-to-chip communication. It must also be constructed from static RAM, which

3808-519: The second-generation Pentium III began to incorporate on-die L2 cache, which at the time was advertised as Advanced Transfer Cache , but Intel continued to refer to the Dual Independent Bus till the end of Pentium III. [REDACTED] BSB is an improvement over the older practice of using a single system bus , because a single bus typically became a severe bottleneck as CPUs and memory speeds increased. Due to its dedicated nature,

3876-400: The term RAM refers solely to solid-state memory devices (either DRAM or SRAM), and more specifically the main memory in most computers. In optical storage, the term DVD-RAM is somewhat of a misnomer since, it is not random access; it behaves much like a hard disc drive if somewhat slower. Aside, unlike CD-RW or DVD-RW , DVD-RAM does not need to be erased before reuse. The memory cell is

3944-419: The theoretical effective signaling rate (which is commonly called megatransfers per second or MT/s). For example, if a motherboard (or processor) has its bus set at 200 MHz and performs 4 transfers per clock cycle, the FSB is rated at 800 MT/s. The specifications of several generations of popular processors are indicated below. Back-side bus In personal computer microprocessor architecture,

4012-547: The time the Pentium Pro and Pentium II products were announced, in the 1990s. "Front side" refers to the external interface from the processor to the rest of the computer system, as opposed to the back side, where the back-side bus connects the cache (and potentially other CPUs). A front-side bus (FSB) is mostly used on PC-related motherboards (including personal computers and servers). They are seldom used in embedded systems or similar small computers. The FSB design

4080-558: The transistor leakage current increases, leading to excess power consumption and heat... Secondly, the advantages of higher clock speeds are in part negated by memory latency, since memory access times have not been able to keep pace with increasing clock frequencies. Third, for certain applications, traditional serial architectures are becoming less efficient as processors get faster (due to the so-called von Neumann bottleneck ), further undercutting any gains that frequency increases might otherwise buy. In addition, partly due to limitations in

4148-466: The user to change the multiplier or FSB settings due to the probability of erratic behavior or failure. Motherboards purchased separately to build a custom machine are more likely to allow the user to edit the multiplier and FSB settings in the PC's BIOS. The front-side bus had the advantage of high flexibility and low cost when it was first designed. Simple symmetric multiprocessors place a number of CPUs on

SECTION 60

#1732798161967

4216-498: Was Samsung's 64   Mbit DDR SDRAM chip, released in June 1998. GDDR (graphics DDR) is a form of DDR SGRAM (synchronous graphics RAM), which was first released by Samsung as a 16   Mbit memory chip in 1998. The two widely used forms of modern RAM are static RAM (SRAM) and dynamic RAM (DRAM). In SRAM, a bit of data is stored using the state of a six- transistor memory cell , typically using six MOSFETs. This form of RAM

4284-537: Was a performance improvement over the single system bus designs of the previous decades, but these front-side buses are sometimes referred to as the "system bus". Front-side buses usually connect the CPU and the rest of the hardware via a chipset , which Intel implemented as a northbridge and a southbridge . Other buses like the Peripheral Component Interconnect (PCI), Accelerated Graphics Port (AGP), and memory buses all connect to

4352-532: Was criticized by AMD as being an old and slow technology that limits system performance. More modern designs use point-to-point and serial connections like AMD's HyperTransport and Intel's DMI 2.0 or QuickPath Interconnect (QPI). These implementations remove the traditional northbridge in favor of a direct link from the CPU to the system memory, high-speed peripherals, and the Platform Controller Hub , southbridge or I/O controller. In

4420-445: Was expected that memory latency would become an overwhelming bottleneck in computer performance. Another reason for the disparity is the enormous increase in the size of memory since the start of the PC revolution in the 1980s. Originally, PCs contained less than 1 mebibyte of RAM, which often had a response time of 1 CPU clock cycle, meaning that it required 0 wait states. Larger memory units are inherently slower than smaller ones of

4488-410: Was introduced in 1965, used a form of capacitor-bipolar DRAM, storing 180-bit data on discrete memory cells , consisting of germanium bipolar transistors and capacitors. While it offered higher speeds than magnetic-core memory, bipolar DRAM could not compete with the lower price of the then dominant magnetic-core memory. Capacitors had also been used for earlier memory schemes, such as the drum of

4556-460: Was possible. Magnetic core memory was the standard form of computer memory until displaced by semiconductor memory in integrated circuits (ICs) during the early 1970s. Prior to the development of integrated read-only memory (ROM) circuits, permanent (or read-only ) random-access memory was often constructed using diode matrices driven by address decoders , or specially wound core rope memory planes. Semiconductor memory appeared in

4624-433: Was released in 1970. The earliest DRAMs were often synchronized with the CPU clock (clocked) and were used with early microprocessors. In the mid-1970s, DRAMs moved to the asynchronous design, but in the 1990s returned to synchronous operation. In 1992 Samsung released KM48SL2000, which had a capacity of 16   Mbit . and mass-produced in 1993. The first commercial DDR SDRAM ( double data rate SDRAM) memory chip

#966033