System bus

From Wikipedia, the free encyclopedia
Example of a single system computer bus

A system bus is a single computer bus that connects the major components of a computer system, combining the functions of a data bus to carry information, an address bus to determine where it should be sent or read from, and a control bus to determine its operation. The technique was developed to reduce costs and improve modularity, and although popular in the 1970s and 1980s, more modern computers use a variety of separate buses adapted to more specific needs.

The system level bus (as distinct from a CPU's internal datapath busses) connects the CPU to memory and I/O devices.[1] Typically a system level bus is designed for use as a backplane.[2]

Background scenario[edit]

Many of the computers were based on the First Draft of a Report on the EDVAC report published in 1945. In what became known as the Von Neumann architecture, a central control unit and arithmetic logic unit (ALU, which he called the central arithmetic part) were combined with computer memory and input and output functions to form a stored program computer.[3] The Report presented a general organization and theoretical model of the computer, however, not the implementation of that model.[4] Soon designs integrated the control unit and ALU into what became known as the central processing unit (CPU).

Computers in the 1950s and 1960s were generally constructed in an ad-hoc fashion. For example, the CPU, memory, and input/output units were each one or more cabinets connected by cables. Engineers used the common techniques of standardized bundles of wires and extended the concept as backplanes were used to hold printed circuit boards in these early machines. The name "bus" was already used for "bus bars" that carried electrical power to the various parts of electric machines, including early mechanical calculators.[5] The advent of integrated circuits vastly reduced the size of each computer unit, and buses became more standardized.[6] Standard modules could be interconnected in more uniform ways and were easier to develop and maintain.

Description[edit]

To provide even more modularity with reduced cost, memory and I/O buses (and the required control and power buses) were sometimes combined into a single unified system bus.[7] Modularity and cost became important as computers became small enough to fit in a single cabinet (and customers expected similar price reductions). Digital Equipment Corporation (DEC) further reduced cost for mass-produced minicomputers, and memory-mapped I/O into the memory bus, so that the devices appeared to be memory locations. This was implemented in the Unibus of the PDP-11 around 1969, eliminating the need for a separate I/O bus.[8] Even computers such as the PDP-8 without memory-mapped I/O were soon implemented with a system bus, which allowed modules to be plugged into any slot.[9] Some authors called this a new streamlined "model" of computer architecture.[10]

Many early microcomputers (with a CPU generally on a single integrated circuit) were built with a single system bus, starting with the S-100 bus in the Altair 8800 computer system in about 1975.[11] The IBM PC used the Industry Standard Architecture (ISA) bus as its system bus in 1981. The passive backplanes of early models were replaced with the standard of putting the CPU and RAM on a motherboard, with only optional daughterboards or expansion cards in system bus slots.

Simple symmetric multiprocessing using a system bus

The Multibus became a standard of the Institute of Electrical and Electronics Engineers as IEEE standard 796 in 1983.[12] Sun Microsystems developed the SBus in 1989 to support smaller expansion cards.[13] The easiest way to implement symmetric multiprocessing was to plug in more than one CPU into the shared system bus, which was used through the 1980s. However, the shared bus quickly became the bottleneck and more sophisticated connection techniques were explored.[14]

Even in very simple systems, at various times the data bus is driven by the program memory, by RAM, and by I/O devices. To prevent bus contention on the data bus, at any one instant only one device drives the data bus. In very simple systems, only the data bus is required to be a bidirectional bus. In very simple systems, the memory address register always drives the address bus, the control unit always drives the control bus, and an address decoder selects which particular device is allowed to drive the data bus during this bus cycle. In very simple systems, every instruction cycle starts with a READ memory cycle where program memory drives the instruction onto the data bus while the instruction register latches that instruction from the data bus. Some instructions continue with a WRITE memory cycle where the memory data register drives data onto the data bus into the chosen RAM or I/O device. Other instructions continue with another READ memory cycle where the chosen RAM, program memory, or I/O device drives data onto the data bus while the memory data register latches that data from the data bus.

More complex systems have a multi-master bus—not only do they have many devices that each drive the data bus, but also have many bus masters that each drive the address bus. The address bus as well as the data bus in bus snooping systems is required to be a bidirectional bus, often implemented as a three-state bus. To prevent bus contention on the address bus, a bus arbiter selects which particular bus master is allowed to drive the address bus during this bus cycle.

Dual Independent Bus[edit]

Intel has used the term Dual Independent Bus (DIB) for two different purposes. The first one came when Intel changed from a single local bus to the DIB, using the external front-side bus to the main system memory and I/O devices, and the internal back-side bus to the L2 CPU cache. This was introduced in the Pentium Pro in 1995.[15][16][17]

In 2005 and 2006 Intel introduced the 8500 and 5000 chipsets, where DIB referred to the two front-side buses on a chipset, which doubles the system bandwidth compared to having just one FSB shared by all the CPUs. However, the information needed to guarantee the cache coherence of shared data located in different caches have to be sent in broadcast (snooped) to check the other FSB's CPUs' cache state, reducing the available bandwidth. To reduce the coherency traffic, a snoop filter was included in the higher-end chipsets, in order to have cache state information available on-chipset. In 2007 Intel extended the idea of multiple buses in the 7300 chipset with four independent FSBs, calling it dedicated high-speed interconnects (DHSI).[18]

The system bus approach is obsolete in the modern personal and server computers, which instead use higher-performance interconnection technologies such as HyperTransport and Intel QuickPath Interconnect, while the system bus architecture continued to be used on simpler embedded microprocessors. The systems bus can even be internal to a single integrated circuit, producing a system-on-a-chip. Examples include AMBA, CoreConnect, and Wishbone.[19]

See also[edit]

References[edit]

  1. ^ Edward Bosworth. "Chapter 10 – Overview of Busses".
  2. ^ Hui Wu. "Computer Buses and Parallel Input/Output". 2006.
  3. ^ John von Neumann (June 30, 1945). "First Draft of a Report on the EDVAC" (PDF). Archived from the original (PDF) on March 14, 2013. Retrieved May 27, 2011. Introduction and editing by Michael D. Godfrey, Stanford University, November 1992.
  4. ^ Michael D. Godfrey; D. F. Hendry (1993). "The Computer as von Neumann Planned It" (PDF). IEEE Annals of the History of Computing. 15 (1): 11–21. doi:10.1109/85.194088. S2CID 569933. Archived from the original (PDF) on 2011-08-25.
  5. ^ U.S. Patent 3,470,421 "Continuous Bus Bar for Connector Plate Back Panel Machine Wiring" Donald L. Shore et al., Filed August 30, 1967, issued September 30, 1969.
  6. ^ U.S. Patent 3,462,742 "Computer System Adapted to be Constructed of Large Integrated Circuit Arrays" Henry S. Miller et al., Filed December 21, 1966, issued August 19, 1969.
  7. ^ Linda Null; Julia Lobur (2010). The essentials of computer organization and architecture (3rd ed.). Jones & Bartlett Learning. pp. 36, 199–203. ISBN 978-1-4496-0006-8.
  8. ^ C. Gordon Bell; R. Cady; H. McFarland; B. Delagi; J. O'Laughlin; R. Noonan; W. Wulf (1970). "A New Architecture for Mini-Computers—The DEC PDP-11" (PDF). Spring Joint Computer Conference: 657–675.
  9. ^ Small Computer Handbook (PDF). Digital Equipment Corporation. 1973. pp. 2–9.
  10. ^ Miles J. Murdocca; Vincent P. Heuring (2007). Computer architecture and organization: an integrated approach. John Wiley & Sons. p. 11. ISBN 978-0-471-73388-1.
  11. ^ Herbert R. Johnson. "Origins of S-100 computers".
  12. ^ "796-1983 — IEEE Standard Microcomputer System Bus". Institute of Electrical and Electronics Engineers. 1983. Retrieved May 25, 2011.
  13. ^ Frank, E.H. (1990). "The SBus: Sun's high performance system bus for RISC workstations". Digest of Papers Compcon Spring '90. Thirty-Fifth IEEE Computer Society International Conference on Intellectual Leverage. pp. 189–194. doi:10.1109/CMPCON.1990.63672. ISBN 0-8186-2028-5. S2CID 25815415.
  14. ^ Donald Charles Winsor (1989). Bus and Cache Memory Organization for Multiprocessors (PDF). University of Michigan Electrical Engineering department. Ph.D. dissertation.
  15. ^ Intel's CEO Reveals New Bus Architecture To Be Implemented In Upcoming Pentium® II Microprocessor
  16. ^ Todd Langley and Rob Kowalczyk (January 2009). "Introduction to Intel Architecture: The Basics" (PDF). White paper. Intel Corporation. Archived from the original (PDF) on June 7, 2011. Retrieved May 25, 2011.
  17. ^ "Accelerated Graphics Port". Next Generation. No. 37. Imagine Media. January 1998. pp. 94–96.
  18. ^ An Introduction to the Intel® QuickPath Interconnect, Figures 4 and 5.
  19. ^ Rudolf Usselmann (January 9, 2001). "OpenCores SoC Bus Review" (PDF). Retrieved May 30, 2011.