Scalable Coherent Interface

From Wikipedia, the free encyclopedia
  (Redirected from Scalable Coherent Interconnect)
Jump to: navigation, search
Scalable Coherent Interface and Serial Express Users, Developers, and Manufacturers Association
SCIzzL logo.gif
Group that supports the standard
Abbreviation SCIzzL
Formation 1996
Type Non-profit
Website www.scizzl.com

The Scalable Coherent Interface or Scalable Coherent Interconnect (SCI), was a high-speed interconnect standard for shared memory multiprocessing and message passing used in the 1990s. The goal was to scale well, provide system-wide memory coherence and a simple interface; i.e. a standard to replace buses in multiprocessor systems without the inherent scalability and performance limitations of buses. The IEEE Std 1596-1992, IEEE Standard for Scalable Coherent Interface (SCI) was approved by the IEEE standards board on March 19, 1992.[1]

History[edit]

Soon after the Fastbus (IEEE 960) follow-on Futurebus (IEEE 896) project in 1987, some engineers predicted it would already be too slow for the high performance computing marketplace by the time it would be released in the early 1990s. In response, a "Superbus" study group was formed in November 1987. Another working group of the standards association of the Institute of Electrical and Electronics Engineers (IEEE) spun off to form a standard targeted at this market in July 1988.[2] It was essentially a subset of Futurebus features that could be easily implemented at high speed, along with minor additions to make it easier to connect to other systems, such as VMEbus. Most of the developers had their background from high-speed computer buses. Representatives from companies in the computer industry and research community included Amdahl, Apple Computer, BB&N, Hewlett-Packard, CERN, Dolphin Server Technology, Cray Research, Sequent, AT&T, Digital Equipment Corporation, McDonnell Douglas, National Semiconductor, Stanford Linear Accelerator Center, Tektronix, Texas Instruments, Unisys, University of Oslo, University of Wisconsin.

The original intent was a single standard for all buses in the computer.[3] The working group soon came up with the idea of using point-to-point communication in the form of insertion rings. This avoided the lumped capacitance, limited physical length/speed of light problems and stub reflections in addition to allowing parallel transactions. The use of insertion rings is credited to Manolis Katevenis who suggested it at one of the early meetings of the working group. The working group for developing the standard was led by David B. Gustavson (chair) and David V. James (Vice Chair).[4]

David V. James was a major contributor for writing the specifications including the executable C-code.[citation needed] Stein Gjessing’s group at the University of Oslo used formal methods to verify the coherence protocol and Dolphin Server Technology implemented a node controller chip including the cache coherence logic.

Block diagram of one example

Different versions and derivatives of SCI were implemented by companies like Dolphin Interconnect Solutions, Convex, Data General AViiON (using cache controller and link controller chips from Dolphin), Sequent and Cray Research. Dolphin Interconnect Solutions implemented a PCI and PCI-Express connected derivative of SCI that provides non-coherent shared memory access. This implementation was used by Sun Microsystems for its high-end clusters, Thales Group and several others including volume applications for message passing within HPC clustering and medical imaging. SCI was often used to implement non-uniform memory access architectures. It was also used by Sequent Computer Systems as the processor memory bus in their NUMA-Q systems. Numascale developed a derivative to connect with coherent HyperTransport.

The standard[edit]

The standard defined two interface levels; the physical level that deals with electrical signals, connectors, mechanical and thermal conditions and the logical level that describes the address space, data transfer protocols, cache coherence mechanisms, synchronization primitives, control and status registers, and initialization and error recovery facilities. This structure allows new developments in physical interface technology to be easily adapted without any redesign on the logical level.

Scalability for large systems is achieved through a distributed directory based cache coherence model. (The other popular models for cache coherency are based on system-wide eavesdropping (snooping) of memory transactions – a scheme which is not very scalable.) In SCI each node contains a directory with a pointer to the next node in a linked list that shares a particular cache line.

SCI defines a 64-bit flat address space (16 exabytes) where 16 bits are used for identifying a node (65,536 nodes) and 48 bits for address within the node (256 terabytes). A node can contain many processors and/or memory. The SCI standard defines a packet switched network.

Topologies[edit]

SCI can be used to build systems with different types of switching topologies from centralized to fully distributed switching. With a central switch, each node is connected to the switch with a ringlet (in this case a two-node ring). In distributed switching systems, each node can be connected to a ring of arbitrary length and either all or some of the nodes can be connected to two or more rings. The most common way to describe these multi-dimensional topologies is k-ary n-cubes (or tori). The SCI standard specification mentions several such topologies as examples.

The 2-D torus is a combination of rings in two dimensions. Switching between the two dimensions requires a small switching capability in the node. This can be expanded to three or more dimensions. The concept of folding rings can also be applied to the Torus topologies to avoid any long connection segments.

Transactions[edit]

SCI sends information in packets. Each packet consists of an unbroken sequence of 16-bit symbols. The symbol is accompanied by a flag bit. A transition of the flag bit from 0 to 1 indicates the start of a packet. A transition from 1 to 0 occurs 1 (for echoes) or 4 symbols before the packet end. A packet contains a header with address command and status information, payload (from 0 through optional lengths of data) and a CRC check symbol. The first symbol in the packet header contains the destination node address. If the address is not within the domain handled by the receiving node, the packet is passed to the output through the bypass FIFO. In the other case, the packet is fed to a receive queue and may be transferred to a ring in another dimension. All packets are marked when they pass the scrubber (a node is established as scrubber when the ring is initialized). Packets without a valid destination address will be removed when passing the scrubber for the second time to avoid filling the ring with packets that would otherwise circulate indefinitely.

Cache coherence[edit]

Cache coherence ensures data consistency in multiprocessor systems. The simplest form applied in earlier systems was based on clearing the cache contents between context switches and disabling the cache for data that were shared between two or more processors. These methods were feasible when the performance difference between the cache and memory were less than one order of magnitude. Modern processors with caches that are more than two orders of magnitude faster than main memory would not perform anywhere near optimal without more sophisticated methods for data consistency. Bus based systems use eavesdropping (snooping) methods since buses are inherently broadcast. Modern systems with point-to point links use broadcast methods with snoop filter options to improve performance. Since broadcast and eavesdropping are inherently non-scalable, these are not used in SCI.

Instead, SCI uses a distributed directory-based cache coherence protocol with a linked list of nodes containing processors that share a particular cache line. Each node holds a directory for the main memory of the node with a tag for each line of memory (same line length as the cache line). The memory tag holds a pointer to the head of the linked list and a state code for the line (three states – home, fresh, gone). Associated with each node is also a cache for holding remote data with a directory containing forward and backward pointers to nodes in the linked list sharing the cache line. The tag for the cache has seven states (invalid, only fresh, head fresh, only dirty, head dirty, mid valid, tail valid).

The distributed directory is scalable. The overhead for the directory based cache coherence is a constant percentage of the node’s memory and cache. This percentage is in the order of 4% for the memory and 7% for the cache.

Legacy[edit]

SCI is a standard for connecting the different resources within a multiprocessor computer system and it is not as widely known to the public as for example the Ethernet family for connecting different systems. Different system vendors implemented different variants of SCI for their internal system infrastructure. These different implementations interface to very intricate mechanisms in processors and memory systems and each vendor has to preserve some degrees of compatibility for both hardware and software.

Gustavson led a group called the Scalable Coherent Interface and Serial Express Users, Developers, and Manufacturers Association and maintained a web site for the technology starting in 1996.[3] A series of workshops were held through 1999. After the first 1992 edition,[1] follow-on projects defined shared data formats in 1993,[5] a version using low-voltage differential signaling in 1996,[6] and a memory interface known as Ramlink later in 1996.[7] In January 1998, the SLDRAM corporation was formed to hold patents on an attempt to define a new memory interface that was related to another working group called SerialExpress or Local Area Memory Port.[8][9] However, by early 1999 the new memory standard was abandoned.[10]

In 1999 a series of papers was published as a book on SCI.[11] An updated specification was published in July 2000 by the International Electrotechnical Commission (IEC) of the International Organization for Standardization (ISO) as ISO/IEC 13961.[12]

See also[edit]

References[edit]

  1. ^ a b "IEEE Standard for Scalable Coherent Interface (SCI)". IEEE Std 1596-1992 (IEEE Standards Board). 1992. ISBN 9780738129501. 
  2. ^ David B. Gustavson (September 1991). "The Scalable Coherent Interface and Related Standards Projects". SLAC Publication 5656. Stanford Linear Accelerator Center. Retrieved August 31, 2013. 
  3. ^ a b "Scalable Coherent Interface and Serial Express Users, Developers, and Manufacturers Association". Group web site. Retrieved August 31, 2013. 
  4. ^ "1596 WG - Working Group for Scalable Coherent Interface". Working group web site. Retrieved August 31, 2013. 
  5. ^ "IEEE Standard for Shared-Data Formats Optimized for Scalable Coherent Interface (SCI) Processors". IEEE 1596.5-1993 (IEEE Standards Board). April 25, 1994. ISBN 9780738112091. 
  6. ^ "IEEE Standard for Low-Voltage Differential Signals (LVDS) for Scalable Coherent Interface (SCI)". IEEE IEEE 1596.3-1996 (IEEE Standards Board). July 31, 1996. ISBN 9780738131368. 
  7. ^ "EEE Standard for High-Bandwidth Memory Interface Based on Scalable Coherent Interface (SCI) Signaling Technology (RamLink)". IEEE IEEE 1596.4-1996 (IEEE Standards Board). September 16, 1996. ISBN 9780738131375. 
  8. ^ David B. Gustavson (February 10, 1999). "Organizing for Alternatives". 
  9. ^ David V. James, David B. Gustavson, B. Fleischer (May–Jun 1998). "SerialExpress-a high performance workstation interconnect". IEEE Micro (IEEE): 54–65. doi:10.1109/40.683105. 
  10. ^ David Lammers (February 19, 1999). "ISSCC: SLDRAM group morphs to DDR II". EE Times. 
  11. ^ Hermann Hellwagner, Alexander Reinefeld, ed. (1999). SCI: Scalable Coherent Interface: Architecture and Software for High-Performance Compute Clusters. Lecture Notes in Computer Science. Springer. ASIN 3540666966. ISBN 978-3540666967. 
  12. ^ "Scalable Coherent Interface (SCI)". International Standard ISO/IEC 13961 IEEE Std 1596. July 10, 2000.