RapidIO

From Infogalactic: the planetary knowledge core
Jump to: navigation, search
RapidIO
200px
RapidIO - the unified fabric for Performance Critical Computing
Year created 2000; 24 years ago (2000)
Width in bits Port widths of 1, 2, 4, 8, and 16 lanes
Number of devices Sizes of 256, 65,536, and 4,294,967,296
Speed Per lane (each direction):
  • v1.x: 1.25, 2.5, and 3.125 GBd
  • v2.x: added 5.0, 6.25 GBd
  • v3.x: added 10.3125 GBd
Style Serial
Hotplugging interface Yes
External interface Yes, Chip-Chip, Board-Board (Backplane), Chassis-Chassis

The RapidIO architecture is a high-performance packet-switched, interconnect technology. RapidIO supports messaging, read/write and cache coherency semantics. RapidIO fabrics guarantee in-order packet delivery, enabling power- and area- efficient protocol implementation in hardware. Based on industry-standard electrical specifications such as those for Ethernet, RapidIO can be used as a chip-to-chip, board-to-board, and chassis-to-chassis interconnect. The protocol is marketed as: RapidIO - the unified fabric for Performance Critical Computing,[1] and is used in many applications such as Data Center & HPC, Communications Infrastructure, Industrial Automation and Military & Aerospace that are constrained by at least one of size, weight, and power (SWaP).

History

RapidIO has its roots in energy-efficient, high-performance computing. The protocol was originally designed by Mercury Computer Systems and Motorola (Freescale) as a replacement for Mercury’s RACEway proprietary bus and Freescale's PowerPC bus.[2] The RapidIO Trade Association was formed in February 2000, and included telecommunications and storage OEMs as well as FPGA, processor, and switch companies. The protocol was designed to meet the following objectives:

  • Low latency
  • Guaranteed, in order, packet delivery
  • Support for messaging and read/write semantics
  • Could be used in systems with fault tolerance/high availability requirements
  • Flow control mechanisms to manage short-term (less than 10 microseconds), medium-term (tens of microseconds) and long-term (hundreds of microseconds to milliseconds) congestion
  • Efficient protocol implementation in hardware
  • Low system power
  • Scales from two to thousands of nodes

The RapidIO Specification Revision 1.1, released in 2001, defined a wide, parallel bus. This specification did not achieve extensive commercial adoption.

The RapidIO Specification Revision 1.2, released in 2002,[3] defined a serial interconnect based on the XAUI physical layer. Devices based on this specification achieved significant commercial success within wireless baseband,[4] imaging and military compute. [5]

The RapidIO Specification Revision 2.0, released in 2008,[6] added more port widths (2×, 8×, and 16×) and increased the maximum lane speed to 6.25 Gbd (gigabaud). Revision 2.1 has repeated and expanded the commercial success of the 1.2 specification.[7]

The RapidIO Specification Revision 3.0, released in 2013,[8] has the following changes and improvements compared to the 2.x specifications:

  • Based on industry-standard Ethernet 10GBASE-KR electrical specifications for short (20 cm + connector) and long (1 m + 2 connector) reach applications
  • Directly leverages the Ethernet 10GBASE-KR DME training scheme for long-reach signal quality optimization
  • Defines a 64b/67b encoding scheme (similar to the Interlaken standard) to support both copper and optical interconnects and to improve bandwidth efficiency
  • Dynamic asymmetric links to save power (for example, 4× in one direction, 1× in the other)
  • Addition of a time synchronization capability similar to IEEE 1588, but much less expensive to implement
  • Support for 32-bit device IDs, increasing maximum system size and enabling innovative hardware virtualization support
  • Revised routing table programming model simplifies network management software
  • Packet exchange protocol optimizations

RapidIO dominates wireless infrastructure

RapidIO fabrics enjoy dominant market share in global deployment of cellular infrastructure 3G, 4G & LTE networks with millions of RapidIO ports shipped[9] into wireless base stations worldwide. RapidIO fabrics were originally designed to support connecting different types of processors from different manufacturers together in a single system. This flexibility has driven the widespread use of RapidIO in wireless infrastructure equipment where there is a need to combine heterogeneous, DSP, FPGA and communication processors together in a tightly coupled system with low latency and high reliability.

RapidIO used in data center / HPC analytics

Data Center and HPC Analytics systems have been deployed using a RapidIO 2D Torus Mesh Fabric,[10] that provides a high speed general purpose interface among the system cartridges for applications that benefit from high bandwidth, low latency node-to-node communication. The RapidIO 2D Torus unified fabric is routed as a torus ring configuration connecting up to 45 server cartridges capable of providing 5Gbs per lane connections in each direction to its north, south, east and west neighbors. This allows the system to meet many unique HPC applications where efficient localized traffic is needed.

Also, using an open modular data center and compute platform,[11] a heterogeneous HPC system has showcased the low latency attribute of RapidIO to enable real-time analytics.[12] In March 2015 a top-of-rack switch was announced to drive RapidIO into mainstream data center applications.[13]

RapidIO in aerospace

The interconnect or "bus" is one of the critical technologies in the design and development of spacecraft avionic systems that dictates its architecture and level of complexity. There are a host of existing architectures that are still in use given their level of maturity. These existing systems are sufficient for a given type of architecture need and requirement. Unfortunately, for next generation missions a more capable avionics architecture is desired; which is well beyond the capabilities levied by existing architectures. A viable option toward the design and development of these next generation architectures is to leverage existing commercial protocols capable of accommodating high levels of data transfer.

In 2012, RapidIO was selected by the Next Generation Spacecraft Interconnect Standard (NGSIS) working group to serve as the foundation for standard communication interconnects to be used in spacecraft. The NGSIS is an umbrella standards effort that includes RapidIO Version 3.1 development, and a box level hardware standards effort under VITA 78 called SpaceVPX or High ReliabilityVPX. The NGSIS requirements committee developed extensive requirements criteria with 47 different elements for the NGSIS interconnect. Independent trade study results by NGSIS member companies demonstrated the superiority of RapidIO over other existing commercial protocols, such as InfiniBand, Fibre Channel, and 10G Ethernet. As a result, the group decided that RapidIO offered the best overall interconnect for the needs of next-generation spacecraft.[14]

RapidIO specification 3.1 released

The RapidIO Specification Revision 3.1, released in 2014,[15] was developed through a collaboration between the RapidIO Trade Association and NGSIS. Revision 3.1 has the following enhancements compared to the 3.0 specification:

  • MECS Time Synchronization protocol for smaller embedded systems. MECS Time Synchronization supports redundant time sources. This protocol is lower cost than the Timestamp Synchronization Protocol introduced in revision 3.0
  • PRBS test facilities and standard register interface.
  • Structurally Asymmetric Link behavioral definition and standard register interface. Structurally Asymmetric Links carry much more data in one direction than the other, for applications such as sensors or processing pipelines. Unlike dynamic asymmetric links, Structurally Asymmetric Links allow implementers to remove lanes on boards and in silicon, saving size, weight, and power. Structurally asymmetric links also allow the use of alternative lanes in the case of a hardware failure on a multi-lane port.
  • Extended error log to capture a series of errors for diagnostic purposes
  • Space device profiles for endpoints and switches, which define what it means to be a space-compliant RapidIO device.

PHY roadmap

The RapidIO roadmap aligns with Ethernet PHY development. RapidIO specifications for 25 GBd and higher links are in development.[16]

Terminology

Link Partner
One end of a RapidIO link.
Endpoint
A device that can originate and/or terminate RapidIO packets.
Processing Element
A device which has at least one RapidIO port
Switch
A device that can route RapidIO packets.

Protocol overview

The RapidIO protocol is defined in a 3-layered specification:

  • Physical: Electrical specifications, PCS/PMA, link-level protocol for reliable packet exchange
  • Transport: Routing, multicast, and programming model
  • Logical: Logical I/O, messaging, global shared memory (CC-NUMA), flow control, data streaming

System specifications include:

  • System Initialization
  • Error Management/Hot Swap

Physical layer

The RapidIO electrical specifications are based on industry-standard Ethernet and Optical Interconnect Forum standards:

  • XAUI for lane speeds of 1.25, 2.5, and 3.125 GBd
  • OIF CEI 6+ Gbit/s for lane speeds of 5.0 and 6.25 GBd
  • 10GBASE-KR 802.3-ap (long reach) and 802.3-ba (short reach) for lane speeds of 10.3125 GBd

The RapidIO PCS/PMA layer supports two forms of encoding/framing:

  • 8b/10b for lane speeds up to 6.25 GBd
  • 64b/67b, similar to that used by Interlaken for lane speeds over 6.25 GBd

Every RapidIO processing element transmits and receives three kinds of information: Packets, control symbols, and an idle sequence.

Packets

Every packet has two values that control the physical layer exchange of that packet. The first is an acknowledge ID (ackID), which is the link specific, unique 5, 6, or 12 bit value that is used to track packets exchange on a link. Packets are transmitted with serially increasing ackID values. Because the ackID is specific to a link, the ackID is not covered by CRC, but by protocol. This allows the ackID to change with each link it passes over, while the packet CRC can remain a constant end-to-end integrity check of the packet. When a packet is successfully received, it is acknowledged using the ackID of the packet. A transmitter must retain a packet until it has been successfully acknowledged by the link partner.

The second value is the packet's physical priority. The physical priority is composed of the Virtual Channel (VC) identifier bit, the Priority bits, and the Critical Request Flow (CRF) bit. The VC bit determines if the Priority and CRF bits identify a Virtual Channel from 1 through 8, or are used as the priority within Virtual Channel 0. Virtual Channels are assigned guaranteed minimum bandwidths. Within Virtual Channel 0, packets of higher priority can pass packets of lower priority. Response packets must have a physical priority higher than requests in order to avoid deadlock.

The physical layer contribution to RapidIO packets is a 2-byte header at the beginning of each packet that includes the ackID and physical priority, and a final 2-byte CRC value to check the integrity of the packet. Packets larger than 80 bytes also have an intermediate CRC after the first 80 bytes. With one exception a packet's CRC value(s) acts as an end-to-end integrity check.

Control symbols

RapidIO control symbols can be sent at any time, including within a packet. This gives RapidIO the lowest possible in-band control path latency, enabling the protocol to achieve high throughput with smaller buffers than other protocols.

Control symbols are used to delimit packets (Start of Packet, End of Packet, Stomp), to acknowledge packets (Packet Acknowledge, Packet Not Acknowledged), reset (Reset Device, Reset Port) and to distribute events within the RapidIO system (Multicast Event Control Symbol). Control symbols are also used for flow control (Retry, Buffer Status, Virtual Output Queue Backpressure) and for error recovery.

The error recovery procedure is very fast. When a receiver detects a transmission error in the received data stream, the receiver causes its associated transmitter to send a Packet Not Accepted control symbol. When the link partner receives a Packet Not Accepted control symbol, it stops transmitting new packets and sends a Link Request/Port Status control symbol. The Link Response control symbol indicates the ackID that should be used for the next packet transmitted. Packet transmission then resumes.

IDLE sequence

The IDLE sequence is used during link initialization for signal quality optimization. It is also transmitted when the link does not have any control symbols or packets to send.

Transport layer

Every RapidIO endpoint is uniquely identified by a Device Identifier (deviceID). Each RapidIO packet contains two device IDs. The first is the destination ID (destID), which indicates where the packet should be routed. The second is the source ID (srcID), which indicates where the packet originated. When an endpoint receives a RapidIO request packet that requires a response, the response packet is composed by swapping the srcID and destID of the request.

RapidIO switches use the destID of received packets to determine the output port or ports that should forward the packet. Typically, the destID is used to index into an array of control values. The indexing operation is fast and low cost to implement. RapidIO switches support a standard programming model for the routing table, which simplifies system control.

The RapidIO transport layer supports any network topology, from simple trees and meshes to n-dimensional hypercubes, multi-dimensional toroids, and more esoteric architectures such as entangled networks.

The RapidIO transport layer enables hardware virtualization (for example, a RapidIO endpoint can support multiple device IDs). Portions of the destination ID of each packet can be used to identify specific pieces of virtual hardware within the endpoint.

Logical layer

The RapidIO logical layer is composed of several specifications, each providing packet formats and protocols for different transaction semantics.

Logical I/O

The logical I/O layer defines packet formats for read, write, write-with-response, and various atomic transactions. Examples of atomic transactions are set, clear, increment, decrement, swap, test-and-swap, and compare-and-swap.

Messaging

The Messaging specification defines Doorbells and Messages. Doorbells communicate a 16-bit event code. Messages transfer up to 4K of data, segmented into up to 16 packets each with a maximum payload of 256 bytes. Response packets must be sent for each Doorbell and Message request. The response packet status value indicates done, error, or retry. A status of retry requests the originator of the request to send the packet again. The logical level retry response allows multiple senders to access a small number of shared reception resources, leading to high throughput with low power.

Flow control

The Flow Control specification defines packet formats and protocols for simple XON/XOFF flow control operations. Flow control packets can be originated by switches and endpoints. Reception of a XOFF flow control packet halts transmission of a flow or flows until an XON flow control packet is received or a timeout occurs. Flow Control packets can also be used as a generic mechanism for managing system resources.

CC-NUMA

The Globally Shared Memory specification defines packet formats and protocols for operating a cache coherent shared memory system over a RapidIO network.

Data streaming

The Data Streaming specification supports messaging with different packet formats and semantics than the Messaging specification. Data Streaming packet formats support the transfer of up to 64K of data, segmented over multiple packets. Each transfer is associated with a Class of Service and Stream Identifier, enabling thousands of unique flows between endpoints.

The Data Streaming specification also defines Extended Header flow control packet formats and semantics to manage performance within a client-server system. Each client uses extended header flow control packets to inform the server of the amount of work that could be sent to the server. The server responds with extended header flow control packets that use XON/XOFF, rate, or credit based protocols to control how quickly and how much work the client sends to the server.

System initialization

Systems with a known topology can be initialized in a system specific manner without affecting interoperability. The RapidIO system initialization specification supports system initialization when system topology is unknown or dynamic. System initialization algorithms support the presence of redundant hosts, so system initialization need not have a single point of failure.

Each system host recursively enumerates the RapidIO fabric, seizing ownership of devices, allocating device IDs to endpoints and updating switch routing tables. When a conflict for ownership occurs, the system host with the larger deviceID wins. The "losing" host releases ownership of its devices and retreats, waiting for the "winning" host. The winning host completes enumeration, including seizing ownership of the losing host. Once enumeration is complete, the winning host releases ownership of the losing host. The losing host then discovers the system by reading the switch routing tables and registers on each endpoint to learn the system configuration. If the winning host does not complete enumeration in a known time period, the losing host determines that the winning host has failed and completes enumeration.

System enumeration is supported in Linux by the RapidIO subsystem.

Error management

RapidIO supports high availability, fault tolerant system design, including hot swap. The error conditions that require detection, and standard registers to communicate status and error information, are defined. A configurable isolation mechanism is also defined so that when it is not possible to exchange packets on a link, packets can be discarded to avoid congestion and enable diagnosis and recovery activities. In-band (port-write packet) and out-of-band (interrupt) notification mechanisms are defined.

Form factors

The RapidIO specification does not discuss the subjects of form factors and connectors, leaving this to specific application-focussed communities. RapidIO is supported by the following form factors:

Software

Processor-agnostic RapidIO support is found in the Linux kernel.

Applications

The RapidIO interconnect is used extensively in the following applications:

  • Wireless base stations
  • Aerospace and Military single-board computers, as well as radar, acoustic and image processing systems
  • Video
  • Storage
  • Supercomputing
  • Medical imaging
  • Industrial control and data path applications

RapidIO is expanding into supercomputing, server, and storage applications.

Competing protocols

PCI Express is targeted at the host to peripheral market, as opposed to embedded systems. Unlike RapidIO, PCIe is not optimized for peer-to-peer multi processor networks. PCIe is ideal for host to peripheral communication. PCIe does not scale as well in large multiprocessor peer-to-peer systems, as the basic PCIe assumption of a "root complex" creates fault tolerance and system management issues.

Another alternative interconnect technology is Ethernet. Ethernet is a robust approach to linking computers over large geographic areas, where network topology may change unexpectedly, the protocols used are in flux, and link latencies are large. To meet these challenges, systems based on Ethernet require significant amounts of processing power, software and memory throughout the network to implement protocols for flow control, data transfer, and packet routing. RapidIO is optimized for energy efficient, low latency, processor-to-processor communication in fault tolerant embedded systems that span geographic areas of less than one kilometre.

SpaceFibre is a competing technology for space applications.[17]

See also

References

  1. http://www.rapidio.org
  2. Lua error in package.lua at line 80: module 'strict' not found.
  3. Lua error in package.lua at line 80: module 'strict' not found.
  4. Lua error in package.lua at line 80: module 'strict' not found.
  5. Lua error in package.lua at line 80: module 'strict' not found.
  6. Lua error in package.lua at line 80: module 'strict' not found.
  7. Lua error in package.lua at line 80: module 'strict' not found.
  8. Lua error in package.lua at line 80: module 'strict' not found.
  9. http://www.rcrwireless.com/20121203/opinion/reader-forum-cloud-radio-access-small-cell-networks-based-rapidio
  10. http://www.hpcwire.com/2014/09/24/paypal-finds-order-chaos-hpc/
  11. http://prodrive-technologies.com/prodrive-technologies-announces-datacenter-hpc-system-dccp-280-rapidio-10-gigabit-ethernet/
  12. http://www.businesswire.com/news/home/20141118005342/en/IDT-Orange-Silicon-Valley-NVIDIA-Accelerate-Computing#.VQqdHuF0Uso
  13. http://prodrive-technologies.com/prodrive-technologies-launches-prsb-760g2-large-rapidio-networks/
  14. Lua error in package.lua at line 80: module 'strict' not found.
  15. Lua error in package.lua at line 80: module 'strict' not found.
  16. Lua error in package.lua at line 80: module 'strict' not found.
  17. Lua error in package.lua at line 80: module 'strict' not found.

External links

  • Lua error in package.lua at line 80: module 'strict' not found..
  • Lua error in package.lua at line 80: module 'strict' not found..
  • Lua error in package.lua at line 80: module 'strict' not found..