Deliver at 100G: The impact of smart memory
Michael Miller, Mosys
EDN (July 24, 2014)
Middleware and full function appliance box design teams face the daunting challenge of developing and meeting the performance requirements for next generation 100Gbps. Using general-purpose multi-core CPU arrays provides the flexibility needed to support emerging trends like SDN and NFV. However the packet inspection and buffering functions point to the need to direct traffic to achieve load-balanced cores. Meeting these requirements in the same form factor as earlier generation 10Gbps and 40Gbps products can seem next to impossible. With new developments in serial memory solutions, middleware systems can now buffer, process packets, and steer traffic to general purpose CPU arrays 2-10 times faster than previous 10G or 40G solutions.
The Form Factor Challenge: Fitting 100G in a 40G Box
To address the form factor issue, design teams at both the device and appliance level need to understand the importance of the trends in 100G networks. Although 10G and 40G architectures are similar to 100G, 100G (and higher) designs pose trade-off challenges compounded by multiple factors. These include higher bandwidths and an order of magnitude increase in the rate of lookups. This increase is being driven by requirements for more statistics to monitor performance, increased security measures, and expanded functionality.
Since the rate of lookups is growing faster than the CPU can process, this growth fuels the need for pre-filtering. Like every other electronic design, 100G (and up) architectures face the same trade-offs of area, power, performance, and pin count. Traditional memory solutions exhibit ongoing access rate issues versus processing speeds. If you combine traditional memory access rates with the maximum 2 Gbits per pin capacity, it leads to the conclusion that there simply aren’t enough available pins. In addition, if the design were able to have sufficient pins, the amount of memory required poses a significant challenge because of the area consumed.
To read the full article, click here
Related Semiconductor IP
- NPU IP Core for Mobile
- NPU IP Core for Edge
- Specialized Video Processing NPU IP
- HYPERBUS™ Memory Controller
- AV1 Video Encoder IP
Related White Papers
- MIPI in next generation of AI IoT devices at the edge
- The benefit of non-volatile memory (NVM) for edge AI
- The Growing Importance of AI Inference and the Implications for Memory Technology
- DDR Memory Systems at the Heart of Consumer Electronics
Latest White Papers
- Ramping Up Open-Source RISC-V Cores: Assessing the Energy Efficiency of Superscalar, Out-of-Order Execution
- Transition Fixes in 3nm Multi-Voltage SoC Design
- CXL Topology-Aware and Expander-Driven Prefetching: Unlocking SSD Performance
- Breaking the Memory Bandwidth Boundary. GDDR7 IP Design Challenges & Solutions
- Automating NoC Design to Tackle Rising SoC Complexity