"Early and accurate" power analysis: myth or reality?
By Preeti Gupta, Director RTL Product Management, Apache Design, Inc. (a subsidiary of ANSYS)
EETimes (4/11/2012 6:43 PM EDT)
Power is receiving a mounting share of attention. Innovation, fueled by the information and internet age, poses new challenges for electronic systems across a spectrum of applications. Mobile devices continue to break new frontiers of functional integration. Phones are now your email, social networking interface, video and music player, gaming device, camera, GPS, and more – all rolled into one. Yet the smart phone must survive through the day, and hopefully longer, without having to recharge the battery. Data centers and cloud computing grapple with power and carbon footprints as they move and process incredible amounts of data back and forth, consuming electricity to the order of 1-2% of the total that the entire world consumes. Advances in fabrication technology have made it possible for processors and system-on-chips (SoCs) to boast of over three-billion transistors, also pushing the limits of power density, integrity and reliability.
To read the full article, click here
Related Semiconductor IP
- Process/Voltage/Temperature Sensor with Self-calibration (Supply voltage 1.2V) - TSMC 3nm N3P
- USB 20Gbps Device Controller
- SM4 Cipher Engine
- Ultra-High-Speed Time-Interleaved 7-bit 64GSPS ADC on 3nm
- Fault Tolerant DDR2/DDR3/DDR4 Memory controller
Related White Papers
- Throttle IP Core Power Dissipation: Use RTL Power Analysis Early and Often
- Low Power Analysis and Verification of Super Speed Inter-Chip (SSIC) IP
- Accurate and Efficient Power estimation Flow For Complex SoCs
- Estimate power at RTL to identify problems early
Latest White Papers
- Fault Injection in On-Chip Interconnects: A Comparative Study of Wishbone, AXI-Lite, and AXI
- eFPGA – Hidden Engine of Tomorrow’s High-Frequency Trading Systems
- aTENNuate: Optimized Real-time Speech Enhancement with Deep SSMs on RawAudio
- Combating the Memory Walls: Optimization Pathways for Long-Context Agentic LLM Inference
- Hardware Acceleration of Kolmogorov-Arnold Network (KAN) in Large-Scale Systems