How to Reduce Code Size (and Memory Cost) Without Sacrificing Performance
Embedded.com
Nov 29 2005 (17:55 PM)
Today's intelligent compilers offer many options for squeezing more performance out of application code. Many of these optimizations, however, tend to increase overall code size.
As a result, once developers of optimized application code have reached the required performance specifications, there still remains the challenge of bringing code size back under control.
Through an iterative process of building application code using different compiler optimization options and profiling the result, developers can hone in and identify infrequently used and non-critical sections of code to trade off performance where it matters least for reduced code size, providing minimal impact on system performance. Often, varying compiler options to reduce code size can enable developers to decrease the amount of on-chip and external memory an application requires without adversely affecting performance, thereby reducing the overall bill of materials (BOM).
To read the full article, click here
Related Semiconductor IP
- Root of Trust (RoT)
- Fixed Point Doppler Channel IP core
- Multi-protocol wireless plaform integrating Bluetooth Dual Mode, IEEE 802.15.4 (for Thread, Zigbee and Matter)
- Polyphase Video Scaler
- Compact, low-power, 8bit ADC on GF 22nm FDX
Related White Papers
- How silicon and circuit optimizations help FPGAs offer lower size, power and cost in video bridging applications
- How to Reduce FPGA Logic Cell Usage by >x5 for Floating-Point FFTs
- How to use snakes to speed up software without slowing down the time-to-market?
- How a voltage glitch attack could cripple your SoC or MCU - and how to securely protect it
Latest White Papers
- Reimagining AI Infrastructure: The Power of Converged Back-end Networks
- 40G UCIe IP Advantages for AI Applications
- Recent progress in spin-orbit torque magnetic random-access memory
- What is JESD204C? A quick glance at the standard
- Open-Source Design of Heterogeneous SoCs for AI Acceleration: the PULP Platform Experience