Measuring the complexity of processor bugs to improve testbench quality
I am often asked the question “When is the processor verification done?” or in other words “how do I measure the efficiency of my testbench and how can I be confident in the quality of the verification?”. There is no easy answer. There are several common indicators used in the industry such as coverage and bug curve. While they are absolutely necessary, these are not enough to reach the highest possible quality. Indeed, such indicators do not really unveil the ability of verification methodologies to find the last bugs. With experience, I learned that measuring the complexity of processor bugs is an excellent indicator to use throughout the development of the project.
What defines the complexity of a processor bug and how to measure it?
Experience taught me that we can define the complexity of a bug by counting the number of independent events or conditions that are required to hit the bug.
To read the full article, click here
Related Blogs
- From DIY To Advanced NoC Solutions: The Future Of MCU Design
- Half of the Compute Shipped to Top Hyperscalers in 2025 will be Arm-based
- Enhancing Edge AI with the Newest Class of Processor: Tensilica NeuroEdge 130 AICP
- Evolution of AMBA AXI Protocol: An Introduction to the Issue L Update
Latest Blogs
- Rivian’s autonomy breakthrough built with Arm: the compute foundation for the rise of physical AI
- AV1 Image File Format Specification Gets an Upgrade with AVIF v1.2.0
- Industry’s First End-to-End eUSB2V2 Demo for Edge AI and AI PCs at CES
- Integrating Post-Quantum Cryptography (PQC) on Arty-Z7
- UA Link PCS customizations from 800GBASE-R Ethernet PCS Clause 172