To install click the Add extension button. That's it.

The source code for the WIKI 2 extension is being checked by specialists of the Mozilla Foundation, Google, and Apple. You could also do it yourself at any point in time.

4,5
Kelly Slayton
Congratulations on this excellent venture… what a great idea!
Alexander Grigorievskiy
I use WIKI 2 every day and almost forgot how the original Wikipedia looks like.
Live Statistics
English Articles
Improved in 24 Hours
Added in 24 Hours
What we do. Every page goes through several hundred of perfecting techniques; in live mode. Quite the same Wikipedia. Just better.
.
Leo
Newton
Brights
Milds

Statistical static timing analysis

From Wikipedia, the free encyclopedia

Conventional static timing analysis (STA) has been a stock analysis algorithm for the design of digital circuits for a long time. However the increased variation in semiconductor devices and interconnect has introduced a number of issues that cannot be handled by traditional (deterministic) STA. This has led to considerable research into statistical static timing analysis, which replaces the normal deterministic timing of gates and interconnects with probability distributions, and gives a distribution of possible circuit outcomes rather than a single outcome.

YouTube Encyclopedic

  • 1/3
    Views:
    536 651
    148 908
    595 957
  • Excel - Time Series Forecasting - Part 1 of 3
  • Panel Data Models with Individual and Time Fixed Effects
  • How To Calculate and Understand Analysis of Variance (ANOVA) F Test.

Transcription

Comparison with conventional STA

Deterministic STA is popular for good reasons:

  • It requires no vectors, so it does not miss paths.
  • The run time is linear in circuit size (for the basic algorithm).
  • The result is conservative.
  • It typically uses some fairly simple libraries (typically delay and output slope as a function of input slope and output load).
  • It is easy to extend to incremental operation for use in optimization.

STA, while very successful, has a number of limitations:

  • Cannot easily handle within-die correlation, especially if spatial correlation is included.
  • Needs many corners to handle all possible cases.
  • If there are significant random variations, then in order to be conservative at all times, it is too pessimistic to result in competitive products.
  • Changes to address various correlation problems, such as CPPR (Common Path Pessimism Removal) make the basic algorithm slower than linear time, or non-incremental, or both.

SSTA attacks these limitations more or less directly. First, SSTA uses sensitivities to find correlations among delays. Then it uses these correlations when computing how to add statistical distributions of delays.

There is no technical reason why determistic STA could not be enhanced to handle correlation and sensitivities, by keeping a vector of sensitivities with each value as SSTA does. Historically, this seemed like a big burden to add to STA, whereas it was clear it was needed for SSTA, so no-one complained. See some of the criticism of SSTA below where this alternative is proposed.

Methods

There are two main categories of SSTA algorithms – path-based and block-based methods.

A path-based algorithm[1] sums gate and wire delays on specific paths. The statistical calculation is simple, but the paths of interest must be identified prior to running the analysis. There is the potential that some other paths may be relevant but not analyzed so path selection is important.

A block-based algorithm[2] generates the arrival times (and required) times for each node, working forward (and backward) from the clocked elements. The advantage is completeness, and no need for path selection. The biggest problem is that a statistical max (or min) operation that also considered correlation is needed, which is a hard technical problem.

There are SSTA cell characterization tools that are now available such as Altos Design Automation's Variety tool.

Criticism

A number of criticisms have been leveled at SSTA:

  • It's too complex, especially with realistic (non-gaussian) distributions.
  • It's hard to couple to an optimization flow or algorithm.
  • It's hard to get the data the algorithm needs. Even if you can get this data, it is likely to be time-varying and hence unreliable.
  • If used seriously by the customers of a fab, it restricts the changes the fab might make, if they change that statistical properties of the process.
  • The benefit is relatively small, compared to an enhanced deterministic STA that also takes into account sensitivities and correlation.[3]

Tools that perform static timing analysis

FPGAs

  • Altera Quartus II
  • Xilinx ISE

ASICs

See also

References

  1. ^ Orshansky, M.; Keutzer, K., 2002, A general probabilistic framework for worst case timing analysis, Design Automation Conference, 2002. Proceedings. 39th, Vol., Iss., 2002, Pages: 556–561.
  2. ^ Visweswariah, C.; Ravindran, K.; Kalafala, K.; Walker, S.G.; Narayan, S.; Beece, D.K.; Piaget, J.; Venkateswaran, N.; Hemmett, J.G., 2006, First-Order Incremental Block-Based Statistical Timing Analysis, IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems, Vol.25, Iss.10, Oct. 2006, Pages: 2170–2180
  3. ^ Noel Menezes. "The Good, the Bad, and the Statistical" (PDF). ISPD 2007.
This page was last edited on 6 March 2024, at 14:07
Basis of this page is in Wikipedia. Text is available under the CC BY-SA 3.0 Unported License. Non-text media are available under their specified licenses. Wikipedia® is a registered trademark of the Wikimedia Foundation, Inc. WIKI 2 is an independent company and has no affiliation with Wikimedia Foundation.