Creating Value Through Test
Creating Value Through Test
Erik Jan Marinissen1 Bart Vermeulen1 Robert Madge2 Michael Kessler3 Michael Müller3
1Philips Research Laboratories 2 3
LSI Logic Corp. IBM Deutschland Entwicklung GmbH
IC Design – Digital Design & Test Product Engineering Hardware Development
Prof. Holstlaan 4 – WAY-41 23400 N.E. Glisan Street Schönaicherstrasse 220
5656 AA Eindhoven Gresham, OR 71032 Böblingen
The Netherlands United States of America Germany
[email protected] [email protected] [email protected]
[email protected] [email protected]
Abstract
Test is often seen as a necessary evil; it is a fact of life that ICs have manufacturing defects and those need to be filtered out by testing
before the ICs are shipped to the customer. In this paper, we show that techniques and tools used in the testing field can also be
(re-)used to create value to (1) designers, (2) manufacturers, and (3) customers alike. First, we show how the test infrastructure can
be used to detect, diagnose, and correct design errors in prototype silicon. Secondly, we discuss how test results are used to improve
the manufacturing process and hence production yield. Finally, we present test technologies that enable systems of high reliability
for safety-critical applications.
Program Breakpoint
Breakpoint hit?
N
core 1 core 2 Y
TCB TCB
IC Done?
N
Y
Figure 1: Scan-based silicon debug architecture. End
The architecture in Figure 1 implements the “ABC” of scan-based sil- Figure 2: Flow used for scan-based silicon debug.
icon debug; Access to the scan chains, Breakpoints to detect one or
more internal events, and Clock control. These are explained below. Clock Control
Access The chip is stopped by gating the on-chip clocks. Stopping the clocks
The scan chains are controlled via dedicated TAP data registers. One effectively freezes the content of all flip flops and embedded memo-
clock domain is scanned out at a time, as the TAP has only one serial ries. After the clocks have stopped, the circuit can be safely switched
output. In debug mode, the scannable flip flops are concatenated into to debug scan mode. In debug scan mode, each internal clock is se-
debug scan probes, one in each clock domain. This concatenation is lected in turn to allow the content of the corresponding scan probe to
performed at core-level using a debug shell, shown in gray in Figure 1. be scanned out. The stop and re-activation functions for the internal
The debug shell hides all core-specific details, such as number of scan clocks for scan chain shifting are added to the on-chip clock genera-
chains and clock domains inside the core, and provides a single, uni- tion unit and controlled from a Clock Control DCB.
form hardware interface for debug at the integration level. Each core The scan-based silicon debug architecture presented above uses scal-
provides one serial input and one serial output to its debug scan probes, able modules to implement the required silicon debug functionality.
and a standardized debug interface to an Access Control Debug Con- These modules are added to each of the submodules in the design, al-
trol Block (AC-DCB), that controls the scan probe multiplexing. At lowing a core-based design, test, and debug methodology to be fol-
the integration level, all serial inputs and outputs are daisy-chained. lowed. The advantage of this is that the debug architecture can be
During a silicon debug session, the AC-DCB is used to select each completely tailored to suit a particular design. As an example, a de-
core-level scan probe in turn, while its input and output are connected sign may contain cores with a different number of scan chains and/or
to the chip’s TDI and TDO pins. To the user this complexity in access- clock domains. The debug shell hides these test-details by providing
ing the various scan probes is hidden. The debugger software takes one core-level debug interface. Each of these cores has the same debug
care of issuing the proper TAP commands to select each probe in turn, interface, allowing a design-for-debug tool to automatically perform
and translating the bit-streams received on the chip’s TDO output to the core interconnect at integration level. This reduces the time re-
individual flip flop, multi-bit register, and memory content. quired for a designer to add design-for-debug hardware to a design. In
addition, because of the scalable nature of the debug architecture, an
Breakpoints estimation tool can be used to make an educated trade-off between for
To examine the behavior of the chip in detail using state dumps, it is example breakpoint granularity and hardware cost.
required to first determine at which point during the chip’s execution a
state dump has to be made. An on-chip breakpoint mechanism is added
to the design to allow the chip to be stopped at regular intervals during 1.2 Silicon Debug Successes
its execution. This regularity is important, as it allows the state dumps This debug methodology has been successfully applied to a number of
to provide a clear insight in the data and control processing going on in- large digital system chips within Philips.
side the chip over time. If the breakpoint mechanism does not provide
The debug facilities on the CPA chip [4] proved essential in verify-
enough temporal resolution, the ‘blind spots’ in between state dumps,
ing its video-processing capabilities. The first silicon exhibited several
where no information can be obtained, might seriously complicate and
problems during initialization, causing the chip to malfunction after 50
lengthen the debug process.
to 75 video frames. The source of the problems was detected by exam-
During a silicon debug session, the breakpoint mechanism is pro- ining scan dumps taken at each cycle during the initialization sequence.
grammed via a Breakpoint DCB to stop the chip at a certain point in Using the debug controllability, we replaced the faulty ROM-based ini-
time. After stopping, a scan dump via the TAP is made and compared tialization code by loading corrected code into an external SDRAM and
to golden reference data. Based on this comparison, the end user can instructing the chip to fetch its code from that SDRAM. After this fix,
use the debugger software to re-program the breakpoint mechanism to the designers could successfully verify all image processing functions
without further silicon spins.
For the PNX8525/Viper chip [5], and more recently for the within Philips.
PNX7100/Chrysalis chip, the state dumping functionality allowed de-
signers to correctly diagnose the faulty behavior of subcomponents. In
one case, the flow shown in Figure 2 was repeatedly used to back-track 1.3 Conclusion
mismatches between simulation states and silicon states back to the The presented debug methodology successfully relies on the existing
output of a single gate. Under specific circumstances, the output of scan chain access to debug prototype silicon. With only little extra
this gate was not able to drive an internal signal to the correct value in hardware to create debug access, breakpoints, and clock control, it be-
time, which ultimately resulted in erroneous behavior. Once this was comes possible to obtain state dumps while the chip is in its applica-
discovered, the fix was easy to implement and verify. tion. These state dumps provide the debug engineer with essential in-
As a result of these successes, a standardization activity is currently on- formation to locate design errors still left in the chip, and overall help
going to make this debug methodology available to all digital designs to reduce the number of silicon spins and time-to-market of the chip.
Figure 5: IDDQ data from two silicon lots from the same process. Note
outlier die from both lots which cannot be efficiently screened with the
on-tester limit.
Figure 6: Min-VDD vs. device speed for two different lots, clearly
showing Min-VDD outliers and lot-to-lot intrinsic variation. 2.4 Test Cost Reduction/Adaptive Testing
Test engineers have always utilized raw test data to isolate correlation
or repeatability issues in the test program or hardware. Recent trends,
however, have been towards collection and analysis of large volumes of
test data over longer periods of time. Test times can be reduced through
elimination of unnecessary or redundant tests or vectors and adaptive
test methods can be introduced where the results of certain tests can
determine the need for more extensive testing based on probability of
failure. More recent trends towards foundry wafer manufacturing has
led to the importance of adaptive testing due to the potentially variable
quality of silicon coming from the fabs. Test time improvements of
26% have been reported with adaptive test methods while also improv-
ing product quality [17].
2.5 Conclusion
ATE testing and the raw data results provide ever-increasing value in
the manufacturing of complex integrated circuits. No longer are the
raw data results ignored in favor of pass/fail ‘bin’ results, rather the
results are a critical part of the yield learning, quality and reliability
Figure 7: Statistical Post-Processing (SPP) data flow showing the use improvement and cost reduction process in all areas of manufacturing
of raw parametric ATE data in off-tester pass/fail decision making and and process or product development.
inkless re-binning.
References [13] R. Madge et al. Screening Min-VDD Outliers using Feed-Forward Volt-
age Testing. In Proceedings IEEE International Test Conference (ITC),
[1] E.J. Aas et al. Quantifying Design Quality Through Design Experiments. pages 673–682, October 2002.
IEEE Design & Test of Computers, Vol. 11(No. 1):27–38, 1994.
[14] A. Kinra, H. Balachandran, R. Thomas, and J. Carulli. Logic Mapping
[2] K. Holdbrook, S. Joshi, S. Mitra, J. Petolino, R. Raman, and M. Wong. on a Microprocessor. In Proceedings IEEE International Test Conference
microSPARC: A case-study of scan based debug. In Proceedings IEEE (ITC), pages 701–710, September 2000.
International Test Conference (ITC), pages 70–75, 1994.
[3] B. Vermeulen, T. Waayers, and S.K. Goel. Core-Based Scan Architecture [15] C. Hora, R. Segers, S. Eichenberger, and M. Lousberg. An Effective Di-
for Silicon Debug. In Proceedings IEEE International Test Conference agnosis Method to Support Yield Improvement. In Proceedings IEEE
(ITC), pages 638–647, October 2002. International Test Conference (ITC), pages 260–269, October 2002.
[4] B. Vermeulen and G.J. Rootselaar. Silicon Debug of a Co-Processor Ar- [16] B. Benware. Logic Mapping on ASIC Products. In Proceedings Interna-
ray for Video Application. In Digest of IEEE Intnl. High-Level Design tional Symposium for Testing and Failure Analysis (ISTFA), pages 579–
Verification and Test Workshop (HLDVT), pages 47–52, November 2000. 586, 2002.
[5] B. Vermeulen, S. Oostdijk, and F. Bouwman. Test and Debug Strategy of [17] S. Benner and O. Boroffice. Optimal Production Test Times Through
the PNX8525 Nexperia Digital Video Platform System Chip. In Proceed- Adaptive Test Programming. In Proceedings IEEE International Test
ings IEEE International Test Conference (ITC), pages 121–130, 2001. Conference (ITC), pages 908–915, October 2001.
[6] W. Maly. Yield Diagnostics Through Interpretation of Test Data. In Pro-
ceedings IEEE International Test Conference (ITC), pages 10–20, October [18] M. Kusko, B. Robbins, T. Koprowski, and W. Huott. 99%- AC Test Cov-
1987. erage Using only LBIST on the 1-GHz IBM S/390 zseries 900 Micropro-
cessor. In Proceedings IEEE International Test Conference (ITC), pages
[7] M.A. Merino. SmartBit : Bitmap to Defect Correlation Software for
586–592, October 2001.
Yield Improvement. In Proceedings IEEE/SEMI Advanced Semiconduc-
tor Manufacturing Conference, pages 194–198, 2000. [19] H. Bardell, W.H. McAnney, and J. Savir. Built-In Test for VLSI: Pseudo-
[8] C. Gloor. Embedded Memory Analysis for Standard Cell ASIC Yield random Techniques. John Wiley & Sons, Chichester, 1987.
Enhancement. In Proceedings International Symposium for Testing and [20] R.F. Rizzolo et al. System performance management for the S/390 Parallel
Failure Analysis (ISTFA), pages 69–76, 2000. Enterprise Server G5 1999. IBM Journal of Research and Development,
[9] A. Gattiker and W. Maly. Current Signatures. In Proceedings IEEE VLSI Vol. 43(No. 5/6):651–660, 1999.
Test Symposium (VTS), pages 112–117, April 1996.
[21] M. Müller et al. RAS Strategy for IBM S/390 G5 and G6. IBM Journal
[10] T.J. Powell, J. Pair, M. St.John, and D. Counce. Delta IDDQ for Test-
of Research and Development, Vol. 43(No. 5/6):875–888, 1999.
ing Reliability. In Proceedings IEEE VLSI Test Symposium (VTS), pages
439–443, April 2000. [22] B.W. Curran et al. IBM eServer z900 High Frequency Microprocessor
[11] W.R. Daasch, R. Madge, K. Cota, and J. McNames. Neighbor Selection Technology, Circuits, and Design Methodology. IBM Journal of Research
for Variance Reduction in IDDQ and Other Parametric Data. In Proceed- and Development, Vol. 46(No. 4/5):631–644, 2002.
ings IEEE International Test Conference (ITC), pages 92–100, October [23] L. Spainhower and T. Gregg. IBM S/390 Parallel Enterprise Server G5
2001. Fault Tolerance: A Historical Perspective. IBM Journal of Research and
[12] R. Madge, M. Rehani, K. Cota, and W.R. Daasch. Statistical Post- Development, Vol. 43(No. 5/6):863–873, 1999.
Processing at Wafersort – An Alternative to Burn-in and a Manufacturable
Solution to Test Limit Setting for Sub-micron Technologies. In Proceed- [24] L.C. Alves et al. RAS Design for the IBM eServer z900. IBM Journal of
ings IEEE VLSI Test Symposium (VTS), pages 69–74, April 2002. Research and Development, Vol. 46(No. 4/5):503–521, 2002.