Você está na página 1de 6

Advances in Variation-Aware Modeling, Verication, and Testing of Analog ICs

Dimitri De Jonghe , Elie Maricau , Georges Gielen , Trent McConaghy , Bratislav Tasi , and Haralampos Stratigopoulos c
Leuven, Heverlee, Belgium. georges.gielen@esat.kuleuven.be Design Automation Inc., Canada. trentmc@solidodesign.com NXP Semiconductor Inc., Eindhoven, The Netherlands. bratislav.tasic@nxp.com TIMA Laboratory, Grenoble, France. haralampos.stratigopoulos@imag.fr
Solido K.U.

AbstractThis tutorial paper describes novel scalable, nonlinear/generic, and industrially-oriented approaches to perform variation-aware modeling, verication, fault simulation, and testing of analog/custom ICs. In the rst section, Dimitri De Jonghe, Elie Maricau, and Georges Gielen present a new advance in extracting highly nonlinear, variation-aware behavioral models, through the use of data mining and a re-framing of the modelorder reduction problem. In the next section, Trent McConaghy describes new statistical machine learning techniques that enable new classes of industrial EDA tools, which in turn are enabling designers to perform fast and accurate PVT / statistical / highsigma design and verication. In the third section, Bratislav Tasi presents a novel industrially-oriented approach to analog c fault simulation that also has applicability to variation-aware design. In the nal section, Haralampos Stratigopoulos describes describes state-of-the-art analog testing approaches that address process variability.

points are retrieved from the database and stitched together by interpolation [3]. Transfer Function Trajectories (TFT) is a recent approach that transforms MNA matrices into a mixed state space/frequency description using nonlinear regression [4], [5]. The resulting models contain a compact set of equations, which are more easily ported to general-purpose simulators than an interpolation database. They have good scalability, accuracy, and need minimal training. For industrial usage, process variations must be considered; to handle them, TFT exploits modern machine learning techniques [6]. We now describe MNA-based modeling, then regressionbased model generation, and nally experimental results. B. MNA-Based Modeling A general expression for MNA-modelled circuits is given by the nonlinear branch equation: d q(v) + i(v) = Biin (1) dt which is the typical representation of a system using a SPICE simulator. v = v(t) RN are the N unknown node voltages and inductor currents in the circuit. iin = iin (t) RMi are the external inputs to the circuit. q(.) and i(.) RN N are matrix-valued functions describing the charges and currents of nonlinear components. B RN Mi is a constant incidence matrix mapping the inputs to the internal nodes of the circuit. A general trajectory sampling method stores the linearized conductance and capacitance matrices, G and C, that are returned at each time point k by the Newton-Rhapson algorithm during transient simulation [1], [2]: G(k) = i(v)|vk , C (k) = q(v)|vk (2)

I. VARIATION -AWARE B EHAVIORAL M ODELING OF A NALOG C IRCUITS A. Introduction We present a new technique for accurate, reliable, behavioral model extraction for large analog circuits under process variations. For a designer, the ideal tool does oneclick macromodel generation from a given circuit netlist, with sufcient speed and accuracy, and without having to go through procedural difculties like model selection and training. However, todays analog/custom design ows have a high degree of specialization, and increasingly have variation / reliability issues on advanced technology nodes. Meeting all these requirements at once is a challenge for EDA developers. Nonlinear Model Order Reduction (MOR) techniques, which extract the necessary data from the internal circuit description, are becoming a standard for accurate and efcient model extraction [1]. In MOR techniques, data mining is applied on the Modied Nodal Analysis (MNA) matrices while the circuit is simulated in transient analysis [2]. Before the matrix samples are stored in a database, useful information is reduced to a minimum by matrix projection. The resulting model is a collection of low-rank / low-order local expansion points of the original system. During evaluation, the model
0

After running projection-based MOR, the reduced matrices can be used as a low-order expansion point in a piecewise model [2]. Alternatively, a compact nonlinear model can be found by transforming the MNA samples into the frequency domain: H (k) (s) = V (k) (s)
(k) Iin (s)

978-3-9810801-8-6/DATE12/ c 2012 EDAA

= G(k) + s C (k)

(3)

where H (k) (s) is the frequency response at time step k. The data transformation is depicted more in detail in Fig. 1.

Fig. 2. Fig. 1. Transformation of MNA trajectory samples into the frequency domain.

Variation-aware TFT modeling ow.

The Transfer Function Trajectory (TFT) approach models the state-dependent transfer functions as a hyperplane T (x(k) , s) in a mixed state space/frequency domain [5]: H (k) (s) T (x(k) , s) = rp (x(k) ) s + ap (x(k) ) p=1
P

D. Experimental Results The variation-aware TFT approach has been veried on the high-speed output buffer example, described in detail in [5]. The average normalized error (NMSE) for FFX, MARS and LS-SVM is 1.3%, 2.7% and 2.5% respectively.

(4)

Here, x(k) = x[k] is a state estimator that links each state k to l1 delayed inputs (FIR), l2 delayed autoregressive outputs (AR) and PVT parameters such as process variations p , mismatch m and temperature T : x[t] = iin [t], ..., iin [t l1 ], v[t 1], ..., v[t l2 ], p , m , T (5) A Hammerstein model breaks an n-dimensional nonlinear mapping into three sequential blocks: an input nonlinearity, then a linear mapping, and a nonlinear output mapping. TFT casts the approximation of system (4) into such a model [5], [7]. For the linear block, the Vector Fitting Algorithm extracts a common set of P N stable poles aP . The nonlinear blocks are derived from the indenite integral of the residue data {p (x[k])}. The next section explores suitable highr dimensional nonlinear regressors for these residue samples. C. Multivariate Nonlinear Residue Regression The computational complexity of the nonlinear Hammerstein functions is determined both by the number of deterministic input parameters and stochastic process parameters. The dimensionality of the problem is l1 + l2 + K + 1, where K is the number of explanatory PVT parameters. Moreover, strongly nonlinear dynamic behavior is expected for large signal input waveforms. We consider these regression approaches: multivariate adaptive regression splines (MARS) [8], least-squares support vector machines (LS-SVM) [9] and a recent deterministic symbolic regression technique, fast function extraction (FFX) [6]. Interpolation algorithms are not considered due to their poor extrapolation performance. Fig. 2 shows the model generation ow. First, a set of PVT samples is generated with, for example, Monte-Carlo sampling. Of course, this can be replaced by a more advanced active learning strategy. The data is further tted with Vector Fitting and Residue Fitting and nally translated to a Hammerstein structure.
FIR AR PVT

Fig. 3. Output buffer results: Left: Normalized model error (NMSE) for FFX as a function of process variations. Right: Comparison of the TFT regressors for the static nonlinear function. The fast (FF), typical (TT) and slow (SS) corner are also plotted.

II. I NDUSTRIAL VARIATION -AWARE D ESIGN AND V ERIFICATION OF C USTOM IC S As Moores Law progresses [10] and variation gets worse, the traditional approaches are becoming inadequate. We now describe the issues, and how modern variation-aware ows and tools can help manage PVT and statistical variation. A. Fast PVT Design and Verication Process, voltage, and temperature (PVT) variations are often modeled as a set of PVT corners. Traditionally, only a few were needed: {FF, SS} x {min V, max V} x {min T, max T} = 8. But modern processes have more modelsets, and tighter margins mean that more intermediate values must be considered. Consider the reference VCO from the TSMC AMS Ref. Flow 2.0 [11] on TSMC 28nm. A reasonable setup has 15 modelset values, 3 values for temperature, and 5 values for each of its three voltage variables, totalling 3375 corners. HSPICET M takes 70 s to simulate this, therefore it takes 66 hours to evaluate all corners. Designers may cope by guessing which corners cause the worst-case performance, but a wrong guess could mean failure in testing (leading to a re-spin), or failure in the eld. One alternative is linear sensitivity analysis, but it will miss the worst-case corners whenever the response is not linear enough. Quadratic modeling is a bit more general, but will fail if the response is not quadratic enough.

A new alternative is FastPVT [12], which is both fast and accurate. The idea is to cast the PVT problem as a global optimization problem: minimize performance in the space of PVT variation; then to solve it reliabily using adaptive machine learning with arbitrarily nonlinear models [13] and SPICE in the loop. On the VCO listed above, FastPVT found the worstcase PVT corner, with condence, in 371 simulations (versus 3375, for a 9.1x speedup). We have benchmarked FastPVT on a suite of 108 representative industrial problems. Figure 4 left shows the distribution of speedups. The average speedup was 11.3x, and the maximum speedup was 43.1x. For the 56 problems with >200 corners, the average speedup was 19.3x. A few problems had speedup of just 1.0x, which simply indicated that there was so little structure in the mapping from PVT variables to outputs that adaptive modeling could not help. FastPVT is part of the rapid-iteration design ow shown in Figure 5 left. The idea is to rst extract one or a few design-specic PVT corners; then to design against them with feedback from SPICE; and nally to verify. Design iterations are fast because there are so few corners. If verication succeeds, then the designer can go to layout; otherwise he/she adds the failed corners and goes back to the design step.

on statistical corners will improve the whole performance distribution. This is followed up by verication. Figure 5 middle and right illustrates the statistical ows.

Fig. 5. Fast, accurate, scalable corner-based design ows for PVT (left), 3-sigma statistical (middle), and high-sigma statistical (right).

Fig. 4. Left: Distribution of FastPVT speedups on 108 industrial problems. Right: QQ plot of sense amp power distribution, comparing 1M MC samples (1M simulations) to 100M HSMC samples (<10K simulations).

B. Fast Statistical Design and Verication Of course, PVT is not always the way. Some designers have access to sufciently good statistical MOS models to consider doing statistical analysis, which is inherently more accurate than PVT. However, since Monte Carlo (MC) simulations are far too slow within the design loop, MC is traditionally run as a verication afterthought. For high-sigma, the challenge is even greater, since it is not feasible to do the 5 billion MC simulations to verify at 6 sigma yield. Finally, designers tend to think in terms of corners, not statistics. There is a way for handle these challenges. The key is to extract statistical corners that actually bound the 3-sigma or 6-sigma output performances for the circuit, versus traditional MOS corners like FF which bound the performances of the device. Then, changing device sizes to improve performance

The statistical ows in Figure 5 require corner extraction / verication tools that are fast, accurate, and scalable. These lead to subproblems that can be addressed via advanced statistical, optimization, and machine learning techniques. 3-sigma corner extraction uses nonparametric density estimation [14] on a small set of MC samples to identify a 3-sigma target output value, then optimizes in process variable space to nd a process point that meets the target value. 3-sigma verication relies on scalable low-discrepancy sampling [15] to minimize the number of samples to verify 3-sigma yield. For high-sigma, importance sampling [16] and statistical blockade [17] are popular in the literature; unfortunately, those have yet to demonstrate scalability to more than 6-12 process variables. High-sigma MC (HSMC) [18] generates a huge set of MC samples (e.g. 5 billion), then uses adaptive machine learning to rapidly identify the MC samples that have extreme output values and simulate them. HSMC has bounded complexity because it operates on a nite set of samples; and has been applied to industrial problems with >1000 variables. Figure 4 right shows example results on a sense amp with 150 process variables; we see that HSMC is able to correctly identify the tail (top right hand corner) from 100M samples generated in <10K simulations. These advanced techniques enable corner-based, accurate, and scalable ows for PVT, statistical, and high-sigma variation. As Figure 5 illustrates, the ows are are identical, except for the corner extraction / verication tool. The result is a unied, designer-friendly approach for variation-aware design. III. FAST FAULT S IMULATION A LGORITHM FOR FAULT AND M ONTE -C ARLO S IMULATIONS A. Introduction Simulating modern Analog Mixed Signal circuits typically requires signicantly large CPU time, especially for Defect Oriented Test (DOT) Fault and Monte-Carlo (MC) simulations, where one needs to perform a signicant number of simulations per circuit, see [19]. Fault simulations for DOT involve calculations of the impact of probable defects (typically bridges or opens) that are injected into the netlist of the circuit and performing analog simulation (e.g. DC, Transient, and AC) of the predened test benches. Since the

number of possible defects is large (typically 1000-50000 for moderately large IC) CPU time needed to complete the analog simulations is hardly feasible. Therefore, multiple ideas have been introduced to overcome this issue, see e.g. [20] [22]. In [23], [24] we introduced applications of the novel simulation technique that provides a signicant speedup for fault simulations. Here we provide an overview of the Fast Fault Simulation algorithm with the extension, results obtained and possible applications for statistical MC simulations. B. Fast Fault Simulation Algorithm The standard way of performing the fault simulations is launching a series of sequential simulation runs. The runs are unrelated, i.e. every fault run simulates the entire time interval without any reuse of the knowledge obtained from the golden (fault free) and previous fault runs. The main idea of the Fast Fault Simulation approach is the reuse of the already obtained simulation data by dening a parallel run during which all the circuits are simulated simultaneously, as shown in Figure 6.

CAN Transceivers, while CDAC stands for Control Digital to Analog Converter, an IP block used in multiple NXP products.
Test case TJA1050 DC TJA1050 TR TJA1055 TR CDAC Test1 CDAC Test2 Number of faults 923 923 200 100 100 CPU Time [s] fault-free circuit 17278.56 2750.54 117984 216756 6852513 CPU Time [s] faulty circuits 40.91 139.36 36419.72 56477.76 662074.77 Speedup 422.8 19.8 3.2 3.9 10.4

TABLE I FFS SPEEDUP RESULTS ON SEVERAL NXP DESIGNS

The results shown in Table I are produced by using NXP in-house simulator Pstar. However, the idea is independent of simulator or simulation technique, since the algorithm can be easily superposed to any analysis of choice. C. Fault Sensitivity Analysis Although FFS algorithm introduces a signicant speedup, for the already mentioned DOT setups the performance is not always sufcient. Therefore, for most frequently used and time-consuming transient simulations we exploit two additional characteristics of the fault simulation: 1) the main objective of the DOT method is to establish the detection status of a defect and not to determine the actual output values of the faulty circuit, 2) the output value is often measured at a few time points and therefore it is sufcient to compute the output value only for these time points while many time points have to be computed in a standard transient analysis. Therefore, we introduce a notion of the numerical discrete bridge model that is present in the topology of the circuit at measurement points only, as shown in Figure 8.

Fig. 6.

Comparison between standard approach and Fast Fault Simulation

The Fast Fault Simulation concept is based on usage of the already obtained golden circuit solution as the initial value of the Newton-Raphson iterative process (enabling faster convergence) and the bypassing of the unchanged circuit parts which are unaffected by the fault impact. The benet of two speedup concepts can be illustrated by an example, shown in Figure 7, where an NXP automotive product with 61 faults has been analyzed by applying a transient test.

Fig. 8. Time points reduction: (a) standard model (b) bridge present at the measurements points only

Fig. 7. Reuse of golden solution and bypassing benets of FFS algorithm (bypassing = 1 means the whole matrix is solved)

Assuming that all fault simulations require similar effort as the golden simulation (workload corresponding to the rst input on the horizontal axis), it is clear that the algorithm introduce benets from the both aspects which are multiplied to provide the nal speedup. In practical applications, the algorithm performs signicantly better than the standard approach, as shown in Table I, where several designs have been analyzed. TJA1050 and TJA1055 are NXP products from the family of

Of course, such an approach introduces signicantly larger speedups, as well as inaccuracies of the fault solution. To keep, at least partly, the accuracy under control, we introduce the nonlinear correction technique, which is based on NewtonRaphson iterative correction steps. The algorithm has been applied on already mentioned TJA1055 with 18 tests and 38000 extracted faults. Matching vs. speedup results are shown in Table II. Since there are typically a large number of tests, overall accuracy is at the acceptable level for practical application. For example, total of 150 tests for TJA1055 has 95.1% average matching, which proved accurate enough when compared to 1.5M IC samples measured.

Algorithm Standard Linear FSA Nonlinear FSA (5 iter) Nonlinear FSA (10 iter)

Worst match 100% 72% 89% 92%

Best match 100% 97% 97% 97%

Speedup 1 1050 238 173

CPU time 3 years 1 day 4.2 days 5.8 days

TABLE II FAULT SENSITIVITY ANALYSIS PERFORMANCE

D. Monte-Carlo Simulations Comparing MC and fault simulation problems, it is clear that there are similarities. For example, many MC samples are quite close to the nominal value or at least to each other. Therefore, combining multiple samples in a joined run would allow to exploit similar benets as introduced for fault simulation problem. The main difculty here is how to superpose the simulation algorithm to dynamic sampling techniques, i.e. the ones that choose new samples based on the outcome of the previous ones. IV. A DVANCES IN VARIATION -AWARE T ESTING OF A NALOG IC S A. Analog Test Challenges in the SoC Era Todays circuit design trends, namely channel-length scaling, dense integration of heterogeneous systems onto a single die, reduction of the form factor, etc., as well as the requirement to meet stringent specications, result inadvertently in increased process variability. This reality places an ever increasing emphasis on efcient test strategies to screen out and discard circuits that fail to meet the desired specications. To this end, there has been a shift in the analog test paradigm in recent years towards full specication-based test suites. The current practice nowadays is to verify indiscriminately, oneby-one all the performances that are promised in the datasheet. As a result, the cost of high-volume testing has been increased dramatically in the recent years. According to recent industry reports [25], testing the analog, mixed-signal, and RF functions of a modern SoC may reach more than 50% of the overall production cost, including silicon and packaging costs. The test cost is mainly due to the sophisticated automatic test equipment that is required and the lengthy test times that are involved. Thus, the reduction of test cost can be achieved by minimizing the test time per device, by increasing the number of devices that are tested in parallel, and by alleviating the dependence on expensive test equipment. B. Towards Analog Test Cost Reduction 1) Statistical test: A generic approach to reduce the test cost is to replace the specication-based tests by alternate measurements that can be extracted rapidly on a low-cost assortment of test equipment. Thereafter, the performances can be implicitly inferred from the alternate measurements, as long as there exists a high level of correlation between them. Typically, however, it is not possible to obtain a closed-form mathematical expression of the mapping between the alternate measurements and the performances. For this purpose, the

mapping is derived through statistical learning which employs a representative sample of devices that is collected during the ramp-up phase across different wafers and lots. Once the mapping is derived, it can be readily used to test future devices based solely on alternate measurements. There are two types of mappings that we can establish. One possibility is to perform a direct go/no-go test by learning the mapping f1 : x pass/f ail, where x denotes the vector of alternate measurements. The simplest approach is to assign test limits individually to each of the alternate measurements which effectively results in examining the footprint of x with respect to the position of a hyper-rectangular in the alternate measurement space. However, a hyper-rectangular acceptance region is often a crude approximation and, thus, a more sophisticated approach is to allocate a non-linear hyper-surface [26]. The second possibility, known as alternate test, is to predict the values of the individual performances by learning mappings of the form f2j : x pj , where pj is the jth performance, [27]. In this case, we learn n regression functions, where n is the number of performances. 2) Built-in test: Built-in test consists of adding auxiliary circuitry into the device with the aim to perform part of the test on-chip and, thereby, to provide simple digital, DC or low-frequency measurements to the test equipment or even just the pass/fail information. Built-in test facilitates the test of embedded blocks, enables parallel test, reduces the complexity of test instrumentation, and helps to diagnose the source of failure so as to provide valuable information for yield enhancement. One built-in test approach is to migrate some features of the test equipment into the circuit, i.e. build a miniature tester on-chip, in order to perform various curve tracing, oscilloscope, and spectrum analysis tasks. Several integrated test cores have been demonstrated aiming at characterizing the baseband frequency response [28], generating arbitrary band-limited waveforms [29], and digitizing arbitrary periodic analog waveforms [29]. Built-in test can also rely on reconguring the device into an easily testable form. For example, a loop-back connection can be established between the transmitter and the receiver chains of an RF transceiver, in order to use purely baseband test signals [30]. Another idea is to connect the device in a negative feedback loop during the test mode such that sustainable oscillations are produced at the output [31]. The amplitude and the frequency of the oscillation can be used to detect the presence of defects within the device. Another built-in test approach is to integrate sensors into the device with the aim to extract digital, DC or low-frequency test signatures that nevertheless carry higher frequency information. Several such sensors have been demonstrated for RF devices, including envelope detectors [32], current sensors [33], and process sensors [34]. Process sensors are particularly attractive because they are non-intrusive, i.e. they do not degrade the performance of the device, whereas the envelope detectors and current sensors necessarily require to be codesigned with the device. The process sensors are basic analog

stages that mimic part of the architecture of the device (i.e. bias stage, current mirrors, etc.) or basic layout components (i.e. transistor, capacitor, etc.). They are laid out in close proximity to the device and they monitor it by virtue of being subjected to the same process variations. In particular, any degradation in the performances of the device reects on the outputs of the sensors that shift away from their nominal values. 3) Adaptive test: Adaptive test is the dynamic adjustment of the test program (e.g. test limits, test content, and test ow) based on historical and real-time test data. In effect, the aim of adaptive test is to dene a decision-tree structure that in the extreme could result in each device being uniquely tested by a ow dictated by the observed response and by historical test data. Adaptive test consists in (a) eliminating specication tests on a lot-to-lot basis [35], (b) adapting the test limits based on the test results of the preceding device in order to obtain better quality control [36], (c) changing the order of specication tests on-the-y to move forward tests that have proven to detect many failing devices [37], (d) assessing on-the-y the condence of test decisions based on alternate measurements and, in case the test decision is deemed to be prone to error, forward the device to the standard specication test approach [38], (e) skip specication tests if their pass probability is beyond a condence level threshold [39]. R EFERENCES
[1] R. Rutenbar, G. Gielen, and J. Roychowdhury, Hierarchical modeling, optimization, and synthesis for system-level analog and RF designs, Proc. of the IEEE, vol. 95, no. 3, pp. 640 669, 2007. [2] S. Tiwary and R. Rutenbar, Faster, parametric trajectory-based macromodels via localized linear reductions, in Proc. IEEE/ACM Intern. Conf. On Comput.-Aided Des., Nov. 2006, pp. 876883. [3] B. Bond and L. Daniel, Stable reduced models for nonlinear descriptor systems through piecewise-linear approximation and projection, IEEE Trans. Comput.-Aided Des. Integr. Circuits. Syst., vol. 28, no. 10, pp. 14671480, Oct. 2009. [4] D. De Jonghe and G. Gielen, Compact trajectory-based behavioural models for analogue circuits, IEEE Electr. Lett., vol. 46, no. 15, pp. 10581060, 2010. [5] , Efcient analytical macromodeling of large analog circuits by transfer function trajectories, in Proc. IEEE/ACM Intern. Conf. On Comput.-Aided Des., 2011. [6] T. McConaghy, High-dimensional statistical modeling and analysis of custom integrated circuits, in IEEE Custom Integr. Circuits Conf., 2011. [7] B. Gustavsen, Improving the pole relocating properties of vector tting, in IEEE Trans. Power Del., vol. 21, no. 3, 2006. [8] J. Friedman, Multivariate adaptive regression splines, Annals Stat., vol. 19, no. 1, pp. 1 67, 1991. [9] J. Suykens and J. Vandewalle, Least squares support vector machine classiers, Neural Process. Lett., vol. 9, no. 3, pp. 293 300, 1999. [10] ITRS authors, International technology roadmap for semiconductors, ITRS, Tech. Rep., 2011. [Online]. Available: http://www.itrs.net/Links/2011ITRS/Home2011.htm [11] TSMC Public Relations, TSMC completes 28nm design infrastructure, design partners show solutions at DAC, 2011. [Online]. Available: http://www.prnewswire.com/news-releases/tsmccompletes-28nm-design-infrastructure-design-partners-show-solutionsat-dac-122641853.html [12] Solido Design Automation, Fast PVT Product, Variation Designer Platform, 2011. [Online]. Available: http://www.solidodesign.com [13] D. R. Jones, M. Schonlau, and W. J. Welch, Efcient global optimization of expensive black-box functions, J. of Global Optimization, vol. 13, pp. 455492, 1998. [14] A. Bowman, An alternative method of cross-validation for the smoothing of density estimates, Biometrika, vol. 71, pp. 353360, 1984.

[15] X. Wang and I. Sloan, Low discrepancy sequences in high dimensions: How well are their projections distributed? J. Comput. Appl. Math., vol. 213, no. 2, pp. 366386, Apr. 2008. [16] R. Kanj, R. Joshi, and S. Nassif, Mixture importance sampling and its application to the analysis of SRAM designs in the presence of rare failure events, in Proc. Des. Autom. Conf., 2006. [17] A. Singhee and R. Rutenbar, Statistical blockade: Very fast statistical simulation and modeling of rare circuit events and its application to memory design, IEEE Trans. Comput.-Aided Des. Integr. Circuits Syst., vol. 28, no. 8, pp. 11761189, Aug. 2009. [18] Solido Design Automation, HSMC Product, Variation Designer Platform, 2011. [Online]. Available: http://www.solidodesign.com [19] L. Fang, Y. Zhong, H. van de Donk, and Y. Xing, Implementation of defect oriented testing and ICCQ testing for industrial mixed-signal IC, in Proc. IEEE Asian Test Symp., Oct. 2007, pp. 404412. [20] G. C. Temes, Efcient methods of fault simulation, in Proc. IEEE Midwest Symp. On Circuits and Syst., 1977, pp. 191194. [21] Z. Yang and M. Zwolinski, Fast, robust DC and transient fault simulation for nonlinear analogue circuits, in Proc. Des. Autom. and Test in Europe, 1999, pp. 244248. [22] J. Hou and A. Chatterjee, Concurrent transient fault simulation for analog circuits, IEEE Trans. Comput.-Aided Des. Integr. Circuits Syst., vol. 22, no. 10, pp. 1385 1398, Oct. 2003. [23] H. Hashempour, J. Dohmen, B. Tasi , B. Kruseman, C. Hora, M. van c Beurden, and Y. Xing, Test time reduction in analogue/mixed-signal devices by defect oriented testing: An industrial example, in Proc. Des. Autom. and Test in Europe, 2011, pp. 16. [24] B. Kruseman, B. Tasi , C. Hora, J. Dohmen, H. Hashempour, M. van c Beurden, and Y. Xing, Defect oriented testing for analog/mixed-signal devices, in Proc. IEEE Intern. Test Conf., 2011. [25] F. Poehl, F. Demmerle, J. Alt, and H. Obermeir, Production test challenges for highly integrated mobile phone SOCs; a case study, in Proc. IEEE European Test Symp., 2010, pp. 1722. [26] H.-G. Stratigopoulos and Y. Makris, Error moderation in low-cost machine-learning-based analog/RF testing, IEEE Trans. Comput.-Aided Des. Integr. Circuits Syst., vol. 27, no. 2, Feb. 2008. [27] R. Voorakaranam, S. S. Akbay, S. Bhattacharya, S. Cherubal, and A. Chatterjee, Signature testing of analog and RF circuits: Algorithms and methodology, IEEE Trans. Circuits and Syst. I, May 2007. [28] A. Valdes-Garcia, F.-L. Hussien, J. Silva-Martinez, and E. SanchezSinencio, An integrated frequency response characterization system with a digital interface for analog testing, IEEE J. Solid-State Circuits, vol. 41, no. 10, pp. 23012313, Oct. 2006. [29] M. Hafed, N. Abaskharoun, and G. Roberts, A 4-GHz effective sample rate integrated test core for analog and mixed-signal circuits, IEEE J. Solid-State Circuits, vol. 37, no. 4, pp. 499514, Apr. 2002. [30] A. Valdes-Garcia, J. SiIva-Martinez, and E. Sanchez-Sinencio, Onchip testing techniques for RF wireless transceivers, IEEE Des. Test Comput., vol. 23, no. 4, pp. 268277, Apr. 2006. [31] G. Huertas, D. Vazquez, E. Peralias, A. Rueda, and J. Huertas, Testing mixed-signal cores: a practical oscillation-based test in an analog macrocell, IEEE Des. Test Comput., Nov./Dec. 2002. [32] Y.-C. Huang, H.-H. Hsieh, and L.-H. Lu, A build-in self-test technique for RF low-noise ampliers, IEEE Trans. Microw. Theory Tech., vol. 56, no. 5, pp. 1035 1042, May 2008. [33] M. Cimino, H. Lapuyade, Y. Deval, T. Taris, and J.-B. Begueret, Design of a 0.9 V 2.45 GHz self-testable and reliability-enhanced CMOS LNA, IEEE J. Solid-State Circuits, vol. 43, no. 5, May 2008. [34] L. Abdallah, H.-G. Stratigopoulos, S. Mir, and C. Kelma, RF front-end test using built-in sensors, IEEE Des. Test Comput., Nov./Dec. 2011. [35] S. Benner and O. Borofce, Optimal production test times through adaptive test programming, in Proc. IEEE Intern. Test Conf., 2001. [36] L. Fang, M. Lemnawar, and Y. Xing, Cost effective outliers screening with moving limits and correlation testing for analogue ICs, in Proc. IEEE Intern. Test Conf., Oct. 2006, pp. 1 10. [37] M. Chen and A. Orailoglu, Test cost minimization through adaptive test development, in Proc. IEEE Int. Conf. On Comput. Des., Oct. 2008. [38] H.-G. Stratigopoulos and S. Mir, Adaptive alternate analog test, IEEE Des. Test Comput., 2012 (to appear). [39] E. Yilmaz and S. Ozev, Adaptive test elimination for analog/RF circuits, in Proc. Des. Autom. Conf., Jul. 2009, pp. 720725.

Você também pode gostar