Circuit reliability prediction based on deep autoencoder network

Circuit reliability prediction based on deep autoencoder network

ARTICLE IN PRESS JID: NEUCOM [m5G;September 10, 2019;23:55] Neurocomputing xxx (xxxx) xxx Contents lists available at ScienceDirect Neurocomputin...

3MB Sizes 0 Downloads 73 Views

ARTICLE IN PRESS

JID: NEUCOM

[m5G;September 10, 2019;23:55]

Neurocomputing xxx (xxxx) xxx

Contents lists available at ScienceDirect

Neurocomputing journal homepage: www.elsevier.com/locate/neucom

Circuit reliability prediction based on deep autoencoder network Jie Xiao a, Weifeng Ma a, Jungang Lou b,∗, Jianhui Jiang c, Yujiao Huang a, Zhanhui Shi a, Qing Shen b, Xuhua Yang a a

College of Computer Science and Technology, Zhejiang University of Technology, Hangzhou, China College of Information Science, Huzhou University, Huzhou, China c Software Engineering Department, Tongji University, Shanghai, China b

a r t i c l e

i n f o

Article history: Received 1 March 2019 Revised 14 June 2019 Accepted 7 July 2019 Available online xxx Communicated by Dr. Ma Jiayi Keywords: Logic circuit Input vector Feature set Deep auto encoder network Reliability prediction

a b s t r a c t As semiconductor feature size continues to decrease and the density of integration continues to increase, highly reliable circuit design is experiencing many challenges, including reliability evaluation, which is one of the most important steps in circuit design. However, faced with the very large scale of integrated circuits at present, traditional simulation-based methods are slightly inadequate in terms of computational complexity and do not apply to the circuits at the concept stage. To solve this problem, this paper presents a new prediction method for circuit reliability based on deep auto encoder networks. Firstly, we analyze and extract the main features associated with circuit reliability. Next, we construct an efficient method for data collection by combining the characteristics of the feature set with the requirements of deep auto encoder networks. Then, we build a deep auto encoder network model oriented to circuit reliability prediction in a supervised learning manner. Simulation results on 74-series circuits and ISCAS85 benchmark circuits show that although the accuracy of the proposed method is slightly lower than that of both the Monte Carlo (MC) method and the fast probabilistic transfer matrix (F-PTM) model, its timespace consumption is approximately constant on different circuits, and it is 102,458,469 times faster than the MC method, and approximately 4,383 times faster than the F-PTM model. Furthermore, the proposed method could be used to predict circuit reliability at the conceptual stage, and it is a very efficient approximation method that could greatly reduce the power consumption of the calculation. © 2019 Published by Elsevier B.V.

1. Introduction With the extensive application of integrated circuit products in various fields, such as aviation, medical, and industry, high reliability is a requirement [1]. However, with the continued reduction in semiconductor feature size and the increase in integration density, uncertainty failure, such as intermittent and transient faults, has a more pronounced effect on circuit reliability [2]. Therefore, improving the reliability of circuits is a very important task that is directly related to the safety of people’s lives and property. Past analyses have found that reliability evaluation is one of the most important bases for improving circuit reliability [3]. Faced with the current very-large-scale integrated circuits, we need a quick and efficient evaluation method to measure their reliability at the early stages of circuit design so as to make timely decisions, thereby shortening the product development cycle to improve market



Corresponding author. E-mail addresses: [email protected] (J. Xiao), [email protected] (J. Lou), [email protected] (J. Jiang).

competitiveness. Circuit reliability refers to the probability that the circuit can correctly perform the specified tasks in a certain environment [2]. At present, the evaluation approaches for circuit reliability mainly include the measurement-based, the simulation-based and the analytical approach. There are two basic schemes in the measurement-based approach: sequential truncation tests, and timed or fixed-number truncation tests. Although the test results are very practical, the schemes usually require a large number of samples and have a large lag in the quality evaluation. The simulation-based approach models a fault randomly and injects the fault into the various gates by a given rule at a certain abstract level, and then counts the number of failures of the circuit to obtain its reliability. The typical model is the Monte Carlo (MC) method, which can be applied to the various abstract levels of circuits to achieve high accuracy but has time-consuming. For large-scale integrated circuits, the MC method is more commonly used as a verification method. The analytical approach is commonly implemented at a specified abstract level to evaluate the reliability of circuits in a multi-fault environment. The typical methods include the probabilistic transfer matrix model (PTM)

https://doi.org/10.1016/j.neucom.2019.07.100 0925-2312/© 2019 Published by Elsevier B.V.

Please cite this article as: J. Xiao, W. Ma and J. Lou et al., Circuit reliability prediction based on deep autoencoder network, Neurocomputing, https://doi.org/10.1016/j.neucom.2019.07.100

JID: NEUCOM 2

ARTICLE IN PRESS

[m5G;September 10, 2019;23:55]

J. Xiao, W. Ma and J. Lou et al. / Neurocomputing xxx (xxxx) xxx

[4], probabilistic gate model (PGM) [5], Bayesian networks method [6], and stochastic calculation method [7]. They tend to achieve high accuracy but often have large time-space consumption, and require that the object to be analyzed is an existing entity. In summary, the approaches mentioned above are not applicable to the circuits at the conception stage. However, with the acceleration of new technology applications and product revisions, it is necessary to have new technologies to fast evaluate the reliability of circuits at the conception stage to boost the speed of circuit structure design. Noticing that the deep learning model is a method for discovering distributed representations of data by simulating the neural centers of the human brain, and combining the features at a low level to form more abstract representations at a high level [8]. The technology of deep learning can achieve object classification or result prediction by combining classifiers or predictors, and shows approximately constant time-space consumption this meets the new requirements put forward by the existing evaluation methods of circuit reliability. Therefore, this technology is chosen in this study to quickly evaluate the reliability of integrated circuits at the conception stage. According to the label attributes of the training data, the deep learning model can be classified into supervised learning and unsupervised learning. In general, the former is more accurate than the latter because its input data is tagged, while the latter does not have tagged data [9]. Considering that the F-PTM model [10] can provide the reliability of the applied circuits to label the input features of the circuits, a deep learning network with supervised learning is chosen in this study to implement the follow-up research. At present, the research on deep learning, such as convolution neural networks (CNNs) [11], the deep belief networks (DBNs) [12] and deep autoencoder networks (DANs) [13], mainly focuses on the fields of machine vision, speech recognition, natural language processing, and traffic prediction, etc. Some studies have pointed out that the CNN mainly focuses on unstructured data, while the DBN and DAN perform well for both the structured and the unstructured data, especially the structured data [14,15]. Further studies have found that a DBN is more suitable for processing continuous data, while a DAN performs better when dealing with discrete data [16]. Analysis found that the reliability of a circuit is mainly depended on its topological structure, the failure probability of the basic gates, and its application environment (see Section 4.1 for the details). Obviously, the dataset constructed by the feature set has structured and discrete characteristics. Therefore, this paper chooses the DAN model to conduct research on the reliability prediction for circuit structures. The gates in this paper are assumed to be subjected to von-Neumann faults which consists of transient bit-flip faults caused by signal noise, although other stack at fault models can also be used in this paper. However, the structure of the DAN model is influenced by many factors, such as the characteristics, scale, and quality of the dataset, which directly determine the prediction accuracy of the model. Therefore, it is necessary to preprocess the dataset according to the characteristics of the DAN model and the dataset, and to configure the structure reasonably to obtain acceptable predictions, which requires a wealth of expertise and superb skills. Therefore, there is still a lot of work to do in using the DAN model to predict circuit reliability and achieving acceptable accuracy. Firstly, we select the features associated with circuit reliability to determine the feature set, and we construct an efficient method for data collection for the feature set. Next, we preprocess the dataset to meet the requirements of the DAN model. Then, we train and test the constructed DAN model with the constructed training set and the testing set to find a reasonable structure for the DAN model that meets the accuracy requirements. Finally, we predict the reliability of the circuits according to their applied input features.

The rest of the paper is organized as follows. The preparatory work related to this study is introduced in Section 2, Section 3 describes the extraction of labeled data by the F-PTM model, followed by circuit feature analysis and extraction in Section 4. The prediction method for circuit reliability based on the DAN model is presented in Section 5. The experimental results are presented in Section 6, followed by the conclusions in Section 7. 2. Preparatory work To successfully conduct this study, the preparatory work involved is the description of the principles and construction method for the DAN model. Since the feature dataset in this paper has labeled attributes, and supervised learning is superior to unsupervised learning in terms of classification and prediction accuracy, we choose the DAN model with supervised learning to predict circuit reliability. The structure and the construction of the DAN model are presented in Figs. 1 and 2, respectively. It can be seen from Fig. 1 that the DAN is a neural network model with multiple hidden layers that consists of multiple autoencoders (AEs), and each AE is a learning network with only one input layer and one hidden layer, where the output of the AE at the bottom is used as the input to the AE at the top. Based on the representation in the previous layer, each layer in the DAN model learns multiple layer-based representations from the raw data, so as to learn more essential features [17]. It can be seen from Fig. 2 that the construction of the DAN model mainly includes the following parts. (a) Data collection and preprocessing. Firstly, we collect the feature data according to the characteristics of the DAN model. Next, we preprocess the feature dataset to meet the requirements of the DAN model. Then, we divide the dataset into training set and testing set by the previously determined proportions. (b) Construct a DAN-based prediction model. Firstly, we construct a feature learning network based on AEs. Next, we initialize the weight W and the bias vector b for the deep feedforward neural network by the feature learning network with layer-wise training. Then, we fine-tune the DAN model with supervised learning with the labeled training dataset. (c) Model testing and selection. We test and evaluate the DAN model with the test dataset and the error evaluation function, respectively, to determine whether the constructed model meets the accuracy requirements, so as to select an efficient prediction model. 2.1. AE The role of the AE is to recover the raw data from the abstracted data as much as possible, so as to extract the distributed feature representation from the raw input data. The AE is a special three-layer back propagation (BP) neural network consisting of an input layer, an encoder, and a decoder, as shown in Fig. 3, where V=(v1 ,v2 , .,vin ) is the feature dataset, vi (i = 1, 2, …, in) is the ith input feature dimension vector, and xi is the ith dimension input signal for the input layer. In the AE, w(1) and b(1) are the shared weight matrix and the bias vector between the input layer and the hidden layer, respectively, and w(2) and b(2) are the shared weight matrix and bias vector between the hidden layer and the output layer, respectively. The "+1" node is the bias node, and hw,b (x) is the output signal. To extract the essential features from the input data, the trained steps of the AE can be summarized simply, as follows. Firstly, we perform encoding transformation on the input signal x using Eqs. (1) and (2) to obtain another represent ation h of x. Next, we decode h using Eqs. (3) and (4) to obtain the output signal hw,b (x). Then, we update the shared weight matrix w and the bias vector

Please cite this article as: J. Xiao, W. Ma and J. Lou et al., Circuit reliability prediction based on deep autoencoder network, Neurocomputing, https://doi.org/10.1016/j.neucom.2019.07.100

ARTICLE IN PRESS

JID: NEUCOM

[m5G;September 10, 2019;23:55]

J. Xiao, W. Ma and J. Lou et al. / Neurocomputing xxx (xxxx) xxx

3

Fig. 1. The structure of the DAN model.

Fig. 2. Construct an AE-based layer-wise feature-learning network.





h = a (2 ) = f z (2 ) ,

(2)

z(3 ) = w(2 ) a(2 ) + b(2 ) ,

(3)





hw,b (x ) = a(3 ) − f z(3 ) , J AE(w,b) =





(4)



Lw,b a(3 ) , x ,

(5)

x∈V

f (z ) =

1 , 1 + exp (−z )

(6)

2.2. Small batch gradient descent method Fig. 3. Typical AE structure.

b in the AE using the cost function presented in Eq. (5) and the BP algorithm. We repeat the above steps until we find the w and b that minimize Eq. (6), (hw,b (x)≈x), where z(2) is the input to the hidden layer; a(2) is the encoding result in the AE (denoted as the output of hidden layer); a(3) is the input to the output layer, which is the decoding result in AE; f is the sigmoid function (nonlinear activation function) presented in Eq. (6); and L is the reconstruction error function, which is represented by the mean square error function in this study.

z(2) = w(1) x + b(1) ,

(1)

To optimize the parameter θ ={w,b}, the solution is usually achieved by the gradient descent method in error back propagation. However, for a large training dataset, the learning process is slow because the gradient vector J(θ ) is calculated once for each training instance. Therefore, the mini-batch gradient descent (MBGD) is introduced in this study. Unlike the gradient descent algorithm (where all the training data is used to update the parameters in each training), MBGD decomposes one training into T iterations, and each iteration is performed by randomly selecting N samples (mini batch) from the training dataset to calculate the gradient of the loss function, so as to update the parameters. To boost the convergence, the latest updated items are added to the

Please cite this article as: J. Xiao, W. Ma and J. Lou et al., Circuit reliability prediction based on deep autoencoder network, Neurocomputing, https://doi.org/10.1016/j.neucom.2019.07.100

ARTICLE IN PRESS

JID: NEUCOM 4

[m5G;September 10, 2019;23:55]

J. Xiao, W. Ma and J. Lou et al. / Neurocomputing xxx (xxxx) xxx

Fig. 4. Training flowchart of the DAN model.

updating parameters in the momentum algorithm, so that the parameters can be updated in time, where the e-th training can be expressed by Eqs. (7)–(9). Jm (θ ) represents the gradient vector in the loss function, and is calculated by using the m-th batch of training data, m∈{1,2,…,T}; S is the number of samples in the m-th batch of the dataset; θ m (e) is the parameter calculated in the m-th iteration in the e-th training; Gm (e) is the update parameter calculated in the first m iterations in the e-th training, G0 (0) = 0; γ is the momentum super-parameter (its value is usually less than or equal to 0.9), and η is the learning rate.

θm+1 (e ) = θm (e ) − Gm (e ),

(7)

Gm (e ) = γ Gm−1 (e ) + η∇ Jm (θ (e ) ),

(8)

∇ Jm (θ ) =

S 

ji (θ ),

(9)

i=1

2.3. Model training The training of the DAN model is described in Fig. 4, and its main steps are as follows: (a) Construct k AEs, and initialize their relevant parameters in a uniformly distributed manner within the interval β shown in Eq. (10), and initialize h0 = V. (b) Initialize the input signal of the i th AE (denoted as AE_i) with hi -1 (i∈{1, 2, ., k}), train AE_i and implement Eqs. (1)–(9). Then encode hi -1 with the trained AE_i to obtain the output signal hi to the hidden layer. (c) If i