Accès libre

The Comprehensive Diagnostic Method Combining Rough Sets and Evidence Theory

À propos de cet article

Citez

Introduction

Information fusion is applied widely during fault diagnosis and there are many ways. Evidence theory or D-S theory (DST) [1, 2] can compound the uncertain information from several information sources and is a very effective uncertainty reasoning way in information fusion technology. However, the evidence theory itself has some problems [3], such as dependency on the evidence provided by expert knowledge, focus element explosion caused by evidence combination, rigorous demands to combination conditions (mutual independence among evidences), inefficiency to evidential conflicts and the subjectivity in the distribution of credibility and so on. Many scholars have improved the theory to overcome these limitations, but still there are some deficiencies [4,5,6]. Rough set theory is a new mathematical tool to deal with vagueness and uncertain knowledge and is very practical. It can extract unrelated essential characteristics to eliminate redundancies without any initial or additional data [7, 8]. Both the rough set and evidence theories are math tools for uncertain information and they focus on the group ability. There is a strong complementary relationship between these two theories [9,10,11]. This paper provides a new research thought for intelligent fault diagnosis by combining the evidence theory with rough set theory to solve the subjectivity, relativity, integrated focus element explosion and conflict problems of evidences, improve the accuracy of fault diagnosis and at last carry out online diagnostics.

Basic Theories
Rough Set Theory

The paper just gives two definitions about the importance of attribute and the basic ideas about rough set theory, which are referred in the literature [12].

Definition 1

The importance of any attribute a ∈ (CR) about D is defined as: SGF(a,R,D)=γ(R{a},D)γ(R,D) SGF(a,R,D) = \gamma(R\cup\{a\},D)-\gamma(R,D)

Definition 2

To BC, if B is independent to D, and also γ(C,D) = γ(B,D), then B is a D relative reduction of C.

D-S Evidence Theory

Here the paper just gives the D-S evidence theory synthesis formula and the related knowledge, which are referred in the literature [1, 2].

Definition 3

If Bel1,Bel2,...,Beln is the belief function on the same identification framework and Θ and m1,m2,...,mn are the basic probability assignments that correspond to them, then the basic probability assignment after synthesis is: m(A)={0A=φAn=An=1Nmn(An)/Anφn=1Nmn(An)Aφ \begin{equation}m(A) =\begin{cases}0 & \quad A=\varphi\\\sum\limits_{\cap A_{n}=A}\prod\limits_{n=1}^{N} m_{n}(A_{n}) / \sum\limits_{\cap A_{n} \neq \varphi}\prod\limits_{n=1}^{N} m_{n}(A_{n}) & \quad \forall A \neq \varphi\end{cases}\end{equation}

Among them: k=An=φn=1Nmn(An)AΘ k = \sum\limits_{\cap A_{n}=\varphi}\prod\limits_{n=1}^{N} m_{n}(A_{n}) \forall A \subset \Theta reflects the degree of conflict of evidences. Definition 3 is the combination rule of D-S theory (DST).

Improvement of D-S Evidence Synthesis Rule

From the D-S evidence theory synthesis formula: when k = 1, it shows that evidences conflict completely and fusion cannot be done; when 0 < k < 1, synthesis formulas can be applied, while k → 1 and evidences conflict highly, the synthesis formula will lead the perverse conclusion. Many scholars have made some significant achievements about evidence conflict probability distribution function; Sun Quan proposed a new synthesis formula by analyzing the advantages and disadvantages of evidence synthesis rules of DST, and Yager performed the synthesis of conflicts in a better way at different degrees. The synthesis formula is as follows [13]: m(φ)=0;m(A)=An=An=1Nmn(An)+f(A)Aφ;m(Θ)=An=Θn=1Nmn(An)+f(Θ) m(\varphi) = 0 ; m(A) = \sum\limits_{\cap A_{n}=A}\prod\limits_{n=1}^{N} m_{n}(A_{n}) + f(A)~~~~\forall A \neq \varphi ; m(\Theta) = \sum\limits_{\cap A_{n}=\Theta}\prod\limits_{n=1}^{N} m_{n}(A_{n}) + f(\Theta) f(A)=kεq(A)(q(A)=1n1inmi(A)) f(A) = k\varepsilon q(A) (q(A)=\frac{1}{n}\sum\limits_{1 \leq i \leq n} m_{i}(A)) f(Θ)=kεq(Θ)+k(1ε) f(\Theta) = k\epsilon q(\Theta) + k(1 - \varepsilon)

Among them: n is the number of evidences; the reliability of evidence ɛ is ε=ek˜(k˜=1n(n1)/2i<jkij i,jn,kij=AiAj=Ωmi(Ai)mj(Aj)) \varepsilon = e^{-\tilde{k}} (\tilde{k} = \frac{1}{n(n-1)/2}\sum\limits_{i<j} k_{ij}~~~~i, j \leq n, k_{ij} = \sum\limits_{A_{i} \cap A_{j}=\Omega} m_{i}(A_{i})m_{j}(A_{j}))

Eq. (4) has some subjective factors, while the physical meaning of Eqs (2) and (3) is not clear. Since q(A) is the average supporting degree of evidence to A, evidence conflict probability k can be assigned in proportion to A. That is weighing and assigning the probability of evidential conflict according to the supporting degree of each proposition; also because of the differences of importance degree of each evidence source, the evidence weight factor is introduced. In this paper, a new evidence synthesis way is proposed: m(φ)=0;m(A)=An=An=1Nmn(An)+kq(A)Aφ;q(A)=1inwimi(A) m(\varphi) = 0 ; m(A) = \sum\limits_{\cap A_{n}=A}\prod\limits_{n=1}^{N} m_{n}(A_{n}) + kq'(A)~~~~\forall A \neq \varphi ; q'(A) = \sum\limits_{1 \leq i \leq n} w_{i} m_{i} (A)

Among them: k=An=φn=1Nmn(An)AΘ k = \sum\limits_{\cap A_{n}=\varphi}\prod\limits_{n=1}^{N} m_{n}(A_{n}) \forall A \subset \Theta ; wi is the weight of No.i evidence source, and this paper gives the self-conflict of evidence to determine the evidence weight coefficient.

Definition 4

Suppose the conflict sum of evidence i and evidence j(j = 1,2,...,i−1,i+1,...,n) is the amount of self-conflict ϕi of evidence i, the expression is ϕi=j=1,jinAiAj=φmi(Ai)mj(Aj) \phi_{i} = \sum\limits_{j=1, j \neq i}^{n}~~\sum\limits_{A_{i} \cap A_{j}=\varphi} m_{i}(A_{i})m_{j}(A_{j}) mi(Ai)mj(Aj) (n is the number of evidence), then: wi=1ϕi/j=1n1ϕji=1,2,,n w_{i} = \frac{1}{\phi_{i}} / \sum\limits_{j=1}^{n} \frac{1}{\phi_{j}}~~~i=1,2,\ldots,n .

Definition 4 shows the greater the amount of self-conflict for evidence, the smaller its weight factor is. Otherwise, the smaller the amount of self-conflict for evidence, the greater its weight factor is. The weight factor shows the important degree of evidences provided by the information source in the synthesis process and the incidence of the synthetic results.

Here is an example to prove the validity of the way. Suppose Θ = {A,B,C}, three evidences are as follows:

m1 : m1(A) = 0.98, m1(B) = 0.01, m1(C) = 0.01;

m2 : m2(A) = 0, m2(B) = 0.01, m2(C) = 0.99;

m3 : m3(A) = 0.9, m3(B) = 0, m3(C) = 0.1.

Three evidences are synthesised by applying the synthetic methods of DST, Yager [14], Sun Quan, Li Bicheng [15] and the synthesis method proposed in this paper, respectively (evidence weight w1 = 0.37414, w2 = 0.21902, w3 = 0.40684) and the results are shown in Table 1. Table 1 shows the evidence synthetic method with the weight factor raised in this paper that can enhance the reasonability and reliability in the evidence combination. It can avoid the subjectivity and randomness from the weight factor by determining the evidence weight coefficient with the amount of evidence self-conflict, reflect the importance of evidence during synthesis and get a better result than other methods.

Comparison of results of various evidence synthesis methods.

Synthesis formula k m(A) m(B) m(C) m(Θ)
DST 0.99901 0 0 1 0
Yager 0.99901 0 0 0.00099 0.99901
Sun Quan 0.99901 0.3210 0.0030 0.1880 0.4880
Li Bicheng 0.99901 0.6260 0.0067 0.3673 0
This paper 0.99901 0.7321 0.0059 0.2620 0

Application of the attribute reduction of the rough set can effectively optimise the key features as the evidence for diagnosis decisions. Attributes significance of rough set can evaluate the weight of each evidence objectively. And then there comes the formula wi=SGF(i,R,D)/jCSGF(j,R,D) w_{i} = SGF(i,R,D) /\sum\limits_{j \in C} SGF(j,R,D) .

Inference method of rough set and evidence theories

This paper creates a fault diagnosis model based on inference strategy to fuse rough set and evidence theories, as shown in Figure 1. First, the basic thought is building an information decision table by discretising the fault sample data sets with continuous attributes. Then, optimising feature parameters suitable for fault diagnosis as theory body by applying the rough set theory for attribute reduction of the decision table. In practice, for the discretised sample set to diagnose, basic probability assignment of related evidences can be calculated based on the reduction and the diagnosis result can be obtained with the inductive decision by the use of the D-S combination rule.

Fig. 1

Structure of diagnostic model integrated with evidence theory and rough set.

Discretisation of continuous data sample

A rough set theory can deal with discretised data, while original sample data are always continuous, so continuous data should be discretised first. There are many ways to discretise and each has its advantage. In practice, each field seeks a proper algorithm according to the characteristic [16, 17].

Symptom attribute reduction

Symptom attribute reduction can reduce the relativity among evidence and decide the fault with attributes as less as possible, remaining the sorting quality invariant and avoiding focus element explosion. On the other hand, the weight of each attribute can be obtained from the information in the decision table and the subjectivity from experts can be avoided, overcoming the difficulties in practical application efficiently caused by the subjectivity and relativity of evidence during the fault diagnosis with evidence theory. About the reduction algorithm of the rough set theory, still there is no recognised and high-efficient algorithm. With the completion of the reduction algorithm considered, attribute reduction is applied in this paper by combining discernibility matrix, dependability of attribute and the heuristic reduction algorithm of information entropy by improving the importance of attribute. The detailed process of this algorithm is referred in the literature [18, 19].

Basic probability assignment of evidence

Basic probability assignment of evidence is realised with the decision attribute D in the decision table as the recognition framework Θ of evidence theory and with all the condition attribute r(rR) after reduction in the decision table as the corresponding evidence. Many scholars have researched the relation between the reliability calculations of evidence theory and rough set theory [10, 11]. Since initial or additional data are not necessary for rough set theory, the basic probability assignment on rough set theory is more objective.

Theorem 1

If XU, equivalence relation U/R = X1,X2,...,Xn, then the measure of lower approximation of Xq(X) = |apr(X)|/|U|,XU (apr(.) is the lower approximation) is a belief function, the basic probability assignment is: m(xi) = |Xi|/|U|,i = 1,2,...,n; m(A) = 0,A / ∉ U/R.

Theorem 2

If Bel is a belief function meeting the needs of the following condition: the focus element of Bel is equivalence class of U; the basic probability assignment of the focus element of Bel A is: m(A) = |A|/|U|, then there exists a Pawlak rough algebra meeting with q(X) = Bel(X),XU.

The proof of Theorems 1 and 2 can be referred in the literature [10]. From Theorems 1 and 2, it is possible to calculate belief function based on a decision table with rough set theory. It provides the theoretical basis for the fusion reasoning of evidence theory and rough set theory.

The basic probability assignment m(A)=BA(1)|AB|Bel(B) m(A) = \sum\limits_{B \subseteq A} (-1)^{\lvert A - B \rvert} Bel(B) can be calculated by defining the belief function Bel(A)=BAm(B)(AΘ) Bel(A) = \sum\limits_{B \subset A} m(B)~~~~ (\forall A \subset \Theta) . From Theorem 1, the belief function is equivalent with a lower approximation. Considering the inconvenience of this expression in practical application, this paper provides a simple algorism with the potential of set: m(A)={|A|/|U|(AU/P)0(AU/P) \begin{equation}m(A) =\begin{cases}\lvert A \rvert / \lvert U \rvert & \quad (A \in U / P)\\0 & \quad (A \notin U / P)\end{cases}\end{equation}

Here P is indiscernibility relation. It is clear that Eq. (5) meets with Theorems 1 and 2.

Next, a practical algorism of basic probability assignment is given based on a rough set decision table: {Input: decision table after reduction; samples to be diagnosed}; {Output: basic probability assignments of all evidences}.

Quantification of samples to be diagnosed: to discretise the samples to be diagnosed according to the discretisation criterion of original sample data;

Recognition framework Θ is decision attribute set and the reasoning evidence is condition attributes r(rR);

To determine the division U/Θ of Θ to U;

To determine the division U/R of R to U with evidence ri as equivalence relation R;

To determine equivalence class in U/R and to determine each intersection of the equivalence and U/Θ according to the discrete value of the samples to be diagnosed in evidence ri;

To get the basic probability assignment of evidence ri to from Eq. (5).

The discount of evidence and reasoning decision

If there is only 1 − α(α ∈ [0,1]) confidence to the whole evidence, is considered as the discount rate. Considering discount to evidence, the calculation of basic probability assignment is as follows: mα(A)=(1α)m(A)AΘA0;mα(Θ)=(1α)m(Θ)+α m^{\alpha}(A)=(1-\alpha)m(A) ~~\forall A \subset \Theta~~ A \neq {\emptyset} ; m^{\alpha}(\Theta) = (1-\alpha)m(\Theta) + \alpha

The diagnosis can be concluded based on the decision of basic probability assignment after a combination of the D-S rule [20].

Diagnosis examples

The engine is an important equipment of a ship, concerning the survivability and battle effectiveness of a ship. The performance of the engine can influence and restrict the performance of the technique and tactics of a ship. The structure of the engine is complex, while the fault analysis to it is difficult. To ensure the accuracy of fault diagnosis, more characteristic parameters should be used. The number of characteristics to be extracted gets larger with the increase of various kinds of parameters and this makes the amount of information to be processed too large and cannot meet the needs of online diagnosis. Only a few key characteristics are sensitive to a fault. They are independent and provide complementary information for each other to improve the accuracy of diagnosis. While the redundant characteristics are not sensible to fault or they are related to other characteristics but useless. The rough set theory can eliminate redundant information effectively, select the key characteristics, get probability assignment and make reasoning decisions with D-S combination rule to solve the problems such as subjectivity in evidence obtaining, relativity of evidence and focus element explosion of evidence combination.

The extraction of evidence information

After the experiments of five fault phenomena appearing in a kind of ship engine and the management to the diagnostic knowledge, this paper selects 10 sets of data samples to build. The common fault information is shown in Figure 2. In Figure 2, U stands for n (n=10) states of the engine, C={cooling water temperature C1, airflow C2, fuel pressure C3, revolution speed C4, torque C5} stands for the five characteristic (symptom) parameters to describe the state of equipment, D={normal D1, high-temperature D2, airflow meter damage D3, fuel injector fault D4, ignition fault D5} shows the fault symptoms to each state of the equipment. Select three sets of data randomly as the test samples (to be diagnosed) to verify the diagnostic effects from the method proposed in this paper.

The attribute value of the sample data in Table 2 is continuous and needs to be discretised. Here a discretisation for continuous attribute value based on SOFM network classification is used and the details and the computing process of this algorithm can be referred in the literature [21]. The decision table after discretisation is shown in Table 3. After attribute reduction with the method proposed in this paper, only three condition attributes, C1, C2, C5, are kept in the decision table. They are the diagnostic proofs for five fault symptoms: cooling water temperature (E1), air flow (E2), torque (E3).

Table of fault information.

U m(C1) m(C2) m(C3) m(C4) m(C5) D
Original samples 1 0.95 1.40 1.20 5100 0.39 D1
2 0.70 1.74 0.94 3450 0.98 D1
4 0.06 1.73 0.98 3900 0.91 D2
5 0.65 3.31 0.63 1950 0.55 D3
7 0.30 1.31 1.32 4350 0.40 D4
8 0.28 1.23 1.11 4300 0.44 D4
10 0.05 2.47 1.01 3800 0.20 D5

Test samples 3 1.02 1.73 0.57 3500 0.98 D1
6 0.67 3.71 0.77 1900 0.45 D3
9 0.18 0.95 1.14 3600 0.30 D5

Decision table after discretization.

U C1 C2 C3 C4 C5 D
Original samples 1 3 1 3 3 2 D1
2 2 2 2 2 3 D1
4 1 2 2 2 3 D2
5 2 3 1 1 2 D3
7 2 1 3 3 2 D4
8 1 1 2 3 2 D4
10 1 2 2 2 1 D5

Test samples 3 3 2 1 2 3 verify D1
6 2 3 1 1 2 verify D3
9 1 1 2 2 1 verify D5

From Definition 1, the importance of attribute C1, C2, C5 about D can be obtained: SGF(C1,R,D)=γ(R{C1},D)γ(R,D)=5/7SGF(C2,R,D)=SGF(C3,R,D)=2/7 SGF(C_{1},R,D) = \gamma(R\cup\{C_{1}\},D)-\gamma(R,D) = 5/7~~~~SGF(C_{2},R,D) = SGF(C_{3},R,D) = 2/7

It shows C1 is the attribute that influences the accuracy mostly to diagnostic decision rule, followed by C2 and C5. This agrees with the engineering practice. Then the weight of each evidence Ei (i=1,2,3) can be obtained from the following equation wi=SGF(i,R,D)/jCSGF(j,R,D)=(ω1,ω2,ω3)=(0.56,0.22,0.22) w_{i} = SGF(i,R,D) / \sum\limits_{j \in C}SGF(j,R,D) = (\omega_{1}, \omega_{2}, \omega_{3}) = (0.56,0.22,0.22)

Basic probability assignment and D-S evidence theory

For evidence E1, E2, E3, the basic probability assignment can be calculated with α=0.05 as the discount rate. In Table 3, the attribute values of test sample 6 and original sample 5 are the same after discretisation, so the diagnostic decision is the same necessarily and D3 need not be verified anymore. For the two sets of test samples, D1 and D5 are fused with D-S evidence theory and the combining method by improving evidence proposed in the paper, the basic probability assignments are shown in Tables 4 and 5. Θ is the overall uncertainty.

Basic probability assignment about D1.

Verify D1 K m(D1) m(D2) m(D3) m(D4) m(D5) m(Θ)
E1 0 0.9500 0 0 0 0 0.05
E2 0 0.3167 0.3167 0 0 0.3167 0.05
E3 0 0.4750 0.4750 0 0 0 0.05
E1E2 DS 0.9144 0.0397 0 0 0.0397 0.0062
Improve DS 0.6017 0.8269 0.0703 0 0 0.0703 0.0325
E1E3 DS 0.9522 0.0433 0 0 0 0.0045
Improve DS 0.4512 0.8899 0.0850 0 0 0 0.0251
E2E3 DS 0.4771 0.4771 0 0 0.0397 0.0062
Improve DS 0.6017 0.4282 0.4282 0 0 0.1111 0.0325
E1E2E3 DS 0.9488 0.0469 0 0 0.0039 0.0006
Improve DS 0.7972 0.7534 0.1498 0 0 0.0569 0.0399

Basic probability assignment about D5.

Verify D1 K m(D1) m(D2) m(D3) m(D4) m(D5) m(Θ)
E1 0 0 0.3167 0 0.3167 0.3167 0.05
E2 0 0.3167 0 0 0.6334 0 0.05
E3 0 0 0 0 0 0.9500 0.05
E1E2 DS 0.0531 0.0531 0 0.8326 0.0531 0.0081
Improve DS 0.7020 0.0794 0.1746 0 0.5339 0.1746 0.0375
E1E3 DS 0 0.0398 0 0.0398 0.9145 0.0060
Improve DS 0.6017 0 0.1520 0 0.1520 0.6636 0.0324
E2E3 DS 0.1624 0 0 0.3248 0.4872 0.0256
Improve DS 0.9025 0.1587 0 0 0.3175 0.4762 0.0476
E1E2E3 DS 0.0245 0.0245 0 0.3840 0.5634 0.0036
Improve DS 0.9677 0.0688 0.1707 0 0.3191 0.3932 0.0482
The analysis and explanation of diagnosis

From Tables 4 and 5, the case cannot be diagnosed and a diagnostic error will occur if the diagnosis is with single evidence. For example, E2 and E3 in Figure 4 cannot be diagnosed, E2 in Table 5 is a diagnostic error. The diagnostic capability of single evidence is limited. So it is necessary to combine all the evidence to get a better diagnosis. The diagnostic capability can be improved because the overall uncertainty reduces in comparison with a single evidence diagnosis after two evidences are combined. But it is possible for the case that cannot be diagnosed or diagnostic errors because of the limitation of self-capability of single evidence and original training samples, for example, E2E3 in Table 4 cannot be diagnosed, and E1E2 in Table 5 is a diagnostic error. For some evidences, the diagnosis is not certain after fusion, the uncertainty is reduced gradually. After the fusion of all evidences E1, E2, E3, certainty strengthens. With one evidence or two evidences, the case cannot be diagnosed or diagnostic errors may occur, while with three evidences fused the correct diagnosis can be obtained finally.

It shows that the diagnostic capability can be improved by reducing the uncertainty efficiently by fusion reasoning for all the evidences that are not redundant.

The fusion results of the method proposed in this text are not as good as those from D-S evidence theory, according to Tables 4 and 5. Because the former is based on solving the fusion of high conflict evidences (k →1) efficiently and it affirms the openness of the recognition framework. The latter can get a good synthetic effect for normal evidences and think the current recognition framework is complete and close. The synthetic result shows that the method with weight factor considered is also effective for solving the synthesis of normal evidences.

Conclusion

An integrated diagnostic method with evidence theory and rough set fusion reasoning is put forward and it can overcome the problems of subjectivity, dependency and focus element explosion of traditional evidences. It has a significant theoretical significance and practical value.

This paper proposed a synthetic method by considering the weight factor. The method can solve the synthetic problems of the evidences with high conflict and is effective for synthesising normal evidences.

The application effect of an integrated diagnosis of rough set and evidence theories is proved to be good with diagnosis examples. With the increase of original sample data, the accuracy of diagnosis will improve. The method proposed in this text can be promoted and applied in the fault diagnosis to other devices if the sample data are large enough.

The research of this paper will be further deepened. It will focus on developing artificial intelligence diagnostic system for better application in industrial practice to achieve maximum production benefits in the direction of our subsequent research.

eISSN:
2444-8656
Langue:
Anglais
Périodicité:
Volume Open
Sujets de la revue:
Life Sciences, other, Mathematics, Applied Mathematics, General Mathematics, Physics