HOME JOURNALS CONTACT

Information Technology Journal

Year: 2014 | Volume: 13 | Issue: 6 | Page No.: 1140-1146
DOI: 10.3923/itj.2014.1140.1146
Acquisition and Reduction of Incomplete Decision Information System Based on Similarity and Dissimilarity Knowledge Granules
Weiyan Xu, Min Wei, Ming Zhang and Xibei Yang

Abstract: In the incomplete decision information systems, based on similarity relation, the concept of similarity and dissimilarity knowledge is firstly proposed. By the basic knowledge granule, two kinds of rough set model are defined and their properties are discussed. In these rough set models, based on similarity and dissimilarity knowledge, an approach is presented to acquire the positive and negative decision rules, respectively. The certainty factor is introduced to measure their certainty degree. Finally, to simplify these decision rules, this study introduces the heuristic algorithm of the attribute reduction based on the significant attributes and analyzes an illustrative example to prove its effectiveness.

Fulltext PDF Fulltext HTML

How to cite this article
Weiyan Xu, Min Wei, Ming Zhang and Xibei Yang, 2014. Acquisition and Reduction of Incomplete Decision Information System Based on Similarity and Dissimilarity Knowledge Granules. Information Technology Journal, 13: 1140-1146.

Keywords: heuristic algorithm, Similarity relation, knowledge acquisition, decision rule and attribute reduction

INTRODUCTION

Rough set theory presented by Pawlak (1982) and Pawlak and Skowron (2007) is an effective mathematical method to analyze and deal with the imprecise, inconsistent and vague information. With over 30 year’s development, it has been widely used in the fields such as pattern recognition, machine learning, decision analysis, knowledge acquisition and data mining (Qian et al., 2010; Hu et al., 2007; Sen and Pal, 2009; Xu et al., 2012a; Liang et al., 2012; Gu et al., 2011) and so on.

The classical rough set theory is based on the complete information systems and there are no unknown values. However, the Incomplete Information Systems (IIS) (Yang et al., 2009a; Grzymala-Busse, 2004; Stefanowski and Tsoukias, 2001) can be found everywhere for a lot of unpredictable reasons. Therefore, mining rules from incomplete information systems is one of the important directions for the development of rough set. Most of decision rules are positive property (Pawlak, 1982; Xu et al., 2012b; Gu et al., 2011; Yang et al., 2009b; Grzymala-Busse, 2004; Stefanowski and Tsoukias, 2001), but we also need some form of negative rules, it means that if an object doesn’t satisfy the certain attribute-value pairs in the condition part, then we can exclude such object from the decision class, which was firstly proposed and successfully applied in the medical diagnosis expert system by Tsumoto (2000). Yao (2010, 2011) solved the acquisition of three-way decisions rule, such as the positive, negative and obtained decision rules based on the model of probabilistic rough sets, but it’s a pity that he didn’t pay attention to the reduction. To acquire the negative decision rule in IIS, Yang et al. (2009a) put forward the difference relationship based on rough set and Xu et al. (2012a) proposed the variable precision rough set model based on negative support set of descriptor, but their attribute reduction algorithm based on discernibility matrix, was not conducive to automatic realization by computer.

In this study, the notion of similarity and dissimilarity knowledge granule based on similarity relation in incomplete decision information system, is introduced firstly. The lower and upper approximation sets based on similarity and dissimilarity knowledge granule, are constructed to acquire the positive and negative decision rules from the incomplete decision information system. Secondly, to obtain a simplified decision rule, the heuristic algorithm of the attribute reduction based on the significant attributes is proposed. The final results of example analysis show that this algorithm is feasible and practical.

INCOMPLETE DECISION INFORMATION SYSTEM

A Decision Information System (DIS) is defined as a 4-tuple DIS = <U, AT, V, f>, where U is a non-empty finite set of objects, AT = C∪D is a non-empty finite set of attributes, C denotes the set of condition attributes and D denotes the set of decision attributes and C∩D = ø. Each attribute α∈AT is associated with a set Va of its values, Va is called the domain of α and then V = VC∪VD. For any x∈U, f (x,α) is the value that x holds on α (α∈AT). If D = {d}, DIS = <U, C∪{d}, V, f> is a single attribute decision information system, otherwise it is called a multiple attribute decision information system. To simplify our discussion, we only discuss the single attribute decision information system.

An Incomplete Decision Information System (IDIS) is a decision information system IDIS = <U, C∪{d}, V, f>, where the condition-attribute values for some objects are unknown. In this study, we consider the unknown value of the condition attributes as missing (Grzymala-Busse, 2004), which is denoted by the special symbol “*”. For instance, if f (x, α) = ‘>*” (x∈ U, α∈AT), here we assume that “*” only may appear in the values of C, thus, V = VC∪Vd.∪{*}.

DECISION RULES

In the rough set theory, the knowledge hidden in IDIS = <U, C∪{d}, V, f>, will be extracted by the form of decision rules. Through the investigation to the object x∈U, we can get the decision rule such as:

(1)

where, des([x]C) is the condition part of the rule, it shows the description of object x under the condition-attribute set C, i.e., des ([x]C) =∧ c∈C(c, vc) and vc ≠ *; des ([x]d) is the decision part of the rule, it shows the description of object x under the decision attribute d, i.e., des([x]d) = ∨i∈Vd(d, i), here i∈Vd is the label of the class. On the basic of des([x]C)→des([x]d), we can conclude that the object x belongs to some certain decision class, therefore, des ([x]C)→des([x]d) is called a positive decision rule.

For each positive decision rule rx, the certainty factor Cer (rx) is defined to measure it’s decision ability:

(2)

where, card (X) is the cardinal number of the set X, ||[x]C|| is the set of those elements having the same description as x on the condition attribute set C, denoted as ||[x]C|| = {y∈U: f (y, c) = f (x, c), ∀c∈C} and ||[x]d|| is the set of those elements having the same description as x on the decision attribute d, it is recorded as ||[x]d|| = {y∈U: f (y, d) = f (x, d)}.

If Cer (rx) = 1, then, rx is a definite positive decision rule; if 0<Cer (rx)<1, rx is an indefinite positive decision rule. The degree of certainty is measured by the value of Cer (rx).

In addition to the above positive decision rule, we always need such decision rule called negative decision rule as, if an object’s condition attribute-value pairs don’t satisfy some certain condition in the condition part, then we can conclude that the object x is not belong to the certain decision class, in other words, we can exclude it from some decision class. The negative rule can be described as:

(3)

Accordingly, the general form of the negative decision can be expressed as:

(4)

where, des(¬[x]C) is the condition part of the negative decision rule and shows the description of the object x under the condition-attribute set C, that is to say des (¬[x]C) = ∧c∈C¬(c, vc) and vc ≠ *; correspondingly des (¬[x]d) is the decision part of the negative decision rule and shows the description of the object x under the decision attribute d, i.e., des¬ ([x]d) = ∨I∈Vd¬ (d, i), here I∈Vd is the label of the class. Similarly, for each negative decision rule, the certainty factor can be denoted by:

(5)

where, ||¬[x]C|| is the set of those elements with the same description as x on the condition attribute set C, recorded as ||¬[x]C|| = {y∈U: f (y, c) ≠ f (x, c), ∀c∈C} and ||¬[x]d|| is the set of those elements having the same description as x on the decision attribute d, recorded as ||¬[x]d|| = {y∈U: f (y, d) ≠ f (x, d)}. If Ncer (rx) = 1, then, rx is a definite negative decision rule; if 0<NCer (rx)<1, then, rx is an indefinite negative decision rule. The degree of certainty is measured by the value of Ncer (rx).

RULE ACQUISITION

For a lot of unpredictable reasons in real word, the study is mostly in incomplete decision information system. Therefore, the study mainly discussed how to acquire the positive and negative decision rules in incomplete decision information system.

Similarity knowledge and positive rule
Definition 1:
Let IDIS = <U, C∪{d},V, f>, for any A⊆C, S (A) is a similarity relation on the universe U, decided by the attribute subset A and define as:

(6)

Obviously, the binary relation S (A) is reflexivity and transitivity, but not necessarily symmetric. Let SIMA (x) = {y∈U: (x, y)∈S (A)} be the set of those elements similar to x, regarded SIMA (x) as the basic knowledge granule (i.e., similar knowledge granule), the lower and upper approximations of X can be constructed to acquire the positive decision rule.

Definition 2: Suppose IDIS = <U,C∪{d},V, f>, for any X⊆U, A⊆C, the lower and upper approximations of X with regard to A, can be defined with SIMA (x) as the basic knowledge granule:

(7)

(8)

Based on the similar knowledge granule SIMA (x), the ordered pair <, > is definded as the rough set of X with regard to A. Through the lower and upper approximation of X, the positive, boundary and negative region can be, respectively described as follows:

(9)

(10)

(11)

Theorem 1: Suppose IDIS = <U, C∪{d}, V, f>, U/d = {D1, D2,…,Di} is the partition of U induced by decision attribute d, Di = {x∈U: f (x, d) = i} and i is the label of the class. Then, for any x∈U, we have:

If SIMA (x)⊆POSSIM (Di), then, rx: des ([x]C)→ des ([x]d) is a definite positive decision rule
If SIMA (x)⊆BNDSIM(Di), then rx: des ([x]C)→ des ([x]d) is an indefinite positive decision rule

Proof 1: Since, SIMA (x)⊆POSDIM(Di), for any x∈SIMA (x), then SIMA (x)≠ø and SIMA (x)⊆Di. By the definition of SIMA (x), the elements whose description is same as x in X, have the same decision attribute-value i, i∈Vd. So, Cer (rx) = 1 sets up, then, rx is a definite positive decision rule.

Proof 2: The proof is similar to proof 1.

Dissimilarity knowledge and negative rule: Through the similarity knowledge induced by a similarity relation, we can obtain the positive decision rule, but can’t get the negative decision rule. Below is the acquisition of the negative rules.

Definition 3: Suppose IDIS = <U,C∪{d},V, f>, for any A⊆C, S (A) is a similarity relation on the universe U determined by the condition attribute subset A, DIMA (x) is the set of those elements dissimilar to x and defined as:

To acquire the negative decision rules, the following discussion studies how to construct the lower and upper approximations of X, based on DIMA (x) as the basic knowledge granule (that is the dissimilar knowledge granule).

Definition 4: Suppose IDIS = <U,C∪{d},V, f>, for any X⊆U, A⊆C, the lower and upper approximations of X with regard to A, can be definded with DIMA (x) as the basic knowledge granule:

(12)

(13)

Based on the dissimilar knowledge granule DIMA (x), the ordered pair (, is defined as the rough set of X with regard to A. Through the lower and upper approximations of X, the positive, boundary and negative region can be respectively described as follows:

(14)

(15)

(16)

Theorem 2: Let IDIS = <U, C∪{d}, V, f>, U/¬d = {¬D1, ¬D2,…,¬Di} is the cover of U induced by the decision attribute d, ¬Di = {x∈U: f (x,d) ≠ i} and i is the label of the class. Then, for any x∈U, we have:

If DIMA (x)⊆POSDIM (¬Di), then, rx: des (¬[x]C)→ des (¬[x]d) is a definite negative decision rule
If DIMA (x)⊆BNDDIM (¬Di), then, rx: des (¬[x]C)→ des (¬[x]d) is an indefinite negative decision rule

Proof 1: Since, DIMA (x)⊆POSDIM (¬Di), for any x∈DIMA (x), then, DIMA (x) ≠ ø and DIMA (x)⊆¬Di. By the definition of DIMA (x), the elements whose description is different from x in X, all don’t take the decision-value i, i∈Vd. So, Ncer (rx) = 1 sets up, then, rx is a definite negative decision rule.

Proof 2: The proof is similar to proof 1.

REDUCTION

Reduction is also called attribute reduction or feature selection and it is the key issue to study the rough set theory [3-4,6,9]. Through the attribute reduction, the simplified decision rule can be obtained. In this section, the heuristic algorithm based on the importance is introduced to simplify the positive and negative decision rules.

Definition 5: Suppose IDIS = <U,C∪{d},V, f>, A⊆C, D = {D1,D2,…,Dr} is the partition of U induced by the decision attribute d, where Di = {x∈U: f (x,d) = i} and ¬Di = {x∈U: f (x,d)≠i}. we define:


If LSIM (A) = LSIM (C), then, A is called a lower approximation distribution consistent set about similar knowledge granule. And if any proper subset of A is not a lower approximation distribution consistent set, then we called that A is the lower approximation distribution reduction about the similar knowledge granule
If USIM (A) = USIM (C), then A is called an upper approximation distribution consistent set about similar knowledge granule. And if any proper subset of A is not an upper approximation distribution consistent set, then we called that A is the upper approximation distribution reduction about the similar knowledge granule
If LSIM (A) = LSIM (C) and USIM (A) = USIM(C), then, A is called a distribution consistent set about the similar knowledge granule. And if any proper subset of A is not a distribution consistent set about similar knowledge granule, then we called that A is the distribution reduction about the similar knowledge granule
If LDIM (A) = LDIM (C), then, A is called a lower approximation distribution consistent set about dissimilar knowledge granule. And if any proper subset f A is not a lower approximation distribution consistent set, then, we called that A is the lower approximation distribution reduction about the dissimilar knowledge granule
If UDIM (A) = UDIM (C), then, A is called an upper approximation distribution consistent set about dissimilar knowledge granule. And if any proper subset of A is not an upper approximation distribution consistent set, then we called that A is the upper approximation distribution reduction about the dissimilar knowledge granule
If LDIM (A) = LDIM (C) and UDIM (A) = UDIM (C), then, A is called a distribution consistent set about the dissimilar knowledge granule. And if any proper subset of A is not a distribution consistent set about dissimilar knowledge granule, then we called that A is the distribution reduction about the dissimilar knowledge granule

How to obtain the distribution reduction based on similar and dissimilar knowledge granule, it is focused on how to look for all of the lower and upper approximation distribution reductions of the decision classes. The heuristic algorithm of the attribute reduction based on the significant attributes is introduced in the following discussion.

Suppose DIS = <U, C∪{d},V, f>, D = {D1,D2,…,Dr} is the partition of U induced by decision attribute d, where Di = {x∈U: f (x,d) = i} and ¬Di = {x∈U: f (x,d) ≠ i}. For any condition attribute α∈C , based on the similar knowledge granule and dissimilar knowledge granule, it’s dependence degrees on the lower and upper approximation of decision class D are, respectively expressed as follows:

Obviously, Sa (SIM,D), Sa (SIM,D), Sa (DIM,D) and Sa (DIM,D) are all greater than or equal to zero.

If Sa (SIM,D) = 0 (or Sa (SIM,D) = 0), after eliminating attribute α, we can conclude that the lower (or upper) approximation distribution of decision class D based on the similar knowledge granule doesn’t change.

If Sa (SIM,D)>0 (or Sa (SIM,D)>0), after eliminating attribute α, we can find that the lower (or upper) approximation distribution of decision class D based on the similar knowledge granule will be changed, then the attribute α can’t be reducted.

If Sa (DIM,D) = 0 (or Sa (DIM,D) = 0), after eliminating attribute α, we can conclude that the lower (or upper) approximation distribution of decision class D based on the dissimilar knowledge granule doesn’t change.

If Sa (DIM,D)>0 (or Sa (DIM,D)>0), after eliminating attribute α, we can find that the lower (or upper) approximation distribution of decision class D based on the dissimilar knowledge granule will be changed, then the attribute α can’t be reducted.

Based on the definition 5, to obtain the distribution reduction of similar and dissimilar knowledge, the key is finding out their lower and upper approximation distribution reduction. Below is a heuristic algorithm of the attribute reduction based on the significant attributes.

Algorithm 1: The lower/upper approximation distribution reduction about the similar knowledge.

Suppose IDIS = <U, C∪{d}, V, f>, C = {α1, α2,…,αi} is the set of attributes, D = {D1,D2,…,Dr} is the partition of U induced by decision attribute d. The steps to obtain the lower approximation distribution reduction of C which is denoted by R, can be described as follows:

Step 1: Suppose R = ø
Step 2: If LSIM (R) = LSIM (C), then, return to Step 4, otherwise return to Step 3
Step 3: For any αi∈(C-R), calculate Sai (SIM,D), when find the minimal value Sai (SIM,D), get ai and replace R = R∪{αi}, then, return to Step 2
Step 4: Export R. R is a lower approximation distribution reduction of C about the similar knowledge

Note: If the judgment condition LSIM (R) = LSIM (C) in step 2 is replaced by USIM (R) = USIM (C) and the Sai (SIM,D) is replaced by Sai (SIM,D), then we can get R, where R is an upper approximation distribution reduction of condition attribute set C about the similar knowledge.

Algorithm 2: The lower/upper approximation distribution reduction about the dissimilar knowledge.

Suppose IDIS = <U,C∪{d}V,f>, C = {α1, α2,…,αi} is the set of attributes, U/¬d = {¬D1, ¬D2,…, ¬Di} is the cover of U induced by decision attribute d and ¬Di = {x∈U: f (x,d) ≠ i}. The lower approximation distribution reduction of condition attribute set C, is denoted by R. The steps of acquiring Reduction R are as follows:

Step 1: Suppose R=ø£+
Step 2: If LDIM (R) = LDIM (C), then return to Step 4, otherwise return to Step 3
Step 3: For any αi∈(C-R), calculate Sai (DIM,D), when find the minimal value Sai (DIM,D), get αi and replace R = R∪{αi}, then, return to Step 2
Step 4: Export R. R is a lower approximation distribution reduction of condition attribute set C about the dissimilar knowledge

Note: If the judgment condition LDIM (R) = LDIM (C) in step 2 is replaced by UDIM (R) = UDIM (C) and Sai (DIM,D) becomes Sai (DIM,D), then we can get R, where R is an upper approximation distribution reduction of condition attribute set C about the dissimilar knowledge.

Example: Table 1 is an IDIS, compute all of the definite positive and negative decision rules.

By Table 1, we know that U = {x1, x2, x3, x4, x5, x6 ,x7} is the domain of discourse, C = {a, b, c} is the set of condition attributes, d is the single decision attribute and Va = Vb = Vc = Vd = {1, 2, 3}. By definition 1, we can draw the following sets:

By the single decision attribute d, U/d = {D1, D2 D3} = {{x1, x7},{x3, x4},{ x2, x5, x6}}

By definition 2,

According to the heuristic algorithm 1, we can consider {a, b} as a reduction of the condition-attribute set C through the program. Combining theorem 1, we can obtain the following definite positive decision rule:

r1: f (x, b) = 1→f (x, d) = 1, // x1, x7∈POSSIM (¬D1)
r2: f (x,a) = 2→ f (x, d) = 2, // x3 ∈POSSIM (¬D2)
r3: f (x,a) = ∧f (x, b) = 2→f (x, d) = 3
r4: f(x,a) = 3∧f (x, b) = 3→f (x, d) = 3
// x1, x7∈POSSIM(¬D3)

Table 1: Incomplete decision information system

By the Table 1, U/¬d = {¬D1, ¬D2, ¬D3} = {{x2,x3, x4, x5, x6}, {x1 x2, x5,x6,x7}, {x1, x3, x4, x7}}, By definition 3, we can draw the following sets:

By definition 4, we have:

According to the heuristic algorithm of lower approximation distribution reduction, we can consider {a, c} as a reduction of the condition-attribute set C through the program. Further, we can obtain the following definite negative decision rule:

CONCLUSION

In this study, the notion of similarity and dissimilarity knowledge granule based on the similarity relation is firstly introduced in IDIS. The lower and upper approximation sets based on similarity and dissimilarity knowledge granule, are constructed to acquire the positive and negative decision rules. By the properties of the new rough set model, we discuss the acquisition of the positive and negative decision rules, respectively. Finally, to obtain a simplified decision rule, we further introduce the heuristic algorithm of the attribute reduction based on the significant attributes and an example shows its effectiveness and feasibility.

ACKNOWLEDGMENTS

This study is supported by the Natural Science Foundation of China (No. 61100116), Natural Science Foundation of Jiangsu Province of China (No. BK2011492), Natural Science Foundation of Jiangsu Higher Education Institutions of China (No. 11KJB520004).

REFERENCES

  • Pawlak, Z., 1982. Rough sets. Int. J. Comput. Inform. Sci., 11: 341-356.
    CrossRef    Direct Link    


  • Pawlak, Z. and A. Skowron, 2007. Rudiments of rough sets. Inform. Sci., 177: 3-27.
    CrossRef    Direct Link    


  • Qian, Y., J. Liang, W. Pedrycz and C. Dang, 2010. Positive approximation: An accelerator for attribute reduction in rough set theory. Artif. Intell., 174: 597-618.
    CrossRef    


  • Hu, Q., Z. Xie and D. Yu, 2007. Hybrid attribute reduction based on a novel fuzzy-rough model and information granulation. Pattern Recognit., 40: 3509-3521.
    CrossRef    


  • Sen, D. and S.K. Pal, 2009. Histogram thresholding using fuzzy and rough measures of association error. IEEE Trans. Image Process., 18: 879-888.
    CrossRef    


  • Xu, W., Y. Li and X. Liao, 2012. Approaches to attribute reductions based on rough set and matrix computation in inconsistent ordered information systems. Knowledge-Based Syst., 27: 78-91.
    CrossRef    


  • Liang, J., F. Wang, C. Dang and Y. Qian, 2012. An efficient rough feature selection algorithm with a multi-granulation view. Int. J. Approximate Reasoning, 53: 912-926.
    CrossRef    


  • Gu, S.M., W.Z. Wu and Y. Zheng, 2011. Rule acquisition in consistent multi-scale decision systems. Proceedings of the 8th International Conference on Fuzzy Systems and Knowledge Discovery, Volume 1, July 26-28, 2011, Shanghai, China, pp: 386-389.


  • Yang, X., J. Xie, X. Song and J. Yang, 2009. Credible rules in incomplete decision system based on descriptors. Knowledge-Based Syst., 22: 8-17.
    CrossRef    


  • Grzymala-Busse, J.W., 2004. Data with Missing Attribute Values: Generalization of Indiscernibility Relation and Rule Induction. In: Transactions on Rough Sets I, Peters, J.F., A. Skowron, J.W. Grzymala-Busse, B. Kostek, R.W. Swiniarski and M.S. Szczuka (Eds.). Springer, New York, USA., ISBN-13: 9783540223740, pp: 78-95


  • Stefanowski, J. and A. Tsoukias, 2001. Incomplete information tables and rough classification. Comput. Intell., 17: 545-566.
    CrossRef    Direct Link    


  • Tsumoto, S., 2000. Automated discovery of positive and negative knowledge in clinical databases. IEEE Eng. Med. Biol. Mag., 19: 56-62.
    CrossRef    PubMed    Direct Link    


  • Yao, Y., 2010. Three-way decisions with probabilistic rough sets. Inform. Sci., 180: 341-353.
    CrossRef    


  • Yao, Y., 2011. The superiority of three-way decisions in probabilistic rough set models. Inform. Sci., 181: 1080-1096.
    CrossRef    


  • Yang, X.B., D.J. Yu, J.Y. Yang and X. Song, 2009. Difference relation-based rough set and negative rules in incomplete information system. Int. J. Uncertainty Fuzziness Knowledge-Based Syst., 17: 649-665.
    CrossRef    


  • Xu, W., J. Jiang and M. Zhang, 2012. Two types of decision rules based on variable precision rough set. Int. J. Adv. Comput. Technol., 14: 270-278.

  • © Science Alert. All Rights Reserved