Incremental Knowledge Compilation Using the ... - Semantic Scholar

Report 2 Downloads 219 Views
JOURNAL OF NETWORKS, VOL. 8, NO. 10, OCTOBER 2013

2241

Incremental Knowledge Compilation Using the Extension Rule Ying Liu 1, 4, Wenxiang Gu*, 2, 4, Tingting Zou 3, Qingtao Huang 1, and Qiao Li 1 1. College of Humanities & Science of Northeast Normal University, Changchun, 130117, China 2. Department of Basic Subjects Teaching, Changchun Architecture & Civil Engineering College, Changchun, 130607, China 3. College of Computer Science and Technology, Jilin University, Changchun 130012, China 4. School of Computer Science and information technology, Northeast Normal University, Changchun, 130117, China *Corresponding author, Email: [email protected]

Abstract—Knowledge compilation by using the extension rule (KCER) has been recognized as a novel compilation approach, although this method can only deal with static knowledge bases. This paper proposes a novel incremental knowledge compilation method by using the extension rule so that it can tackle knowledge compilation problems in the dynamic environment. The method does not recompile the whole knowledge base, while it is to use the information of compiling original knowledge base to speed up the compiling process. The experimental results show that this method can deal with dynamic knowledge bases efficiently. Index Terms—Dynamic Knowledge Base, Extension Rule, EPCCL, Incremental Knowledge Compilation

I.

INTRODUCTION

Automatic reasoning is one of the important research fields of artificial intelligence, and its reasoning complexity is usually NP-complete. Knowledge compilation [1] is effective techniques that have been proposed for addressing such computational difficulties. It is used to network engineering [2], logical reasoning and other fields [3, 4, 5]. The central observation is that in Knowledge representation reasoning problems, the input to a problem is conceptually split in two parts: a knowledge base and a query. Typically, the knowledge base is not modified frequently, and the same knowledge base is applied to answer many queries. The idea of knowledge compilation is to split query answering into two phases: In the first one the knowledge base is preprocessed, thus obtaining an appropriate data structure (such a phase is sometimes called off-line reasoning); In the second phase, the query is actually answered using the output of the first phase (such a phase is sometimes called on-line reasoning). The key is that knowledge compilation needs to be done only once to be accessible for different queries. Hence, the compiling time can be amortized by many queries concerning the compiled knowledge base. In 1987, Reiter and de Kleer [6] firstly put forward the concept of knowledge compilation, namely, compiling the knowledge base into the equivalent of the new one that can be tractable reasoning, and propose a knowledge compiled algorithm by using the prime implicate for © 2013 ACADEMY PUBLISHER doi:10.4304/jnw.8.10.2241-2246

target language. But the knowledge base obtained from this method is not complete for unit resolution. In 1994 Alvaro del Val [7] improved above method, and put forward a new knowledge compilation approach to ensure the completeness of unit resolution. In 1995, Marquis [8] improved Reiter and de Kleer’s method and proposed the knowledge compilation approach based on a theory prime implicate. All of these methods are based on the prime implicate. In 1996, Schrag [9] put forward a knowledge compilation approach by using prime implicant for the target language. The above methods can find the knowledge base which is equivalent to the original knowledge base by knowledge compilation, and then reasoning it, but not all of the knowledge bases exist the equivalent ones, so Selman and Kautz [10, 11] proposed an approximate knowledge compilation approach in 1996, which calculated least upper bound (LUB) and greatest lower bound (GLB) of the knowledge base, and found the Horn knowledge base which is approximate to the original knowledge base, then reasoning, but it didn’t answer all the queries of the original knowledge base. At present, there are many target languages for knowledge compilation, such as the binary decision diagram[12], decomposable negation normal form (DNNF), which supported the target language for polynomial time logical operations, put forward by Darwiche in 1999 [13]. Very resently, Christian Muise1 et al. proposed a new method, DSHARP: Fast d-DNNF Compilation with sharpSAT [14], it is very efficient. Knowledge compilation has become an efficient method for dealing with intractable reasoning problems for NP-complete or co-NP complete problems. Therefore, many researchers have focused their study on knowledge compilation for description logic. Literature [15] proposed a novel knowledge compilation method for description logic, which aims to further improve the knowledge compilation method based on extension rule, so that it can deal with description logic problems. They defined concept extension rule and EPCCCL theory, and then present an algorithm compiling any concept into equivalence EPCCCL theory, and prove that

2242

satisfiabiliy-testing and subsumption-testing can be done in linear time in the size of the compiled concept. Above knowledge compilation methods can only deal with the reasoning problems in static knowledge base, but knowledge base is usually dynamic. For example, when the knowledge base needs to add new clauses, the new knowledge base cannot ensure to keep the properties of tractable reasoning, so it must be recompiled. For those applications which need to knowledge base updated frequently, the above methods need greater time cost. In order to solve these problems, in 1993, Goran Gogic put forward the method for incremental knowledge compilation [16]. For a knowledge base, it gets the approximate LUB and GLB by using the horn approximation. When knowledge base adds a new clause, using the model updating method, it doesn’t recompile the entire knowledge base, but it needs to update the least horn upper bound (Horn-LUB) and greatest horn lower bound (Horn-GLB) of knowledge base and then gets the updated LUB and GLB. Linhai [17] proposed a new method of automated theorem proving, namely extension rule, first it calculated all of maximum terms in the set of clauses, then judged the satisfiability by sets of clauses. This effective method used the including-excluding principle to reduce space complexity of judging satisfiability. Later, they expanded the extension rules to compile knowledge [18, 19], and put forward a new target language EPCCL and an algorithm which can compile arbitrary knowledge base into EPCCL language. Knowledge compilation approach by using the extension rule (KCER) is based on the extension rule whether compilation phase or extension reasoning phase. On the other hand, we applied the extension rule method to the problem of counting model solution, and put forward an algorithm of counting model using extension rule [20, 21]. The purpose of our paper is to promote the knowledge compilation approach based on extension rule, and make it deal with the knowledge compilation problems in dynamic knowledge base. We put forward an algorithm of incremental knowledge compilation by using the extension rule (IKCER), which can update the dynamic knowledge base. It doesn’t attempt to recompile the updated knowledge base, but uses the information which compiled original knowledge base to compile the compiled knowledge base and updated clauses. The experimental results show that the IKCER is able to deal with the update of dynamic knowledge base effectively. Compared with the KCER, our method takes less time, and increases less clauses. This paper is organized as follows. In section 2, we introduce some related concepts and knowledge compilation based on extension rule. In section 3, we put forward the approach of incremental knowledge compilation by using the extension rule, and its algorithm. Section 4 shows the experimental results data through the concrete experimental results. Finally, section 5 summarizes our work.

© 2013 ACADEMY PUBLISHER

JOURNAL OF NETWORKS, VOL. 8, NO. 10, OCTOBER 2013

II.

BASIC KNOWLEDGE

First, we agree with some symbols in this paper. We use ∑ to denote a knowledge base, namely a set of clauses in conjunctive normal form (CNF). C, D... denote the clauses, a, b... denote the atom, a, ¬a... denote the literal, and M denotes the set of all atoms that appear in knowledge base. Definition 2. 1 [17, 22] Given a non-tautology clause C and a set M, C is the maximum item on a set M, iff it contains all atoms in M in either positive form or negative form. Definition 2. 2 [17, 22] Given a clause C and a set M, D  C  a, C  a a  M , a and a do not appear in C , then we call the operation proceeding from C to D as

the extension rule on C. We call D is the result of the extension rule. Theorem 2. 1 [17, 22] Given a clause C and its results D, which is applied extension rule, then a clause C is logically equivalent to D. Theorem 2. 2 [17, 22] Given a knowledge base ∑, and its set of all atoms M ( | M | m ). If all the clauses in ∑ are all maximum terms on M, then ∑ is unsatisfiable, iff m it contains 2 clauses which are different from each other. According to the above theorems, a simple way to judge the satisfiability of knowledge base is to count the m number of maximum terms, if the number is equal to 2 , then the knowledge base is unsatisfiable, otherwise is satisfiable. Therefore literature [10] realized this method using including-excluding principle. Given a knowledge base   {C1 , C2 , , Cn } , and its set of all atoms M ( | M | m ). Let Pi be the set of all maximum terms we can get from Ci by using extension rule, i=1, 2, , n. Let S be the number of distinct maximum terms we can get from ∑ by using the extension rule. Then by using the including-excluding principle, we will have: n

S   Pi -  i 1

1 i j  n

Pi  Pj 

- (1)n 1 P1  P2  where, Pi  2



1 i j l  n

Pi  Pj  Pl

 Pn

m  Ci

For general knowledge base, it needs to calculate 2n1 items of the formula (1). However, when any two clauses all contain the pairs that contain complementary literal(s) in the knowledge base, it only needs to calculate the former n items of formula (1), and reduces the space complexity. Therefore, the literature [18] puts forward EPCCL theory. Definition 2. 3[17, 22] An EPCCL theory is a set of clauses in which each pair (of clauses) contains complementary literal(s). Literature [18] proved EPCCL theory is in the tractable satisfiability class and tractable entailment class. Therefore, EPCCL theory can be used as target language of knowledge compilation. The basic process of the

JOURNAL OF NETWORKS, VOL. 8, NO. 10, OCTOBER 2013

knowledge compilation by using extension rule (KCER) [18] is that: given a proposition knowledge base, we firstly extend a clause by using extension rule, make it and other clauses contain pairs that contain complementary literal(s), and then add the extended clauses into the compiled knowledge base, in this cycle, until all clauses have been processed in original knowledge base. The compiled knowledge base is equivalent to the original knowledge base, and it is an EPCCL theory. III.

INCREMENTAL KNOWLEDGE COMPILATION USING THE EXTENSION RULE

Given a proposition knowledge base ∑, and the compiled knowledge base ∑1 is equivalent to it. The main idea of incremental knowledge compilation is that, when the original knowledge base ∑ add a new clause C, it does not recompile the updated knowledge base ∑∪{C}, but use the compiled information in the knowledge base ∑ʹ which is equivalent to the original knowledge base, and speed up the compilation of the updated knowledge base ∑∪{C}, then get the equivalent knowledge base ∑ʺ. We use the extension rule to the incremental knowledge compilation, and propose a new approach of incremental knowledge compilation by using the extension rule. Given a proposition knowledge base ∑, we choose the EPCCL as the target language of knowledge compilation, and compile the knowledge base ∑, so we can get a knowledge base ∑1 which is equivalent to it, where any two clauses contain complementary literals. If any clauses are compiled by using extension rule in the knowledge base ∑1, the compiled knowledge base is still EPCCL theory. According to the properties, when adding one or more than one clauses to knowledge base, we compile the clauses in knowledge base ∑1 by using extension rule, make them and clause C contain complementary literals, so that we can get a new EPCCL knowledge base ∑2, it is equivalent to the updated knowledge base ∑∪{C}. Now, we give the algorithm for Incremental Knowledge Compilation using Extension Rule (IKCER). First Chosen one clause Ci in knowledge base ∑1, and added it into the knowledge base ∑2, then judged the relationship between Ci and C, deleted a clause or extended Ci by using extension rule, it is ensure that any clauses in ∑2 contain the pairs that contain complementary literal(s) with clause C. Repeating the process until it has handled all the clauses in ∑1. At last, it merges C into ∑3. Using solved the information in original knowledge base, and the method accelerates the compilation for the updated knowledge base, and improves the efficiency of the compilation. Algorithm Ikcer Input: Knowledge base ∑, compiled knowledge base ∑1={C1, C2, …, Cn}, ∑2=∑3=, updated clause C. Output:∑3 is the result of knowledge compilation process While ∑1≠ Loop Select a clause Ci from ∑1, i=1, 2, …, n, ∑2=∑2∪{Ci}; © 2013 ACADEMY PUBLISHER

2243

If Ci and C contain pairs that contain complementary literal(s) Then do-nothing; Else if Ci entails C Then ∑3=∑3∪∑2∪∑1, Return ∑3; Else if C entails Ci Then ∑2=; Else Extend Ci by using extension rule; ∑3=∑3∪∑2, ∑2= ; Endloop ∑3=∑3∪{C}; Return ∑3; Let’s give an example for the ALGORITHM IKCER. Given a knowledge base ∑ and a compiled knowledge base 1  {{a b c d e} {b d} {c d f } {a d f }} , a new clause C  {b c e} , then the process of IKCER as follows: 1. 1  {{a b c d e} {b d} {c d f } {a d f }} ,  2  3   ;

2. 1  {{b d} {c d f } {a d f }} , 2  {{a b c d e}} , 3 :  ;

3. 1  {{b d} {c d f } {a d f }} , 2   , 3   ; 4. 1  {{c d f } {a d f }} , 2  {b d } , 3   ; 5. 1  {{c d f } {a d f }} , 2   , 3  {b d } ; 6. 1  {a d f } , 2  {c d f } , 3  {b d } ; 7. 1  {a d f } , 2  {{b c d f } {b c d e f }} , 3  {b d} ;

8. 1  {a d f } , 2   , 3  {{b d}{b c d f } {b c d e f }} ;

9. 1   , 2  {a d f } , 3  {{b d}{b c d f } {b c d e f }}

10. 1   ,

2  {{a b d f }{a b c d f } {a b c d e f }}

,

11. 3  {{b d}{b c d f } {b c d e f }} ; 12. 1   , 2   , 13.

3  {{b d}{b c d f } {b c d e f } {a b d f }{a b c d f } {a b c d e f }}

;

3  {{b d }{b c d f } {b c d e f }

14. {a b d f }{a b c d f } {a b c d e f }{b c e}}

Theorem 3. 1 Given a knowledge base ∑, a compiled knowledge base ∑1, the updated clause C, and the output results of IKCER ∑3, then ∑3 is equivalent to ∑∪{C}, and ∑3 is an EPCCL theory. Proof. Firstly we need to prove ∑3 equivalent to ∑∪ {C}. In the algorithm of IKCER, rules of the extension and elimination can guarantee equivalence of knowledge base, so ∑3 is equivalent to ∑1∪{C}. While ∑1 is the result of compiled ∑, therefore ∑1 is equivalent to ∑, ∑3 is equivalent to ∑∪{C}. Then we prove ∑3 is an EPCCL theory. In algorithm IKCER, an arbitrary clauses in ∑1 contain the pairs that contain complementary literal(s), after moving a clause Ci into ∑2, extended Ci to make it contains the pairs that contain complementary literal(s) with clause C, now all clauses of expanded Ci still

2244

JOURNAL OF NETWORKS, VOL. 8, NO. 10, OCTOBER 2013

include the pairs with the clauses in ∑1, so before moving the clauses in ∑2 into ∑3, any clauses in ∑2 contain the pairs with clause C, the clauses in ∑2 also contain the pairs with the clauses in ∑1. Although the clauses in ∑1 may be further extended by executing the algorithm, the clauses in ∑3 contain the pairs with all of the extended clauses in ∑1, and with clause C. So, any clauses in the result of IKCER ∑3 contain the pairs that contain complementary literal(s), ∑3 is an EPCCL theory. Theorem 3. 2 The algorithm IKCER is effective and complete. Proof. According to the theorem 3. 1, the algorithm IKCER returns the knowledge base which is an EPCCL theory, so the algorithm is effective. Given a knowledge base ∑ and a clause C, if there is an EPCCL knowledge base ∑ʺ, it is equivalent to ∑∪{C}, the algorithm finally returns ∑ʺ as the result. Therefore the algorithm is complete. Knowledge compilation approach compiles the original knowledge base into the tractable reasoning one, such as Horn knowledge base and EPCCL knowledge base. Then the reasoning problem can be finished in polynomial time, and greatly reduce the time complexity. IV.

update knowledge base, add a new clause, compile the updated knowledge base, and calculate the increased number of clauses and time in knowledge base. Repeat the process until no new clauses add into the knowledge base, and put all of the number of increased clauses and time together finally, as the number of increased clauses and increased time for the random problem.

Figure 2. Random 3-SAT problem, comparing the number of increased clauses of IKCER with KCER on random 3-SAT problem, N=20, M=40.

EXPERIMENTAL RESULTS

In this section, we compare the algorithm IKCER with the original algorithm KCER [18] on random SAT instances. For the update problems of dynamic knowledge base, we test a great number of examples, the efficiency of IKCER is much better than KCER.

Figure 3. Random 4-SAT problem, comparing the number of increased clauses of IKCER with KCER on random 4-SAT problem, N=20, M=40.

Figure 1. Random 2-SAT problem, comparing the number of increased clauses of IKCER with KCER on random 2-SAT problem, N=20, M=40.

The test cases are obtained by a random generator in the experiment. They are < 20, 40, 2 >, < 20, 40, 3 >, < 20, 40, 4 >. The first parameter denotes the number of variables N in knowledge base, the second denotes the number of clauses M, and the third denotes the maximum length of clauses K. Each clause is obtained at random by choosing K variables from the range of the N variables, and each variable’s probability of negative or positive is equal to 0. 5. For each type of difficulty for test cases, it randomly generates 20 instances, and solves them respectively, finally uses the average value as a result. For each instance because of random problem, firstly we compile the knowledge base which contains one clause of original problem, and record the number of clauses and compilation time in compiled knowledge base. Then we © 2013 ACADEMY PUBLISHER

We compare the number of increasing clauses for ICKER and KCER. From the figure 1 to figure 3, we can see that the number of increase clauses for IKCER is less than KCER’s. Usually, KCER’s number of increase clauses is 10-50 times of ICKER’s. The next, we compare the time cost of them. From figure 4 to figure 6, we can see that the algorithm IKCER use less time to compile the incremental knowledge base, on average, it will save 1-10 times than KCER. V.

CONCLUSIONS

This paper proposes an algorithm for incremental knowledge compilation by using the extension rule. For an updated knowledge base, it does not need to recompile the entire knowledge base, but solves the problem of incremental knowledge compilation by using the information in knowledge base, and improves the efficiency. The experimental results show that, for the problems which have different scales and difficulties, our algorithm IKCER is about 1-10 times of KCER on efficiency, and the number of increase clauses is about 1/10 to 1/50 times of KCER’s. Therefore, the approach of

JOURNAL OF NETWORKS, VOL. 8, NO. 10, OCTOBER 2013

incremental knowledge compilation by using the extension rule can be more effective to deal with the compilation problem for knowledge base when the knowledge base changed. In future work, we would like to implement our methods and other methods for description logic, and compare these methods for description logic problems.

Figure 4. Random 2-SAT problem, comparing increased time of IKCER with KCER on random 2-SAT problem, N=20, M=40.

Figure 5. Random 3-SAT problem, comparing increased time of IKCER with KCER on random 3-SAT problem, N=20, M=40.

Figure 6. Random 4-SAT problem, comparing increased time of IKCER with KCER on random 4-SAT problem, N=20, M=40.

ACKNOWLEDGMENTS This research was supported by the National Nature Science Foundation of China under Project Number: 61070084, and the "Twelfth five-year" science and technology research of education department of Jilin under Project Number: 556.

© 2013 ACADEMY PUBLISHER

2245

REFERENCES [1] Cadoli M, Donino FM. A survey on knowledge compilation. AI Communications, 10(1997) pp. 137-150. [2] Sascha Muehlbach, Andreas Koch, Malacoda, Towards High-Level Compilation of Network Security Applications on Reconfigurable Hardware, ANCS '12 Proceedings of the eighth ACM/IEEE symposium on Architectures for networking and communications systems, Pages 247-258. [3] Xiao Zou, Heng Wang, Qiuyu Zhang, “Hand Gesture Target Model Updating and Result Forecasting Algorithm based on Mean Shift, Journal of Multimedia, 2013 Feb, 8(1) pp. 1-7. [4] Ke Xu, Wen Cui, Jun Tie, Xinfang Zhang, “An Algorithm for Detecting Group in Mobile Social Network,” Journal of Networks, 2012 Oct, 7(10) pp. 1584-1591. [5] Ren Wei-wu, Hu Liang, Zhao Kuo, Chu Jianfeng An Efficient Parallel Anomaly Detection Algorithm Based on Hierarchical Clustering. Journal of Networks, 2013 Mar, 8(3) pp. 672-679. [6] Reiter R, de Kleer J. Foundations of assumption-based truth maintenance systems pp. preliminary report. Forbus, K. S. H. In Proceedings of the Sixth National Conference on Artificial Intelligence. Seattle, WA pp. AAAI Press, 1987. 183-188. [7] Del Val A. Tractable databases pp. How to make propositional unit resolution complete through compilation. Jon Doyle, Pietro Torasso, eds. In Proceedings of Fourth International Conference on Principles of Knowledge Representation and Reasoning. San Francisco pp. Morgan Kaufmann Publishers, 1994. 551-561. [8] Marquis P. Knowledge compilation using theory prime implicates. Stefan Wermter, eds. In Proceedings of the Fourteenth International Joint Conference on Artificial Intelligence. Montreal, Quebec, Canada, 1995. 837- 843. [9] Schrag R. Compilation for critically constrained knowledge bases. In Proceedings of the Thirteenth National Conference on Artificial Intelligence. Portland, Oregon pp. AAAI Press, 1996 pp. 510-515. [10] Selman B, Kautz H. Knowledge compilation and theory approximation. Journal of the ACM, 43(1996) pp. 193-224. [11] Del Val A. An Analysis of Approximate Knowledge Compilation. Stefan Wermter, eds. In Proceedings of the Fourteenth International Joint Conference on Artificial Intelligence. Montreal, Quebec, Canada, 1995. 830-836. [12] Guoyong Shi, A survey on binary decision diagram approaches to symbolic analysis of analog integrated circuits. Analog Integrated Circuits and Signal Processing. 2013, Volume 74, Issue 2, pp 331-343 [13] Darwiche A, Marquis P. A Knowledge Compilation Map. Journal of Artificial Intelligence Research, 2002, 17 pp. 229-264. [14] Christian Muise, Sheila A. McIlraith, J. Christopher Beck, et al., DSHARP pp. Fast d-DNNF Compilation with sharpSAT, Advances in Artificial Intelligence. 2012, pp. 356-361 [15] Tingting ZOU, Lei LIU, Shuai LV. Knowledge Compilation for Description Logic Based on Concept Extension Rule. Journal of Computational Information Systems 8:6 (2012) pp. 2409-2416 [16] Gogic G, Papadimitriou C H. Incremental Recompilation of Knowledge. Journal of Artificial Intelligence Research, 1998, 8 pp. 23-37. [17] Lin Hai, Sun Jigui and Zhang Yimin. Theorem proving based on extension rule. Journal of Automated Reasoning, 31(2003) pp. 11-21.

2246

[18] Lin Hai and Sun Jigui. Knowledge compilation using the extension rule. Automated Reasoning, 32(2004) pp. 93-102. [19] Yin Minghao, Sun jigui, Lin Hai, Wu Xia. Possibilistic extension rules for reasoning and knowledge compilation. Journal of Software. 21(2010) pp. 2826-2837. [20] Yin Minghao, Lin Hai and Sun Jigui. Counting Models using Extension. John Collims, Peyman Faratin, eds. In Proceedings of the Twenty-Second AAAI Conference on Artificial Intelligence. Vancouver, British Columbia, Cannada pp. AAAI Press, 2007. 1916-1917. [21] Yin Minghao, Lin Hai and Sun Jigui. Solving #SAT using extension rules. Journal of Software. 20(2009) pp. 1714-1725. [22] Yin Minghao. The research on several issues of automated reasoning and artificial intelligent planning. Ph. D. Thesis. 2008.

Ying Liu was born in JiLin province, China in 1980. She received her bachelor’s degree in JiLin normal University, Jilin, China in 2003. She received her M. Sc. degree in computer science from Northeast Normal University in 2008. Now, she is a Ph. D. candidate at the School of Computer Science and information technology, Northeast Normal University, Changchun, China. Her current research interests include intelligence plan, plan recognition, intelligence scheduling and network security. WenXiang Gu was born in JinLin province, China in 1947. Now, he is a professor at the School of Computer Science and information technology, Northeast Normal University, Changchun, China. His current research interests include intelligence plan, plan recognition and network security. He graduated from the east to Beijing normal university department of mathematics and teaches in 1972. From 1972 to 1992 he was in

© 2013 ACADEMY PUBLISHER

JOURNAL OF NETWORKS, VOL. 8, NO. 10, OCTOBER 2013

the northeast normal university department of mathematics teaching, was appointed associate professor in 1990. Then, in 1996 he was hired as professor, and hired as a doctoral supervisor in 2006. Tingting Zou was born in JiLin Province, China in 1984. She received her bachelor’s degree in northeast normal University, Jilin, China in 2006. She received her M. Sc. degree in computer science from Northeast Normal University in 2009. Now, she is a Ph. D. candidate at the School of Computer Science and technology, JiLin University, Changchun, China. Her current research interests include intelligence plan, plan recognition, date mining and knowledge complication. Qingtao Huang was born in JiLin province, China in 1979. She received her bachelor’s degree in JiLin normal University, Jilin, China in 2001. She received her M. Sc. degree in computer science from Northeast Normal University in 2009. Her current research interests include intelligence plan, plan recognition, date mining and knowledge complication. Qiao Li was born in JiLin province, China in 1984. She received her bachelor’s degree in northeast normal University, Jilin, China in 2007. She received her M. Sc. degree in computer science from Northeast Normal University in 2012. Her current research interests include intelligence plan, plan recognition, intelligence scheduling and network security.

Recommend Documents