• Non ci sono risultati.

by the Nystr¨om method as defined 2: Generate random features X in Eq

N/A
N/A
Protected

Academic year: 2021

Condividi "by the Nystr¨om method as defined 2: Generate random features X in Eq"

Copied!
365
0
0

Testo completo

(1)B. Ma et al. / Randomized Canonical Correlation Discriminant Analysis for Face Recognition. Figure 1.. Framework based on RCCDA for face recognition.. Algorithm 1 RCCDA Input: Labeled data {X, y} and dimension of random feature m. Output: Random optimal discriminant features Z 1: Constructe the matrix of classes Y for label y by one-of-k label encoding defined in Eq. (13). ˆ by the Nystr¨om method as defined 2: Generate random features X in Eq. (16). 3: Compute optimal discriminant vector matrix WX by solving ˆ Y ). [WX , WY ] = CCA(X, ˆ X. 4: Obtain random optimal discriminant features by Z ← XW 5: return Z;. 2.4. RCCDA for Face Recognition. Based on the theory of RCCDA proposed above, as shown in Fig. 1, we present a simple framework for face recognition. Firstly, we preprocess all face images and split images into training set and test set, then represent images with extracted local features. Secondly, Algorithm 1 has been used to obtain random optimal discriminant features for face images. We extract random features and calculate optimal discriminant vectors only by training set, then project all original face representations to optimal discriminant subspace, which can reduce the dimension of face features while preserving as much discriminatory information as possible. Finally, we use random optimal discriminant features of training set to train a classifier to divided test set into different subjects.. 3. 667. Experiment Results. In this section, we evaluate the performance of the proposed RCCDA method on the Extended Yale B[19], AR[24], ORL[29] and FERET[27] face databases. These databases are widely used in evaluating the performances of various face recognition algorithms. As a comparison, we consider the basis methods Nearest Neighbor. (NN), and some state-of-the-art methods: spatially smooth LPP (SLPP)[5], discriminative Graph regularized Extreme Learning Machine (GELM)[26], Relaxed Collaborative Representation (RCR)[42], Regularized Robust Sparse Coding (RRSC)[41], Kernel Sparse Representation (KSR)[15] and Kernel Collaborative Representation (KCR)[38]. The results of GELM and KCR are from their corresponding papers for no source codes found. All face images are transformed into grayscale form and preprocessed by the code proposed in [36]. The local features are WTP descriptors[44] extracted from 3 × 3 image patches. For RCCDA, our shift-invariant kernel is Gaussian and dimension of random features is around n/4 (where n is number of training samples). We adopt the LIBSVM[6] classifier which is used in SLPP, KSR with the same parameter settings. NN used unsupervised nearest neighbor classifier. GELM trained an extreme learning machine as a classifier. Other methods used representation-based classifiers, which can be consider as least squares regressions with different constraints. We measure performance based on recognition accuracy rate, averaged over 10 random training/testing splits. All experiments are carried out on an Intel Pentium D processor with 3.4 GHZ frequency and 4 GB RAM.. 3.1. Computational Complexity Analysis. RCCDA is extremely fast. Obviously, the computational complexity of KCCA and KFDA are both O(n3 ), where n is the size of training samples. CCA is O((p2 + q 2 )n) and RCCA is O((m2x + m2y )n), where p, q are original dimensions of samples and mx , my are dimensions of random features generated by Nystr¨om method. CCA and RCCA are both linear in sample size n. By connecting KCCA and KFDA with KCCDA, RCCA can be used to speed up the caculation of KFDA. That is to say, instead of calculating the KFDA directly whose computational complexity is O(n3 ), RCCDA can be calulated with only O((m2 + c2 )n) computational complexity (where c is number of classes). We study the effect of random feature’s dimensions m on Extended Yale B database, where number of classes c = 38. Randomly 10 images for each subject are selected for training, and the others are.

(2) 668. B. Ma et al. / Randomized Canonical Correlation Discriminant Analysis for Face Recognition. Figure 4. Sample images in the extended Yale B database with variations of illumination.. Figure 2. Recognition rates of RCCDA versus different size of m with 10 training samples on Extended Yale B database.. Figure 5. Samples un-occluded (fisrt row), with sunglasses (second row) and scarves (third row) in the AR database.. 3.3. Figure 3. The face recognition results obtained on Extended Yale B database with different number of training samples.. used for test. As shown in Fig. 2, we can get good performance with a low dimension m = 100. Comparing to the recognition rate with m = 380, which is equal to KCCA or KFDA, the proposed method can yield comparable generalization performance 12 times faster.. The AR face database contains over 4,000 color images corresponding to 126 people’s faces (70 men and 56 women). All images are frontal view with different facial expressions, illumination conditions, and occlusions. In this experiment, we choose a subset that contains 50 men and 50 women. For each subject, 26 images are chosen (14 unoccluded, 6 occluded with sunglasses and 6 occluded with scarves). Each image is cropped to 165 × 120 pixels, as shown in Fig. 5. In order to evaluate the robustness of our method under occlusion, we use 1400 un-occluded images as a training set and randomly splits l images (l = 2, 4, 6, 8, 10) per individual to training a classifier. We first consider the rest of un-occluded images as a test set. Then, 600 images of subjects wearing sunglasses and 600 images of subjects wearing scarves are used to form other two test sets. Fig. 6 shows the recognition rates of occlusion tests, we can learn from them that because of robustness for WTP on expression and occlusion our approach obtains higher accuracy rate than other methods when the number of training samples are 4, 6 and 8.. 3.4 3.2. Experiments on Extended Yale B Database. The extended YaleB database contains 2414 frontal face images of 38 subjects taken under 64 illumination conditions, which has been widely used to evaluate the illumination robustness of face recognition methods. We collect the cropped and normalized face images of 192 × 168 pixels for our experiment (see Fig. 4 for some examples). A random subset with l images per individual is collected to form a training set (l = 5, 10, 20, 30, 40), and the rest is considered as the corresponding test set. Our experimental results are shown in Fig. 3, in which we can observe that our method can achieve better performance than others especially with small number of training samples (l = 5, 10). Such results benefit from that our method retains discriminatory information of samples as much as possible.. Experiments on AR Database. Experiments on ORL Database. ORL database consists of 400 different images of 40 individuals. For each subjects, there are 10 images were taken at different times, varying the lighting, facial expressions and facial details. All image are normalized to 112 × 92 in our experiment (see Fig. 7 for some examples). We randomly select l images (l = 2, 3, 4, 5) of each individual images as training data and the rest as test data. The average recognition rates are reported in Tab. 1. With 5 training samples, proposed method gets close to 100% accuracy rate.. 3.5. Experiments on FERET Database. The FERET database consists of 13539 face images of 1565 subjects who are diverse across age, gender, and ethnicity. We use a subset contains 1400 images from 200 subjects with 7 frontal images per.

(3) 669. B. Ma et al. / Randomized Canonical Correlation Discriminant Analysis for Face Recognition. Table 1. The face recognition results obtained on ORL database with different number of training samples (mean(std-dev))%. Methods. 2 Trains. 3 Trains. 4 Trains. 5 Trains. NN S-LPP GELM RCCDA. 66.9 82.9 84.2 91.2(1.9). 76.6 91.9 90.7 95.6(1.6). 82.1 95.9 94.3 97.6(2.6). 86.3 97.7 96.3 99.3(0.7). Table 2. dev))%.. The face recognition results obtained on FERET subset (mean(std-. Methods. NN. RCR. RRSC. KSR. KCR-l2. RCCDA. Accuracy. 69.5. 82.5. 81.3. 83.0. 88.3. 91.5(0.9). Figure 8.. A subset of normalized images from FERET database.. that recognition rates of the proposed method is higher than other algorithms.. 4. Figure 6. The face recognition results obtained on AR un-occluded, occluded with sunglasses and scarves subset with different number of training samples.. Conclusion and Further Work. In this paper, an efficient RCCDA method is presented for face recognition. Our method extracts local features firstly, then applies RCCDA to reduce the dimension and map the local feature into an optimal discriminant subspace. The main advantage of the proposed method is that RCCDA preserves as much discriminatory information as possible and greatly accelerates the computation through randomized method. The result of experiments on Extended Yale B, AR, ORL and FERET face databases demonstrates that our method achieve better performance than several state-of-the-art methods. As an effective feature extraction method, the proposed RCCDA can also be used to other recognition tasks, such as visual tracking, image retrieval and image classification. For these tasks, the process of feature extraction can be replaced by our method. Necessary distinguishing information can be obtained efficiently with only small number of training samples. Our future work will focus on apply the proposed method to other recognition problems and optimize the combination of local features, kernel functions and other RCCDA encoding methods (such as [33]).. Acknowledgment This work was supported in part by the National Natural Science Foundation of China (No. 61472036). We would like to thank both the editor and the reviewers for the invaluable comments. Many thanks to David Lopez-Paz for very helpful discussions.. Figure 7. pose.. Sample images in ORL database with variations of expression and. subject in this paper. Images are scaled and cropped to the size of 80 × 80. Fig. 8 shows the normalized and cropped images. In our experiment, 5 images per individual are randomly selected to form a training set, and the rest is used for tests. Tab. 2 demonstrates. REFERENCES [1] Yael Adini, Yael Moses, and Shimon Ullman, ‘Face recognition: The problem of compensating for changes in illumination direction’, Pattern Analysis and Machine Intelligence, IEEE Transactions on, 19(7), 721– 732, (1997). [2] Timo Ahonen, Abdenour Hadid, and Matti Pietik¨ainen, ‘Face recognition with local binary patterns’, in Computer vision-eccv 2004, 469–481, Springer, (2004)..

(4) 670. B. Ma et al. / Randomized Canonical Correlation Discriminant Analysis for Face Recognition. [3] Peter N Belhumeur, Jo˜ao P Hespanha, and David J Kriegman, ‘Eigenfaces vs. fisherfaces: Recognition using class specific linear projection’, Pattern Analysis and Machine Intelligence, IEEE Transactions on, 19(7), 711–720, (1997). [4] Roberto Brunelli and Tomaso Poggio, ‘Face recognition: Features versus templates’, IEEE Transactions on Pattern Analysis & Machine Intelligence, (10), 1042–1052, (1993). [5] Deng Cai, Xiaofei He, Yuxiao Hu, Jiawei Han, and Thomas Huang, ‘Learning a spatially smooth subspace for face recognition’, in Computer Vision and Pattern Recognition, 2007. CVPR’07. IEEE Conference on, pp. 1–7. IEEE, (2007). [6] Chih-Chung Chang and Chih-Jen Lin, ‘LIBSVM: A library for support vector machines’, ACM Transactions on Intelligent Systems and Technology, 2, 27:1–27:27, (2011). Software available at http: //www.csie.ntu.edu.tw/˜cjlin/libsvm. [7] Tijl De Bie, Nello Cristianini, and Roman Rosipal, ‘Eigenproblems in pattern recognition’, in Handbook of Geometric Computing, 129–167, Springer, (2005). [8] Petros Drineas and Michael W Mahoney, ‘On the nystr¨om method for approximating a gram matrix for improved kernel-based learning’, The Journal of Machine Learning Research, 6, 2153–2175, (2005). [9] Ronald A Fisher, ‘The use of multiple measurements in taxonomic problems’, Annals of eugenics, 7(2), 179–188, (1936). [10] Yunhui He, Li Zhao, and Cairong Zou, ‘Face recognition based on pca/kpca plus cca’, in Advances in Natural Computation, 71–74, Springer, (2005). [11] Bernd Heisele, Purdy Ho, Jane Wu, and Tomaso Poggio, ‘Face recognition: component-based versus global approaches’, Computer vision and image understanding, 91(1), 6–21, (2003). [12] Harold Hotelling, ‘Relations between two sets of variates’, Biometrika, 28(3/4), 321–377, (1936). [13] Xudong Jiang, Bappaditya Mandal, and Alex Kot, ‘Eigenfeature regularization and extraction in face recognition’, Pattern Analysis and Machine Intelligence, IEEE Transactions on, 30(3), 383–394, (2008). [14] Bj¨orn Johansson, ‘On classification: simultaneously reducing dimensionality and finding automatic representation using canonical correlation’, (2001). [15] Cuicui Kang, Shengcai Liao, Shiming Xiang, and Chunhong Pan, ‘Kernel sparse representation with local patterns for face recognition’, in Image Processing (ICIP), 2011 18th IEEE International Conference on, pp. 3009–3012. IEEE, (2011). [16] Kwang In Kim, Keechul Jung, and Hang Joon Kim, ‘Face recognition using kernel principal component analysis’, Signal Processing Letters, IEEE, 9(2), 40–42, (2002). [17] G Kukharev and E Kamenskaya, ‘Application of two-dimensional canonical correlation analysis for face image processing and recognition’, Pattern Recognition and Image Analysis, 20(2), 210–219, (2010). [18] Olcay Kursun, Ethem Alpaydin, and Oleg V Favorov, ‘Canonical correlation analysis using within-class coupling’, Pattern Recognition Letters, 32(2), 134–144, (2011). [19] Kuang-Chih Lee, Jeffrey Ho, and David J Kriegman, ‘Acquiring linear subspaces for face recognition under variable lighting’, Pattern Analysis and Machine Intelligence, IEEE Transactions on, 27(5), 684–698, (2005). [20] Chengjun Liu, ‘Gabor-based kernel pca with fractional power polynomial models for face recognition’, Pattern Analysis and Machine Intelligence, IEEE Transactions on, 26(5), 572–581, (2004). [21] Chengjun Liu and Harry Wechsler, ‘Gabor feature based classification using the enhanced fisher linear discriminant model for face recognition’, Image processing, IEEE Transactions on, 11(4), 467–476, (2002). [22] David Lopez-Paz, Suvrit Sra, Alex Smola, Zoubin Ghahramani, and Bernhard Sch¨olkopf, ‘Randomized nonlinear component analysis’, arXiv preprint arXiv:1402.0119, (2014). [23] Gui-Fu Lu, Zhong Lin, and Zhong Jin, ‘Face recognition using discriminant locality preserving projections based on maximum margin criterion’, Pattern Recognition, 43(10), 3572–3579, (2010). [24] Aleix M Martinez, ‘The ar face database’, CVC Technical Report, 24, (1998). [25] Brian McWilliams, David Balduzzi, and Joachim M Buhmann, ‘Correlated random features for fast semi-supervised learning’, in Advances in Neural Information Processing Systems, pp. 440–448, (2013). [26] Yong Peng, Suhang Wang, Xianzhong Long, and Bao-Liang Lu, ‘Discriminative graph regularized extreme learning machine and its application to face recognition’, Neurocomputing, 149, 340–353, (2015).. [27] P Jonathon Phillips, Hyeonjoon Moon, Syed A Rizvi, and Patrick J Rauss, ‘The feret evaluation methodology for face-recognition algorithms’, Pattern Analysis and Machine Intelligence, IEEE Transactions on, 22(10), 1090–1104, (2000). [28] Lishan Qiao, Songcan Chen, and Xiaoyang Tan, ‘Sparsity preserving projections with applications to face recognition’, Pattern Recognition, 43(1), 331–341, (2010). [29] Ferdinando S Samaria and Andy C Harter, ‘Parameterisation of a stochastic model for human face identification’, in Applications of Computer Vision, 1994., Proceedings of the Second IEEE Workshop on, pp. 138– 142. IEEE, (1994). [30] Bernhard Scholkopft and Klaus-Robert Mullert, ‘Fisher discriminant analysis with kernels’, Neural networks for signal processing IX, 1(1), 1, (1999). [31] Quan-Sen Sun, Pheng-Ann Heng, Zhong Jin, and De-Shen Xia, ‘Face recognition based on generalized canonical correlation analysis’, in Advances in intelligent computing, 958–967, Springer, (2005). [32] Quan-Sen Sun, Sheng-Gen Zeng, Yan Liu, Pheng-Ann Heng, and DeShen Xia, ‘A new method of feature fusion and its application in image recognition’, Pattern Recognition, 38(12), 2437–2448, (2005). [33] Tingkai Sun and Songcan Chen, ‘Class label versus sample label-based cca’, Applied Mathematics and computation, 185(1), 272–283, (2007). [34] Tingkai Sun, Songcan Chen, Jingyu Yang, Xuelei Hu, and Pengfei Shi, ‘Discriminative canonical correlation analysis with missing samples’, in Computer Science and Information Engineering, 2009 WRI World Congress on, volume 6, pp. 95–99. IEEE, (2009). [35] Xiaoyang Tan and Bill Triggs, ‘Fusing gabor and lbp feature sets for kernel-based face recognition’, in Analysis and Modeling of Faces and Gestures, 235–249, Springer, (2007). [36] Xiaoyang Tan and Bill Triggs, ‘Enhanced local texture feature sets for face recognition under difficult lighting conditions’, Image Processing, IEEE Transactions on, 19(6), 1635–1650, (2010). [37] Matthew Turk and Alex Pentland, ‘Eigenfaces for recognition’, Journal of cognitive neuroscience, 3(1), 71–86, (1991). [38] Dong Wang, Huchuan Lu, and Ming-Hsuan Yang, ‘Kernel collaborative face recognition’, Pattern Recognition, 48(10), 3025–3037, (2015). [39] Christopher Williams and Matthias Seeger, ‘Using the nystr¨om method to speed up kernel machines’, in Proceedings of the 14th Annual Conference on Neural Information Processing Systems, number EPFL-CONF161322, pp. 682–688, (2001). [40] Jian Yang, David Zhang, Alejandro F Frangi, and Jing-yu Yang, ‘Twodimensional pca: a new approach to appearance-based face representation and recognition’, Pattern Analysis and Machine Intelligence, IEEE Transactions on, 26(1), 131–137, (2004). [41] Meng Yang, Lei Zhang, Jian Yang, and Dejing Zhang, ‘Regularized robust coding for face recognition’, Image Processing, IEEE Transactions on, 22(5), 1753–1766, (2013). [42] Meng Yang, Lei Zhang, David Zhang, and Shenlong Wang, ‘Relaxed collaborative representation for pattern classification’, in Computer Vision and Pattern Recognition (CVPR), 2012 IEEE Conference on, pp. 2224–2231. IEEE, (2012). [43] Tianbao Yang, Yu-Feng Li, Mehrdad Mahdavi, Rong Jin, and Zhi-Hua Zhou, ‘Nystr¨om method vs random fourier features: A theoretical and empirical comparison’, in Advances in neural information processing systems, pp. 476–484, (2012). [44] Zuodong Yang, Yinyan Jiang, Yong Wu, Zongqing Lu, Weifeng Li, and Qingmin Liao, ‘Weber binary pattern and weber ternary pattern for illumination-robust face recognition’, in 2015 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA), pp. 1050–1053. IEEE, (2015). [45] Wenchao Zhang, Shiguang Shan, Wen Gao, Xilin Chen, and Hongming Zhang, ‘Local gabor binary pattern histogram sequence (lgbphs): A novel non-statistical model for face representation and recognition’, in Computer Vision, 2005. ICCV 2005. Tenth IEEE International Conference on, volume 1, pp. 786–791. IEEE, (2005). [46] Wenyi Zhao, Rama Chellappa, P Jonathon Phillips, and Azriel Rosenfeld, ‘Face recognition: A literature survey’, ACM computing surveys (CSUR), 35(4), 399–458, (2003)..

(5) ECAI 2016 G.A. Kaminka et al. (Eds.) © 2016 The Authors and IOS Press. This article is published online with Open Access by IOS Press and distributed under the terms of the Creative Commons Attribution Non-Commercial License 4.0 (CC BY-NC 4.0). doi:10.3233/978-1-61499-672-9-671. 671. Reconsidering AGM-Style Belief Revision in the Context of Logic Programs Zhiqiang Zhuang1 and James Delgrande2 and Abhaya Nayak3 and Abdul Sattar1 Abstract. Belief revision has been studied mainly with respect to background logics that are monotonic in character. In this paper we study belief revision when the underlying logic is non-monotonic instead— an inherently interesting problem that is under explored. In particular, we will focus on the revision of a body of beliefs that is represented as a logic program under the answer set semantics, while the new information is also similarly represented as a logic program. Our approach is driven by the observation that unlike in a monotonic setting where, when necessary, consistency in a revised body of beliefs is maintained by jettisoning some old beliefs, in a non-monotonic setting consistency can be restored by adding new beliefs as well. We will define two revision functions through syntactic and modeltheoretic methods respectively and subsequently provide representation theorems for characterising them.. 1. Introduction. The ability to change one’s beliefs when presented with new information is crucial for any intelligent agent. In the area of belief change, substantial effort has been made towards the understanding and realisation of this process. Traditionally, it is assumed that the agent’s reasoning is governed by a monotonic logic. For this reason, traditional belief change is inapplicable when the agent’s reasoning is non-monotonic. Our goal in this research program is to extend the established (AGM) belief set [1] and belief base [13] approaches in belief revision to nonmonotonic setting. In this paper, we focus on disjunctive logic programs, as a well-studied and well-known approach to nonmonotonic reasoning that also has efficient implementations. Much, if not most, of our day-to-day reasoning involves nonmonotonic reasoning. To illustrate issues that may arise, consider the following example. In a university, professors generally teach, unless they have an administrative appointment. Assume we know that John is a professor. Since most faculty do not have an administrative appointment, and there is no evidence that John does, we conclude that he teaches. This reasoning is a classical form of nonmonotonic reasoning, namely using the closed world assumption. It can be represented by the following logic program under the answer set semantics. T each(John) ← P rof (John), not Admin(John). P rof (John) ← .. (1) (2). The answer set {P rof (John), T each(John)} for this logic program corresponds exactly to the facts we can conclude. 1 2 3. IIIS, Griffith University, Australia School of Computing Science, Simon Fraser University, Canada Department of Computing, Macquarie University, Australia. Suppose we receive information that John does not teach, which we can represent by the rule ← T each(John).. (3). Now our beliefs about John are contradictory; and it is not surprising that the logic program consisting of rules (1) – (3) has no answer set. For us or any intelligent agent in this situation to function properly, we need a mechanism to resolve this inconsistency. This is a typical belief revision problem; however, the classical (AGM) approach can not be applied, as we are reasoning non-monotonically. It is not hard to suggest possible causes of the inconsistency and to resolve it. It could be that some of our beliefs are wrong; perhaps professors with administrative duties may still need to do teaching or perhaps John is not a professor. Thus we can restore consistency by removing rule (1) or (2). Alternatively and perhaps more interestingly, it could be that assuming that John is not an administrative staff via the absence of evidence is too adventurous; that is he may indeed be an administrative staff member but we don’t know it. Thus we can also restore consistency by adding the missing evidence of John being an administrative staff member by Admin(John) ← .. (4). The second alternative highlights the distinction for belief revision in monotonic and non-monotonic settings. In the monotonic setting, an inconsistent body of knowledge will remain inconsistent no matter how much extra information is supplied. On the other hand, in the non-monotonic setting, inconsistency can be resolved by either removing old information, or adding new information, or both. Therefore, belief revision functions in a non-monotonic setting should allow a mixture of removal and addition of information for inconsistency-resolution. In this paper, we will define two such revision functions for disjunctive logic programs under the answer set semantics. Our first revision function called slp-revision4 is like belief base revision which takes syntactic information into account. In revising P by Q, a slp-revision function first obtains a logic program R that is consistent with Q and differs minimally from P , then combines R with Q. For example, if P = {(1), (2)} and Q = {(3)}, then R could be {(1)} (i.e., resolving inconsistency by removing (2)); {(2)} (i.e., resolving inconsistency by removing (1)); or {(1), (2), (4)} (i.e., resolving inconsistency by adding (4)). Our second revision function called llp-revision function5 is like AGM belief set revision which ignores syntactic difference and focuses on the logical content of a knowledge base. So in revising P by Q, a llp-revision function 4 5. “s” stands for syntactic and “lp” for logic program. “l” stands for logical content and “lp” for logic program..

(6) 672. Z. Zhuang et al. / Reconsidering AGM-Style Belief Revision in the Context of Logic Programs. will instead obtain a logic program R whose logical content differ the least from that of P where the logical content is characterised by strong equivalent (SE) models [27]. The next section gives logical preliminaries. The following two sections develop our approach to slp-revision and llp-revision, in each case providing postulates, a semantic construction, and a representation result. This is followed by a comparison to other work, and a brief conclusion.. 2. Preliminary Considerations. In this paper, we consider only fully grounded disjunctive logic programs. Thus a logic program (or program for short) here is a finite set of rules of the form: a1 ; . . . ; am ← b1 , . . . , bn , not c1 , . . . , not co where m, n, o ≥ 0, m + n + o > 0, and ai , bj , ck ∈ A for A a finite set of propositional atoms. We denote the set of all logic programs by P. For each rule r, let H(r) = {a1 , . . . , an }, B + (r) = {b1 , . . . , bm }, and B − (r) = {c1 , . . . , co }. The letters P , Q and R are used to denote a logic program throughout the paper. An interpretation is represented by the subset of atoms in A that are true in the interpretation. A classical model of a program P is an interpretation in which all rules of P are true according to the standard definition of truth in propositional logic, and where default negation is treated as classical negation. The set of classical models of P is denoted as M od(P ). Given an interpretation Y , we write Y |= P to mean Y is a classical model of P . The reduct of a program P with respect to an interpretation Y , denoted P Y , is the set of rules: {H(r) ← B + (r) | r ∈ P, B − (r) ∩ Y = ∅}. An answer set Y of P is a subset-minimal classical model of P Y . The set of all answer sets of P is denoted as AS(P ). An SE interpretation [27] is a pair (X, Y ) of interpretations such that X ⊆ Y ⊆ A. The set of all SE interpretations (over A) is denoted SE. The letters M and N are used to denote a set of SE interpretations throughout the paper. An SE interpretation is an SE model of a program P if Y |= P and X |= P Y . The set of all SE models of P is denoted as SE(P ). SE models are first proposed to capture strong equivalence [18] between programs that is SE(P ) = SE(Q) iff P and Q are strongly equivalent, thus they contain more informations than answer sets. For this reason, SE models have been used by many to characterise the logical content of a program [5, 7]. The following definitions and results are given in [10, 8]. A set M of SE interpretations is well-defined if (X, Y ) ∈ M implies (Y, Y ) ∈ M . M is complete if it is well-defined and if (X, Y ) ∈ M , (Z, Z) ∈ M and Y ⊆ Z, then (X, Z) ∈ M . For each disjunctive logic program P , SE(P ) is complete and for each complete set M , there is a unique (up to strong equivalence) disjunctive logic program P such that SE(P ) = M . In this paper, we work with two notions of closure. The closure of a set M under completeness is denoted as Cl(M ). Formally, Cl(M ) is the minimal superset of M such that 1. if (X, Y ) ∈ M , then (Y, Y ) ∈ Cl(M ) and 2. if (X, Y ), (Z, Z) ∈ M and Y ⊆ Z, then (X, Z) ∈ Cl(M ). The closure of a program P which intends to capture all logical consequence of P is denoted as cl(P ). Formally cl(P ) = {r | SE(P ) ⊆ SE({r}) and r is a program rule}. We say P is closed if P = cl(P ). The following two properties of SE models [27] are crucial to this paper:. 1. Y ∈ AS(P ) iff (Y, Y ) ∈ SE(P ) and there is no (X, Y ) ∈ SE(P ) such that X ⊂ Y . 2. (Y, Y ) ∈ SE(P ) iff Y ∈ M od(P ). So SE(P ) = ∅ iff M od(P ) = ∅ but SE(P ) = ∅ does not imply AS(P ) = ∅. This gives rise to two notions of consistency. Definition 1. P is consistent iff AS(P ) = ∅ and P is m-consistent6 iff SE(P ) = ∅. It follows from the SE model properties that consistency implies m-consistency and m-inconsistency implies inconsistency. In other words, a consistent program is m-consistent but not vice versa. For convenience, we say a set M of SE interpretations is consistent iff the program P such that SE(P ) = Cl(M ) is consistent; we say (Y, Y ) is an “answer set” in M iff (Y, Y ) ∈ M and there is no (X, Y ) ∈ M such that X ⊂ Y . Clearly, M is consistent iff Cl(M ) is consistent; the consistency of M indicates the consistency of its corresponding logic program; and (Y, Y ) being an answer set in M indicates Y is an answer set of the corresponding program. In subsequent sections, we will need to describe the difference between two logic programs and between their sets of SE models. For this purpose, we use the symmetric difference operator " which is defined as X " Y = (X \ Y ) ∪ (Y \ X) for any sets X and Y .. 3. SLP-Revision Functions. In this section, we give a syntax-based revision function ∗ : P ×P #→ P for revising one logic program by another. The function maps a logic program P called the original logic program and a logic program Q called the revising logic program, to another logic program P ∗ Q called the revised logic program. Following AGM belief revision, we want to have Q contained in P ∗ Q (i.e., Q ⊆ P ∗ Q) and P ∗ Q is consistent whenever possible. A main task in defining ∗ is to deal with the possible inconsistency between Q and P . As illustrated in the teaching example, one means of ensuring that P ∗ Q is consistent is to remove a minimal set of beliefs from P so that adding Q to the result is consistent. Of course there may be more than one way to remove beliefs from P . Following this intuition, we obtain all maximal subsets of P that are consistent with Q, which we call the s-removal compatible programs of P with respect to Q. Definition 2. The set of s-removal compatible programs of P with respect to Q, denoted P ↓ Q, is such that R ∈ P ↓ Q iff 1. R ⊆ P , 2. R ∪ Q is consistent, and 3. if R ⊂ R ⊆ P , then R ∪ Q is inconsistent. The notion of s-removal compatible programs is not new, classical revision functions [1, 11] are based on more or less the same notion. The difference is that this notion alone is sufficient to capture the inconsistency-resolution strategy of classical belief revision, but there is more that one can do in non-monotonic belief revision. In our non-monotonic setting, we are able to express assumptions (i.e., negation as failure) and to reason with them. Earlier, we assumed John is not an administrator, in the absence of evidence to the 6. “m” stands for “monotonic” which indicates that the notion of mconsistency is based on a monotonic characterisation (i.e., SE models) for logic programs..

(7) Z. Zhuang et al. / Reconsidering AGM-Style Belief Revision in the Context of Logic Programs. 673. contrary. With this, we came to the conclusion that he has to teach. Consequently, if we learn that John does not teach, as in our example, one way of resolving this inconsistency is by adding a minimal set of information so that our assumption does not hold. Following this intuition, we obtain all the minimal supersets of P that are consistent with Q, which we call the s-expansion compatible programs of P with respect to Q.. Definition 5. A function γ is a selection function for P iff for any program Q, γ(P & Q) returns a single element of P & Q whenever P & Q is non-empty; otherwise it returns P .. Definition 3. The set of s-expansion compatible programs of P with respect to Q, denoted P ↑ Q, is such that R ∈ P ↑ Q iff. Definition 6. A function ∗ : P × P #→ P is a slp-revision function iff P ∗ Q = γ(P & Q) ∪ Q. 1. P ⊆ R, 2. R ∪ Q is consistent, and 3. if P ⊆ R ⊂ R, then R ∪ Q is inconsistent. Since the s-expansion and s-removal compatible programs are consistent with Q and are obtained by removing or adding minimal sets of rules from or to P , the union of Q with any of these sets is consistent and comprises a least change made to P in order to achieve consistency. These programs clearly should be candidates for forming the revised logic program P ∗ Q; however, they do not form the set of all candidates. In particular, we can obtain a program that differs the least from P and is consistent with Q by removing some beliefs of P and at the same time adding some new beliefs to P . Thus we consider all those logic programs that differ the least from P and are consistent with Q; these are called the s-compatible programs of P with respect to Q. Definition 4. The set of s-compatible programs of P with respect to Q, denoted P & Q, is such that R ∈ P & Q iff 1. R ∪ Q is consistent and 2. if P " R ⊂ P " R, then R ∪ Q is inconsistent. For example, let P = {a ← b, not c., b., e ← f, not g., f.} and Q = {← a., ← e.}. Then P ∪ Q is inconsistent since a and e can be concluded from P but they contradict the rules of Q. To resolve the inconsistency via making the least change to P , we could remove b ← from P (which eliminates the contradiction about a) and add g ← to P (which eliminates the contradiction about e). The program thus obtained (i.e., (P \ {b.}) ∪ {g.}) is a s-compatible program in P & Q. It is obvious, but worth noting that the notion of s-compatible program subsumes those of s-removal and s-expansion compatible programs. In the above example, P & Q also contains P \ {b., f.} and P ∪ {c., g.}, which are respectively an s-removal and an s-expansion compatible program of P with respect to Q. Proposition 1. (P ↑ Q) ∪ (P ↓ Q) ⊆ P & Q. There are cases in which we cannot resolve inconsistency by only adding new beliefs which means the set of s-expansion compatible programs is empty. For example, if P = {a.} and Q = {← a.}, then P ∪ Q is inconsistent and we cannot restore consistency without removing a ← from P . In these cases, the inconsistency is due to contradictory facts that can be concluded without using any reasoning power beyond that of classical logic. Clearly, the inconsistency is of a monotonic nature, that is, in our terminology, m-inconsistency. Proposition 2. If P ∪ Q is m-inconsistent, then P ↑ Q = ∅. So far, we have identified the candidates for forming P ∗ Q. It remains to pick the “best” one. Such extralogical information is typically modelled by a selection function, which we do next.. The revised logic program P ∗ Q is then formed by combining Q with the s-compatible program picked by the selection function for P . We call the function ∗ defined in this way a slp-revision function for P .. where γ is a selection function for P . In classical belief revision, multiple candidates maybe chosen by a selection function, and their intersection is combined with the new belief to form the revision result. There, a selection function that picks out a single element is called a maxichoice function [1]. In classical logic, maxichoice selection functions lead to undesirable properties for belief set revision but not for belief base revision. In our non-monotonic setting, picking multiple candidates does not make sense, as intersection of s-compatible programs may not be consistent with the revising program. For example, let P = {a ← not b, not c.} and Q = {← a.}. We can restore consistency of P with Q by, for instance, adding the rule b ← to P which corresponds to the s-compatible program P ∪ {b.} or by adding the rule c ← which corresponds to the s-compatible program P ∪ {c.}. However, the intersection of the two s-compatible programs is inconsistent with Q. We turn next to properties of slp-revision functions. Consider the following set of postulates for a function ∗ : P × P #→ P. (s∗s) Q ⊆ P ∗ Q. (s∗c) If Q is m-consistent, then P ∗ Q is consistent. (s∗f) If Q is m-inconsistent, then P ∗ Q = P ∪ Q. (s∗rr) If R = ∅ and R ⊆ P \ (P ∗ Q), then (P ∗ Q) ∪ R is inconsistent. (s∗er) If E = ∅ and E ⊆ (P ∗ Q) \ (P ∪ Q), then (P ∗ Q) \ E is inconsistent. (s∗mr) If R = ∅, R ⊆ P \ (P ∗ Q), E = ∅, and E ⊆ (P ∗ Q) \ (P ∪ Q), then ((P ∗ Q) ∪ R) \ E is inconsistent. (s∗u) If P & Q = P & R, then P \ (P ∗ Q) = P \ (P ∗ R) and (P ∗ Q) \ (P ∪ Q) = (P ∗ R) \ (P ∪ R). These postulates are obtained by adapting the classical belief base revision postulates [13] to the present non-monotonic setting. (s∗s) (Success) states that a revision is always successful in incorporating the new beliefs. (s∗c) (Consistency) states that a revision ensures consistency of the revised logic program whenever possible. In the monotonic setting, a revision results in inconsistency only when the new beliefs are themselves inconsistent. This is not the case in the non-monotonic setting. For example, consider the revision of P = {a.} by Q = {b ← not b}. Although Q is inconsistent, we have P ∪ {b.} as a s-compatible program of P with respect to Q. Thus we can have P ∪ {b.} ∪ Q as the revised logic program, which contains Q and is consistent. Here, a revision results in inconsistency only when the revising logic program is m-inconsistent. In such a case, (s∗f) (Failure) states that the revision corresponds to the union of the original and revising logic program. (s∗rr) (Removal Relevance) states that if some rules are removed from the original logic program for the revision, then adding them to the revised logic program results in inconsistency. It captures the.

(8) 674. Z. Zhuang et al. / Reconsidering AGM-Style Belief Revision in the Context of Logic Programs. intuition that nothing is removed unless its removal contributes to making the revised logic program consistent. (s∗er) (Expansion Relevance) states that if some new rules other than those in the revising logic program are added to the original logic program for the revision, then removing them from the revised logic program results in inconsistency. It captures the intuition that nothing is added unless adding it contributes to making the revised logic program consistent. (s∗mr) (Mixed Relevance) states that if some rules are removed from the original logic program and some new rules other than those in the revising logic program are added to the original logic program for the revision, then adding back the removed ones and removing the added ones result in inconsistency. Its intuition is a mixture of the two above. Note that putting (s∗rr) and (s∗er) together does not guarantee (s∗mr), nor the reverse. In summary, these three postulates express the necessity of adding and/or removing certain belief for resolving inconsistency and hence to accomplish a revision. In classical belief revision, inconsistency can only be resolved by removing old beliefs; the necessity of removing particular beliefs is captured by the Relevance postulate [11].7 The three postulates are the counterparts of Relevance in our non-monotonic setting, and we need all three of them to deal respectively with addition, removal, and a mixture of addition and removal. Finally, (s∗u) (Uniformity) states the condition under which two revising logic programs Q and R trigger the same changes to the original logic program P . That is the rules removed from P (i.e., P \ (P ∗ Q)) and the rules added to P (i.e., (P ∗ Q) \ (P ∪ Q)) for accommodating Q are identical to those for accommodating R. Certainly having Q and R be strongly equivalent (i.e., SE(Q) = SE(R)) is a sufficient condition. However, it is too strong a requirement. Suppose P = {← a.}, Q = {a.}, and R = {a ← b., b.}. Then the minimal change to P we have to made to accommodate Q and R are the same, that is we remove ← a. However Q and R are not strongly equivalent, even though they incur the same change to P . The essential point of this example is that instead of a global condition like strong equivalence, we need a condition that is local to the original logic program P . Unfortunately, it seems there is no existing notion in the logic programming literature that captures this local condition. Thus we use our newly defined notion of s-compatible programs and come up with the local but more appropriate condition in (s∗u). We can show that these postulates are sufficient to characterise all slp-revision functions. Theorem 1. A function ∗ is a slp-revision function iff it satisfies (s∗s), (s∗c), (s∗f), (s∗rr), (s∗er), (s∗mr), and (s∗u).. 4. LLP-Revision Functions. Slp-revision functions preserve the syntactic structure of the original logic program as much as possible. Thus is most useful for scenarios in which syntactic information is prioritised over that of logical content. In this section, we provide a revision function called llp-revision function that prioritises the preservation of logical content over that of syntactic information. The main strategy of llp-revision functions is the same as that of slp-revision functions, which is to first obtain a logic program that differs the least from the original one and that is consistent with the revising one, and then combine it with the revising program. The 7. If ψ ∈ K and ψ ∈ K ∗ φ, then there is some K  such that K ∗ φ ⊆ K  ⊆ K ∪ {φ}, K  is consistent but K  ∪ {ψ} is inconsistent.. distinguishing feature of llp-revision is in the interpretation of “differs the least”. Slp-revision interprets this notion as symmetric difference between the constituent rules whereas llp-revision interprets it as between the logical content. Since the standard approach in characterising the logical content of logic programs is through their SE models, the difference between logical content is represented as the difference between sets of SE models. This brings about the following dual notion of s-compatible program which we call l-compatible program. Note that although we are using SE models, we concern with the stronger notion of consistency (i.e. P is consistent if it has an answer set). Definition 7. The set of l-compatible programs of P with respect to Q, denoted P ' Q, is such that R ∈ P ' Q iff 1. R ∪ Q is consistent, and 2. if SE(P ) " SE(R ) ⊂ SE(P ) " SE(R), then R ∪ Q is inconsistent. A l-compatible program of P with respect to Q is a logic program that is consistent with Q (condition 1) and whose set of SE models differ minimally from that of P (condition 2). We denote by P ⇓ Q and P ⇑ Q the set of l-removal compatible programs and l-expansion compatible programs such that R ∈ P ⇓ Q iff SE(P ) ⊆ SE(R) and R ∈ P ' Q; R ∈ P ⇑ Q iff SE(R) ⊆ SE(P ) and R ∈ P ' Q. It is easy to see that Proposition 1 and 2 also hold for l-compatible programs. We have given a declarative definition for l-compatible programs. The following theorem identifies all the possible constituting SE models of a l-compatible program. Theorem 2. If R ∈ P ' Q, then one of the following holds: 1. P ∪ Q is inconsistent. SE(R) = Cl((SE(P ) ∪ {(Y, Y )}) \ M ) where (Y, Y ) ∈ SE(Q) \ SE(P ) and M = {(X, Z) ∈ SE(P ) | (X, Y ) ∈ SE(Q) and Z ⊆ Y }.8 2. P ∪Q is inconsistent but m-consistent. SE(R) = SE(P )\M where there is (Y, Y ) ∈ SE(P )∩SE(Q) such that (W, W ) ∈ SE(P ) ∩ SE(Q) implies W ⊂ Y and M = {(X, Z) ∈ SE(P ) | (X, Y ) ∈ SE(Q) ∩ SE(P ), X = Y and Z ⊆ Y }.8 3. P ∪ Q is consistent. SE(R) = SE(P ). In the AGM setting, a belief set K is inconsistent with another one K  iff the (classical) models of K disjoint with those of K  . So if a belief set K  differs minimally from K, but is consistent with K  , then the models of K  consist of the models of K together with a single model of K. In our non-monotonic setting, P can be inconsistent with Q even though the (SE) models of P intersect with those of Q. In this setting, there are more options for a program to be consistent with Q and such that its models differ minimally from those of P . Theorem 2 makes explicit all such options. Firstly, the set of SE models of the program could contain SE models that are not in SE(P ) and it does not contain all SE models of P (i.e., condition 1 of Theorem 2). Secondly, set of SE models of the program could contain only SE models in SE(P ) but not all of them (i.e., condition 2 of Theorem 2). In both cases, the exclusion of some SE models of P (i.e., M ) is necessary for guaranteeing the program is consistent with Q. Now we give the definition of llp-revision functions. As for slprevision functions, a selection function γ is assumed for choosing the “best” l-compatible program. Then the closure of the chosen lcompatible program and the revising logic program is returned as the revised logic program. 8. For (X, Z) ∈ M , X may be identical to Z..

(9) Z. Zhuang et al. / Reconsidering AGM-Style Belief Revision in the Context of Logic Programs. Definition 8. A function ∗ : P × P #→ P is a llp-revision function iff P ∗ Q = cl(γ(P ' Q) ∪ Q) where γ is a selection function for P . For properties of llp-revision functions, consider the following set of postulates for a function ∗ : P × P #→ P. (l∗cl) P ∗ Q = cl(P ∗ Q). (l∗s) Q ⊆ P ∗ Q. (l∗c) If Q is m-consistent, then P ∗ Q is consistent. (l∗f) If Q is m-inconsistent, then P ∗ Q = cl(P ∪ Q). (l∗rr) If M = ∅ and M ⊆ (SE(P ) ∩ SE(Q)) \ SE(P ∗ Q), then SE(P ∗ Q) ∪ M is inconsistent. (l∗er) If N = ∅ and N ⊆ SE(P ∗Q)\SE(P ), then either SE(P ∗ Q) \ N is inconsistent or Cl(SE(P ∗ Q) \ N ) = SE(P ∗ Q). (l∗mr) If M = ∅, M ⊆ (SE(P ) ∩ SE(Q)) \ SE(P ∗ Q), N = ∅ and N ⊆ SE(P ∗ Q) \ SE(P ), then (SE(P ∗ Q) ∪ M ) \ N is inconsistent. (l∗u) If P ' Q = P ' R, then SE(P ∗ Q) \ SE(P ) = SE(P ∗ R) \ SE(P ) and {μ ∈ SE(P ) | Cl(SE(P ∗ Q) ∪ {μ}) ∩ SE(Q) is inconsistent} = {μ ∈ SE(P ) | Cl(SE(P ∗ R) ∪ {μ}) ∩ SE(R) is inconsistent}. (l∗cl) (Closure) states that the revised logic program is closed. (l∗s), (l∗c) and (l∗f) are the same as their counterparts for slp-revision function, except for (l∗f) in which the revised logic program has to be closed. Since (l∗s) requires that SE(P ∗ Q) is a subset of SE(Q), SE models outside of SE(Q) is of no concern. Then the principle of minimal change dictates that as much as possible the intersecting models of P and Q are preserved and as least as possible the models outside of SE(P ) are added. (l∗rr) states that if a subset M of SE(P )∩SE(Q) is not in SE(P ∗Q), then adding M to SE(P ∗Q) results in inconsistency. The postulate captures the intuition that no intersecting models of P and Q is excluded in the revision unless this contributes to the consistency of P ∗ Q. (l∗er) states that if a subset N of SE(P ∗ Q) is not in SE(P ), then removing N from SE(P ∗ Q) results in inconsistency or its inclusion in SE(P ∗ Q) is required for SE(P ∗ Q) to be closed under completeness. Recall that the set of SE models of any logic program has to be closed under completeness. The postulate captures the intuition that no SE models outside SE(P ) is included in those of P ∗ Q unless this contributes to the consistency of P ∗ Q or to the closure of the set of SE models of P ∗ Q. (l∗mr) is a mixture of (l∗rr) and (l∗er), capturing the necessity of exclusion of SE models in SE(P ) ∩ SE(Q) and inclusion of SE models not in SE(P ). Finally, (l∗u) states the condition under which the two revising programs Q and R trigger the same changes to the original program in terms of SE models. That is, in accommodating Q, SE models of the revised program that are not those of P (i.e., SE(P ∗Q)\SE(P )) and SE models of P that can not be SE models of the revised program (i.e., {μ ∈ SE(P ) | Cl(SE(P ∗ Q) ∪ {μ}) ∩ SE(Q) is inconsistent}) are identical to those in accommodating R. We can show that these postulates are sufficient to characterise all llp-revision functions. Theorem 3. A function ∗ is a llp-revision function iff ∗ satisfies (l∗cl), (l∗s), (l∗c), (l∗f), (l∗rr), (l∗er), (l∗mr), and (l∗u).. 5. 675. Related Work. There has been much work on belief revision for logic programs. Delgrande et al [7] generalise Satoh’s [22] and Dalal’s [4] revision operators to logic programs. Significantly, they bring SE model into the picture which has inspired several approaches that are based on SE models. Slota and Leite [24, 25, 26] investigate Katsuno and Mendelzon style [14] update for logic programs. Schwind and Inoue [23] provide a constructive characterisation for the revision operators in [7]. Delgrande et al [5] adapt the model-based revision of Katsuno and Mendelzon [15] to logic programs. Finally, Binnewies et al [3] provide a variant of partial meet revision and contraction for logic programs. Comparing with our llp-revision function which also makes use of SE models, these approaches assume a weaker notion of consistency, that is m-consistency. For this reason, some contradictions will not be dealt with in these approaches. For instance, the contradictory rule a ← not a is m-consistent thus is considered to be an acceptable state of belief. Also in our teaching example, as the program consisting of rules (1) – (3) is m-consistent, no attempt will be made to resolve the contradiction about John’s teaching duty by the SE model approaches. Therefore for application scenarios in which such contradictions can not be tolerant, our llp-revision function is clearly a better choice. Apart from the SE model approaches, Kr¨umpelmann and KernIsberner [16] provide a revision function for logic programs that originates from Hansson’s semi-revision [12]. For reference we call the revision function base revision function as the revision they considered is belief base revision. Since they assume the same notion of consistency as ours, all the above mentioned contradictions will be resolved in their approach. As we have noted, classical belief revision is defined for monotonic setting, not for non-monotonic ones. Inconsistency can be caused by wrong assumptions in the non-monotonic setting but not in the monotonic setting. Such causes are not considered in [16]. Consequently, their approach only supports one of the many possible inconsistency-resolution strategies we have developed. Specifically, in [16], inconsistency can be resolved only by removing old beliefs; this strategy is captured by a notion analogous to s-removal compatible programs. The inconsistency-resolution strategies captured by the notion of s-expansion compatible program and s-compatible program in general are not considered. A group of work under the title of update [2, 6, 9, 17, 20, 21, 28, 19, 29, 30] also deals with changes of logic programs. The update however is different from the Katsuno and Mendelzon style update [14]. Following [2, 17], a typical problem setting is to consider a sequence P1 , P2 , . . . , Pn of programs such that 1 ≤ i < j ≤ n implies Pi has higher priority over Pj . The goal of the update then is to obtain a set of answer sets from such a program sequence that in some sense respects the priority ordering. Clearly, these approaches have very different focus from ours which is to obtain a single new logic program.. 6. Conclusion and Future Work. Depending on the application, the logic governing an agent’s beliefs could be either monotonic or non-monotonic. Traditional belief revision assumes that an agent reasons monotonically; therefore, by definition, it is applicable to such situations only. Here we have aimed to study belief revision for situations in which the agent reasons nonmonotonically. To this end, we defined slp-revision function and llprevision function for disjunctive logic programs under the answer set.

(10) 676. Z. Zhuang et al. / Reconsidering AGM-Style Belief Revision in the Context of Logic Programs. semantics, catering respectively for application scenarios that prioritise the preservation of the syntactic structure and that prioritise the preservation of logical content. Inconsistency-resolution is an essential task for belief revision. However, the strategies used in traditional belief revision functions are limited to situations when the agent reasons monotonically. With a logic program we have the luxury of making assumptions via lack of contrary evidence, and we can deduce certain facts from such assumptions. Thus if a set of beliefs is inconsistent, then one possible cause is that we made the wrong assumption. In such cases, we can resolve the inconsistency by adding some new rules so that the assumption can no longer be made. Such a cause of inconsistency and the associated inconsistency-resolution strategy is beyond the scope of traditional belief revision, but is crucial for non-monotonic belief revision. We argue that this rationale, which is encoded in our belief revision functions, captures the fundamental difference between monotonic and non-monotonic belief revision. This paper then has explored AGM-style revision and belief base revision in the non-monotonic setting of disjunctive logic programs; in future work we propose to extend this to a general approach to belief revision in arbitrary non-monotonic settings.. Acknowledgements We would like to thank Sebastian Binnewies, Yifan Jin, and Kewen Wang for some fruitful discussions that help improve the paper.. Appendix: Proof of Results Lemma 1. If X ∈ P & Q, then P ∩ Q ⊆ X and X ∩ (Q \ P ) = ∅. Proof. Let X ∈ P & Q. Assume P ∩ Q ⊆ X. Then P ∩ Q = ∅. By the definition of compatible programs, we have X ∪ Q is consistent. Then since X ∪ Q = (X ∪ (P ∩ Q)) ∪ Q, (X ∪ (P ∩ Q)) ∪ Q is consistent. It then follows from ((X ∪ (P ∩ Q)) " P ) ⊂ (X " P ) that X ∈ P & Q, a contradiction! Assume X ∩ (Q \ P ) = ∅. Then Q \ P = ∅. Since X ∪ Q = (X \ (Q \ P )) ∪ Q, (X \ (Q \ P )) ∪ Q is consistent. It then follows from (X \ (Q \ P )) ⊂ (X " P ) that X ∈ P & Q, a contradiction!. Proof for Proposition 2 Suppose P ∪ Q is m-inconsistent. We need to show P ↑ Q = ∅. Since P ∪Q is m-inconsistent, we have SE(P )∩SE(Q) = ∅. By the definition of expansion compatible program, any element in P ↑ Q has to be a superset of P and consistent with Q. However, for any superset R of P , SE(R) ⊆ SE(P ). Thus SE(R) ∩ SE(Q) = ∅ which implies R ∪ Q is m-inconsistent. Proof for Theorem 1 ⇒: Suppose ∗ is a slp-revision function. Let γ be the selection function for P . We need to show ∗ satisfies (s∗s), (s∗c), (s∗f), (s∗rr), (s∗er), (s∗mr), and (s∗u). (s∗s), (s∗c), and (s∗f) follow immediately from the definition of slp-revision functions. (s∗rr): Suppose there is a program R such that R = ∅ and R ⊆ P \ (P ∗ Q). By the definition of slp-revision functions, we have P ∗ Q = γ(P & Q) ∪ Q, hence γ(P & Q) = P for otherwise R = ∅. Then we have by the definition of selection functions that P & Q = ∅ and γ(P & Q) ∈ P & Q. Let γ(P & Q) = X. Then (P ∗ Q) ∪ R = X ∪ Q ∪ R. Since R ∩ X = ∅, R = ∅, and R ⊆ P , we have ((X ∪ R) " P ) ⊂ (X " P ). Thus we have by the. definition of compatible programs that X ∪ R ∪ Q is inconsistent, that is (P ∗ Q) ∪ R is inconsistent. (s∗er): Suppose there is a program E such that E = ∅ and E ⊆ (P ∗ Q) \ (P ∪ Q). By the definition of slp-revision functions, we have P ∗ Q = γ(P & Q) ∪ Q, hence γ(P & Q) = P for otherwise E∅. Then we have by the definition of selection functions that P & Q = ∅ and γ(P & Q) ∈ P & Q. Let γ(P & Q) = X. Then (P ∗ Q) \ E = (X ∪ Q) \ E. Since E ∩ P = ∅, E = ∅, and E ⊆ X, we have ((X \ E) " P ) ⊂ (X " P ). Thus we have by the definition of compatible programs that (X \ E) ∪ Q is inconsistent. Then since E ∩ Q = ∅, we have (X \ E) ∪ Q = (X ∪ Q) \ E = (P ∗ Q) \ E. Thus (P ∗ Q) \ E is inconsistent. (s∗mr): Can be proved by combining the proving methods for (s∗rr) and (s∗er). (s∗u): Suppose P & Q = P & R. Then γ(P & Q) = γ(P & R). If P & Q = P & R = ∅, then by the definition of slp-revision functions P ∗Q = P ∪Q and P ∗R = P ∪R. Thus P \(P ∗Q) = P \(P ∗R) = ∅ and (P ∗ Q) \ (P ∪ Q) = (P ∗ R) \ (P ∪ R) = ∅. So suppose P & Q = ∅ and P & R = ∅. Let X = γ(P & Q) = γ(P & R). By the definition of slp-revision functions, we have P \ (P ∗ Q) = P \ (X ∪ Q). By basic set theory we have P \ (X ∪ Q) = P \ (P ∩ (X ∪ Q)) = P \ ((P ∩ X) ∪ (P ∩ Q)). Since we have by Lemma 1 that P ∩ Q ⊆ X, P \ ((P ∩ X) ∪ (P ∩ Q)) = P \ (P ∩ X). Again by bas P \ (P ∩ X) = P \ X. Thus P \ (P ∗ Q) = P \ X. It can be shown in the same manner that P \ (P ∗ R) = P \ X. Thus P \ (P ∗ Q) = P \ (P ∗ R). Since we have by Lemma 1 that X ∩ (Q \ P ) = ∅, it follows from basic set theory that (X ∪ Q) \ (P ∪ Q) = X \ P . This means (P ∗ Q) \ (P ∪ Q) = X \ P . It can be shown in the same manner that (P ∗ R) \ (P ∪ R) = X \ P . Thus (P ∗ Q) \ (P ∪ Q) = (P ∗ R) \ (P ∪ R). ⇐: Suppose ∗ : P × P #→ P is a function that satisfies (s∗s), (s∗c), (s∗f), (s∗rr), (s∗er), (s∗mr), and (s∗u). We need to show ∗ is a slp-revision function. Let γ be defined as: γ(P & Q) = ((P ∗ Q) ∩ P ) ∪ ((P ∗ Q) \ Q) for all Q. It suffices to show P ∗ Q = γ(P & Q) ∪ Q and γ is a selection function for P . Part 1: By the definition of γ, we have γ(P & Q)∪Q = ((P ∗Q)∩ P ) ∪ ((P ∗ Q) \ Q) ∪ Q. By basic set theory, ((P ∗ Q) ∩ P ) ∪ ((P ∗ Q) \ Q) ∪ Q = (P ∗ Q) ∪ Q. Since Q ⊆ P ∗ Q follows from (s∗s), we have (P ∗ Q) ∪ Q = (P ∗ Q). Thus (P ∗ Q) = γ(P & Q) ∪ Q. Part 2: We first show γ is well defined. Suppose P & Q = P & R. Then (s∗u) implies P \(P ∗Q) = P \(P ∗R) and (P ∗Q)\(P ∪Q) = (P ∗ R) \ (P ∪ R). Since P \ (P ∗ Q) = P \ (P ∩ (P ∗ Q)) and P \ (P ∗ R) = P \ (P ∩ (P ∗ R)), P \ (P ∗ Q) = P \ (P ∗ R) implies (P ∗ Q) ∩ P = (P ∗ R) ∩ P . Thus (P ∗ Q) \ (P ∪ Q) = (P ∗ R) \ (P ∪ R) implies ((P ∗ Q) ∩ P ) ∪ ((P ∗ Q) \ (P ∪ Q)) = ((P ∗ R) ∩ P ) ∪ ((P ∗ R) \ (P ∪ R)). Then by basic set theory, we have ((P ∗Q)∩P )∪((P ∗Q)\Q) = ((P ∗R)∩P )∪((P ∗R)\R). Thus, γ(P & Q) = γ(P & R). So γ is a function. If P & Q = ∅, then we have to show γ(P & Q) = P . P & Q = ∅ implies Q is m-inconsistent, hence it follows from (s∗f) that P ∗Q = P ∪ Q. Then by the definition of γ, γ(P & Q) = ((P ∗ Q) ∩ P ) ∪ ((P ∗ Q) \ Q) = ((P ∪ Q) ∩ P ) ∪ ((P ∪ Q) \ Q) = P . If P & Q = ∅, then we have to show γ(P & Q) ∈ P & Q. Since P & Q = ∅, Q is m-consistent. Then (s∗c) implies P ∗ Q is consistent. Since P ∗ Q = γ(P & Q) ∪ Q, γ(P & Q) ∪ Q is consistent. Assume there is X such that X ∪ Q is consistent and X " P ⊂ γ(P & Q) " P . Then we have three cases: Case 1, X = γ(P & Q) ∪ R for R = ∅ and R ⊆ P \ γ(P & Q):.

(11) Z. Zhuang et al. / Reconsidering AGM-Style Belief Revision in the Context of Logic Programs. 677. If R ∩ Q = ∅, then since γ(P & Q) ∪ Q = P ∗ Q, R ∩ (P ∗ Q) = ∅. Then it follows from (s∗rr) that (P ∗ Q) ∪ R is inconsistent. Since X ∪ Q = (P ∗ Q) ∪ R, X ∪ Q is inconsistent, a contradiction! If R ∩ Q = ∅, then since R ⊆ P , R ∩ P ∩ Q = ∅. Since (s∗s) implies Q ⊆ P ∗ Q, we have Q ∩ P ⊆ (P ∗ Q) ∩ P , which implies R ∩ ((P ∗ Q) ∩ P ) = ∅. Then since ((P ∗ Q) ∩ P ) ⊆ γ(P & Q), γ(P & Q)∩R = ∅, a contradiction! Thus R∩Q = ∅ is an impossible case. Case 2, X = γ(P & Q) \ E for E ∩ P = ∅, E = ∅, and E ⊆ γ(P & Q): Then since γ(P & Q) ∪ Q = P ∗ Q, E ⊆ P ∗ Q. If E ∩ Q = ∅, then (s∗er) implies (P ∗ Q) \ E is inconsistent. Since X ∪Q = (γ(P & Q)\E)∪Q = (P ∗Q)\E, X ∪Q is inconsistent, a contradiction! If E ∩Q = ∅, then E ⊆ (P ∗Q)\Q. Since E ∩P = ∅, we have E∩(P ∗Q)∩P = ∅. Thus E ⊆ ((P ∗Q)∩P )∪((P ∗Q)\Q). Since ((P ∗ Q) ∩ P ) ∪ ((P ∗ Q) \ Q) = γ(P & Q), we have E ⊆ γ(P & Q), a contradiction! So E ∩ Q = ∅ is an impossible case. Case 3, X = (γ(P & Q) ∪ R) \ E for R = ∅, R ⊆ P \ γ(P & Q), E ∩ P = ∅, E = ∅, E ⊆ γ(P & Q): Then we can show as for Case 1 and 2 that R ∩ P ∗ Q = ∅ and E ⊆ P ∗ Q. If R ∩ Q = ∅ and E ∩ Q = ∅, then (s∗mr) implies ((P ∗ Q) ∪ R) \ E is inconsistent. Thus X ∪ Q = ((γ(P & Q) ∪ R) \ E) ∪ Q = ((P ∗ Q) ∪ R) \ E is inconsistent, a contradiction! Also we can show as for Case 1 and 2 that that R ∩ Q = ∅ and E ∩ Q = ∅ are impossible cases.. Let SE(S) = SE(P ) \ M . (Y, Y ) ∈ SE(S) ∩ SE(Q). Assume S ∪ Q is inconsistent. Then there is (X, Y ) ∈ SE(S) ∩ SE(Q) with X ⊂ Y . But (X, Y ) ∈ M , so (X, Y ) ∈ SE(S), a contradiction! So S ∪ Q is consistent. Assume SE(R) = SE(P ) \ M . Then SE(R) = (SE(P ) \ M ) \ G, SE(R) = (SE(P ) \ M ) ∪ H, or SE(R) = ((SE(P ) \ M ) \ G) ∪ H) where G = ∅, G ⊆ SE(P ) \ M , H = ∅, and H ∩ SE(P ) = ∅. For any of the possibilities, we have SE(P ) " SE(S) ⊂ SE(P ) " SE(R), which means S ∪ Q is inconsistent, a contradiction! So SE(R) = SE(P ) \ M . Assume there is (W, W ) ∈ SE(P ) ∩ SE(Q) such that W ⊂ Y . Let SE(S) = SE(P ) \ M  where M  = {(X, Z) ∈ SE(P ) | (X, W ) ∈ SE(Q) ∩ SE(P ), X = W and Z ⊆ W }. Then we can show as above, S ∪ Q is consistent. By the completeness of SE models, it follows from (X, W ) ∈ SE(Q) ∩ SE(P ) and W ⊂ Y that (X, Y ) ∈ SE(Q) ∩ SE(P ). Then it is easy to see that M  ⊂ M . Thus SE(P ) " SE(S) ⊂ SE(P ) " SE(R) which implies S ∪ Q is inconsistent, a contradiction! So there is no such (W, W ) ∈ SE(P ) ∩ SE(Q). This case corresponds to condition condition 2. Case 3, P ∪ Q is consistent: If SE(R) = SE(P ), then SE(P ) " SE(P ) ⊂ SE(P ) " SE(R) which implies P ∪ Q is inconsistent, a contradiction! Thus SE(R) = SE(P ). This case corresponds to condition 3.. Proof for Theorem 2 Suppose R ∈ P ' Q. Then R ∪ Q is consistent and SE(P ) " SE(S) ⊂ SE(P ) " SE(R) implies S ∪ Q is inconsistent. Since R ∪ Q is consistent, there is (Y, Y ) ∈ SE(R) ∩ SE(Q) and there is no (X, Y ) ∈ SE(R) ∩ SE(Q) with X ⊂ Y . We have three cases: Case 1, P ∪ Q is inconsistent and (Y, Y ) ∈ SE(Q) \ SE(P ): Let M = {(X, Z) ∈ SE(P ) | (X, Y ) ∈ SE(Q) and Z ⊂ Y } and SE(S) = Cl((SE(P ) ∪ {(Y, Y )}) \ M ). Then (Y, Y ) ∈ SE(S) ∩ SE(Q). Assume there is (X, Y ) ∈ SE(S) ∩ SE(Q) such that X ⊂ Y . If (X, Y ) ∈ SE(P ), then (Y, Y ) ∈ SE(P ), a contradiction! Thus (X, Y ) ∈ SE(Q) \ SE(P ) which means there is (X, Z) ∈ SE(S)∩SE(P ) such that Z ⊆ Y . But such (X, Z) is in M and thus (X, Z) ∈ SE(S), a contradiction. So there is no (X, Y ) ∈ SE(S)∩ SE(Q) such that X ⊂ Y , which implies S ∪ Q is consistent. Assume SE(R) = Cl((SE(P ) ∪ {(Y, Y )}) \ M ). Then SE(R) = Cl(((SE(P ) ∪ {(Y, Y )}) \ M ) \ G), SE(R) = Cl(((SE(P )∪{(Y, Y )})\M )∪N ), or SE(R) = Cl((((SE(P )∪ {(Y, Y )}) \ M ) \ G) ∪ N ) where G = ∅, G ⊆ SE(P ) \ M , N = ∅, and N ∩ SE(P ) = ∅. Note that since (Y, Y ) ∈ SE(R), SE(R) ∩ M = ∅ for otherwise R ∪ Q is inconsistent. For any of the possibilities, we have by basic set theory that SE(P ) " SE(S) ⊂ SE(P ) " SE(R), which means S ∪ Q is inconsistent, a contradiction! So SE(R) = Cl((SE(P ) ∪ {(Y, Y )}) \ M ). This case corresponds to condition 1. Case 2, P ∪ Q is inconsistent and (Y, Y ) ∈ SE(Q) ∩ SE(P ): Let M = {(X, Z) ∈ SE(P ) | (X, Y ) ∈ SE(Q) ∩ SE(P ), X = Y and Z ⊆ Y }. Assume SE(P )\M = Cl(SE(P )\M ). Then there is (W, Z) ∈ M such that (X, Z) ∈ Cl(SE(P ) \ M ). If W = Z, then there is (V, Z) ∈ SE(P ) \ M such that V ⊂ Z Since (W, Z) ∈ M , there is (X, Y ) ∈ SE(P ) ∩ SE(Q) such that X = Y and Z ⊆ Y which implies (V, Z) ∈ M , a contradiction! If W ⊂ Z, then there are (W, V ), (Z, Z) ∈ SE(P ) \ M such that V ⊆ Z. Again (W, Z) ∈ M implies there is (X, Y ) ∈ SE(P ) ∩ SE(Q) such that X = Y and Z ⊆ Y which implies (W, V ), (Z, Z) ∈ M , a contradiction! So we have SE(P ) \ M = Cl(SE(P ) \ M ).. Proof for Theorem 3 ⇒: Suppose ∗ is a llp-revision function. Let γ be the selection function for P . We need to show ∗ satisfies (l∗cl), (l∗s), (l∗c), (l∗f), (l∗rr), (l∗er), (l∗mr), and (l∗u). (l∗cl), (l∗s), (l∗f), and (l∗c) follow immediately from the definition of llp-revision functions. (l∗rr): Let O = ∅ and O ⊆ (SE(P ) ∩ SE(Q)) \ SE(P ∗ Q). We need to show SE(P ∗ Q) ∪ O is inconsistent. By the definition of llp-revision functions, we have P ∗ Q = cl(γ(P ' Q) ∪ Q). Let γ(P ' Q) = X. Then we have SE(P ∗ Q) = SE(X) ∩ SE(Q). According to Theorem 2, there are three cases: Case 1, P ∪ Q is inconsistent and SE(X) = Cl((SE(P ) ∪ {(Y, Y )}) \ M ) for M = {(X, Z) ∈ SE(P ) | (X, Y ) ∈ SE(Q) and Z ⊆ Y }: Then we have SE(P ∗ Q) = Cl((SE(P ) ∪ {(Y, Y )})\M )∩SE(Q). Since O ⊆ (SE(P )∩SE(Q))\SE(P ∗ Q), we have O ⊆ M . So there is (X, Y ) ∈ Cl((SE(P ∗ Q) ∪ O) such that X ⊂ Y which means (Y, Y ) is no longer an answer set in Cl(SE(P ∗ Q) ∪ O) and Cl(SE(P ∗ Q) ∪ O) is inconsistent. Case 2, P ∪ Q is inconsistent but m-consistent and SE(X) = SE(P ) \ M for there is (Y, Y ) ∈ SE(P ) ∩ SE(Q) such that (W, W ) ∈ SE(P ) ∩ SE(Q) implies W ⊂ Y and M = {(X, Z) ∈ SE(P ) | (X, Y ) ∈ SE(Q) ∩ SE(P ), X = Y and Z ⊆ Y }: Then SE(P ∗ Q) = (SE(P ) ∩ SE(Q)) \ M . Since O ⊆ (SE(P ) ∩ SE(Q)) \ SE(P ∗ Q), we have O ⊆ M . So there is (X, Y ) ∈ Cl(SE(P ∗ Q) ∪ O) such that X ⊂ Y which means (Y, Y ) is no longer an answer set in Cl(SE(P ∗Q)∪O) and Cl(SE(P ∗Q)∪O) is inconsistent. Case 3, P ∪ Q is consistent and SE(P ∗ Q) = SE(P ) ∩ SE(Q): Then SE(X) = SE(P ) and SE(P ∗Q) = SE(P )∩SE(Q) which means O = ∅, contradiction! So this is an impossible case. (l∗er): Let O = ∅ and O ⊆ SE(P ∗Q)\SE(P ). We need to show SE(P ∗Q)\O is inconsistent or Cl(SE(P ∗Q)\O) = SE(P ∗Q). Let γ(P ' Q) = X. Then we have SE(P ∗Q) = SE(X)∩SE(Q). According to Theorem 2, there are three cases: Case 1: As for (l∗rr) we have SE(P ∗ Q) = Cl((SE(P ) ∩ SE(Q)) \ M ∪ {(Y, Y )}). If (Y, Y ) ∈ O, then SE(P ∗ Q) \ O no longer contains any answer set and SE(P ∗ Q) \ O is inconsis-.

Riferimenti

Documenti correlati

Fig. 10 shows the median evolution of some selected proper- ties as a function of lookback time. In this figure, central galax- ies are divided according to their final stellar mass,

We establish the existence of solvable Lie groups of dimension 4 and left- invariant Riemannian metrics with zero Bach tensor which are neither confor- mally Einstein nor

 The takeOrdered(num, key) action returns a local python list of objects containing the num smallest elements of the considered RDD sorted by. considering a user specified

Based on the big data background, the paper uses the k-means clustering method to divide the load curve of 2,498 residential power users into 5 categories.. On the

Nell’ambito della realizzazione del modulo B del progetto di formazione Stigeac mi è stato affidato il tutoraggio dei quattro borsisti, che ha riguardato il processing

After three days of formula milk nutrition, urine had higher levels of glucose, galactose, glycine and myo-inositol, while up-regulated aconitic acid, aminomalonic acid and adipic

In this work we present NuChart-II, a software that allows the user to annotate and visualize a list of input genes with information relying on Hi-C data, integrating knowledge

L’idea di organizzare un convegno sulla “forma breve” è nata, da un lato, dalla rilevazione della costante presen- za della forma breve in ogni epoca e ambiente culturale,