Lecture Notes in Computer Science Commenced Publication in 1973 Founding and Former Series Editors: Gerhard Goos, Juris ...

This content was uploaded by our users and we assume good faith they have the permission to share this book. If you own the copyright to this book and it is wrongfully on our website, we offer a simple DMCA procedure to remove your content from our site. Start by pressing the button below!

Lecture Notes in Computer Science Commenced Publication in 1973 Founding and Former Series Editors: Gerhard Goos, Juris Hartmanis, and Jan van Leeuwen

Editorial Board David Hutchison Lancaster University, UK Takeo Kanade Carnegie Mellon University, Pittsburgh, PA, USA Josef Kittler University of Surrey, Guildford, UK Jon M. Kleinberg Cornell University, Ithaca, NY, USA Friedemann Mattern ETH Zurich, Switzerland John C. Mitchell Stanford University, CA, USA Moni Naor Weizmann Institute of Science, Rehovot, Israel Oscar Nierstrasz University of Bern, Switzerland C. Pandu Rangan Indian Institute of Technology, Madras, India Bernhard Steffen University of Dortmund, Germany Madhu Sudan Massachusetts Institute of Technology, MA, USA Demetri Terzopoulos University of California, Los Angeles, CA, USA Doug Tygar University of California, Berkeley, CA, USA Moshe Y. Vardi Rice University, Houston, TX, USA Gerhard Weikum Max-Planck Institute of Computer Science, Saarbruecken, Germany

4374

James F. Peters Andrzej Skowron Ivo Düntsch Jerzy Grzymała-Busse Ewa Orłowska Lech Polkowski (Eds.)

Transactions on Rough Sets VI Commemorating the Life and Work of Zdzisław Pawlak, Part I

13

Editors-in-Chief James F. Peters University of Manitoba, Winnipeg, Manitoba R3T 5V6, Canada E-mail: [email protected] Andrzej Skowron Warsaw University, Banacha 2, 02-097 Warsaw, Poland E-mail: [email protected] Volume Editors Ivo Düntsch Brock University, St. Catharines, Ontario L2S 3A1, Canada E-mail: [email protected] Jerzy Grzymała-Busse University of Kansas, Lawrence, KS 66045, USA E-mail: [email protected] Ewa Orłowska National Institute of Telecommunications, ul. Szachowa 1, 04-894 Warsaw, Poland E-mail: [email protected] Lech Polkowski University of Warmia and Mazury and Polish-Japanese Institute of Information Technology Warsaw 10560 Olsztyn, Poland E-mail: [email protected] Library of Congress Control Number: 2007922187 CR Subject Classification (1998): F.4.1, F.1, I.2, H.2.8, I.5.1, I.4 LNCS Sublibrary: SL 1 – Theoretical Computer Science and General Issues ISSN 0302-9743 (Lecture Notes in Computer Science) ISSN 1861-2059 (Transactions on Rough Sets) ISBN-10 3-540-71198-8 Springer Berlin Heidelberg New York ISBN-13 978-3-540-71198-8 Springer Berlin Heidelberg New York This work is subject to copyright. All rights are reserved, whether the whole or part of the material is concerned, specifically the rights of translation, reprinting, re-use of illustrations, recitation, broadcasting, reproduction on microfilms or in any other way, and storage in data banks. Duplication of this publication or parts thereof is permitted only under the provisions of the German Copyright Law of September 9, 1965, in its current version, and permission for use must always be obtained from Springer. Violations are liable to prosecution under the German Copyright Law. Springer is a part of Springer Science+Business Media springer.com © Springer-Verlag Berlin Heidelberg 2007 Printed in Germany Typesetting: Camera-ready by author, data conversion by Scientific Publishing Services, Chennai, India Printed on acid-free paper SPIN: 12028375 06/3142 543210

Preface

Volume VI of the Transactions on Rough Sets (TRS) commemorates the life and work of Zdzislaw Pawlak (1926-2006)1. His legacy is rich and varied. Professor Pawlak’s research contributions have had far-reaching implications inasmuch as his works are fundamental in establishing new perspectives for scientiﬁc research in a wide spectrum of ﬁelds. From a very early age, Zdzislaw Pawlak devoted his life to scientiﬁc research. The pioneering work by Prof. Pawlak included research on the design of computers, information retrieval, modeling conﬂict analysis and negotiation, genetic grammars, and molecular computing. His research led to the introduction of knowledge representation systems during the early 1970s and the discovery of rough sets during the early 1980s. Added to that was Prof. Pawlak’s lifelong interest in painting, photography, and poetry. During his lifetime, he nurtured worldwide interest in approximation, approximate reasoning, and rough set theory and its applications2 . Evidence of the inﬂuence of Prof. Pawlak’s work can be seen in the growth in the rough-set literature that now includes over 4000 publications by more than 1600 authors in the rough set database3 as well as the growth and maturity of the International Rough Set Society4 . Numerous biographies of Zdzislaw Pawlak have been published5 . This volume of the TRS presents papers that reﬂect the profound inﬂuence of a number of research initiatives by Zdzislaw Pawlak. In particular, this volume introduces a number of new advances in the foundations and applications of artiﬁcial intelligence, engineering, logic, mathematics, and science. These advances have signiﬁcant implications in a number of research areas such as the foundations of rough sets, approximate reasoning, bioinformatics, computational intelligence, cognitive science, data mining, information systems, intelligent systems, machine intelligence, and security. In addition, it is evident from the papers included in this volume that rough set theory and its application form a very active research area worldwide. A total of 41 researchers from 8 countries are represented in this volume, namely, Canada, India, France, Norway, Poland, P.R.

1 2

3 4 5

Prof. Pawlak passed away on 7 April 2006. See, e.g., Pawlak, Z., Skowron, A.: Rudiments of rough sets, Information Sciences 177 (2007) 3-27; Pawlak, Z., Skowron, A.: Rough sets: Some extensions, Information Sciences 177 (2007) 28-40; Pawlak, Z., Skowron, A.: Rough sets and Boolean reasoning, Information Sciences 177 (2007) 41-73. http://rsds.wsiz.rzeszow.pl/rsds.php http://roughsets.home.pl/www/ See, e.g., Peters, J.F. and Skowron, A., Zdzislaw Pawlak: Life and Work. Transactions on Rough Sets V, LNCS 4100 (2006) 1-24. See, also, R. Slowi´ nski, Obituary, Prof. Zdzislaw Pawlak (1926-2006), Fuzzy Sets and Systems 157 (2006) 2419-2422.

VI

Preface

China, Sweden, Russia, Thailand, and the USA. Evidence of the vigor, breadth and depth of research in the theory and applications of rough sets can be found in the articles in this volume. Most of the contributions of this commemorative volume of the TRS are on an invitational basis and every paper has been refereed in the usual way. This special issue of the TRS contains 23 papers and extended abstracts that explore a number of research streams that are either directly or indirectly related to research initiatives by Zdzislaw Pawlak. These research streams are represented by papers on propositional logics (Mohua Banerjee and Md. Aquil Khan), intuitionistic rough sets for database applications (Theresa Beaubouef and Fred Petry), missing attribute value problem (Jerzy W. Grzymala-Busse and Witold J. Grzymala-Busse), Zdzislaw Pawlak’s contributions to the study of vagueness (Mihir Chakraborty), data mining (Alicja Wakulicz-Deja and Grzegorz Ilczuk), approximation of concepts (Anna Gomoli´ nska), intelligent systems (Andrzej Jankowski and Andrzej Skowron), acoustics (Bozena Kostek), rule evaluation (Jiye Li, Puntip Pattaraintakorn, and Nick Cercone), rough sets in China (Qing Liu and Hui Sun), four-valued logic (Jan Maluszy´ nski, Andrzej Szalas and Aida Vit´ oria), crisp and fuzzy information systems (Alicja Mieszkowicz-Rolka and Leszek Rolka), artiﬁcial intelligence and rough sets (Tosiharu Munakata), topology and information systems (Piero Pagliani and Mihir K. Chakraborty), conjugate information systems (Maria Semeniuk-Polkowska), incomplete transactional databases (Grzegorz Protaziuk and Henryk Rybinski), classiﬁers, rule induction and rough sets (Jerzy Stefanowski), approximation spaces (Jaroslaw Stepaniuk), relevant attributes in high-dimensional data (Julio J. Vald´es and Alan J. Barton), knowledge discovery in databases (Anita Wasilewska, Ernestina Menasalvas, Christelle Scharﬀ), information quanta and approximation operators (Marcin Wolski), lattice theory for rough sets (Jouni J¨ arvinen). The editors of this volume extend their hearty thanks to reviewers of papers that have been submitted to the TRS during the past 12 months: Manuel OjedaAciego, Mohua Banerjee, Jan Bazan, Mihir Chakraborty, Anna Gomoli´ nska, Etienne Kerre, Pawan Lingras, Victor Marek, Piero Pagliani, Sheela Ramanna, ´ ezak, Jerzy Stefanowski, Jaroslaw Stepaniuk, Piotr Synak, Piotr Dominik Sl¸ Wasilewski and Yiyu Yao. This issue of the TRS has been made possible thanks to the laudable eﬀorts of a great many generous persons and organizations. The editors and authors of this volume also extend an expression of gratitude to Alfred Hofmann, Ursula Barth, Christine G¨ unther and the LNCS staﬀ at Springer for their support in making this volume of the TRS possible. In addition, the editors extend their thanks to Marcin Szczuka for his consummate skill and care in the compilation of this volume. The editors have been supported by the State Committee for Scientiﬁc Research of the Republic of Poland (KBN),

Preface

VII

research grant No. 3T11C00226, and the Natural Sciences and Engineering Research Council of Canada (NSERC) research grant 185986. December 2006

Ivo D¨ untsch Jerzy W. Grzymala-Busse Ewa Orlowska James F. Peters Lech Polkowski Andrzej Skowron

LNCS Transactions on Rough Sets

This journal subline has as its principal aim the fostering of professional exchanges between scientists and practitioners who are interested in the foundations and applications of rough sets. Topics include foundations and applications of rough sets as well as foundations and applications of hybrid methods combining rough sets with other approaches important for the development of intelligent systems. The journal includes high-quality research articles accepted for publication on the basis of thorough peer reviews. Dissertations and monographs up to 250 pages that include new research results can also be considered as regular papers. Extended and revised versions of selected papers from conferences can also be included in regular or special issues of the journal. Honorary Editor: Editors-in-Chief:

Zdzislaw Pawlak – deceased James F. Peters, Andrzej Skowron

Editorial Board M. Beynon G. Cattaneo M.K. Chakraborty A. Czy˙zewski J.S. Deogun D. Dubois I. D¨ untsch S. Greco J.W. Grzymala-Busse M. Inuiguchi J. Jrvinen D. Kim J. Komorowski C.J. Liau T.Y. Lin E. Menasalvas M. Moshkov T. Murai

M. do C. Nicoletti H.S. Nguyen S.K. Pal L. Polkowski H. Prade S. Ramanna R. Slowi´ nski J. Stefanowski J. Stepaniuk Z. Suraj ´ R. Swiniarski M. Szczuka S. Tsumoto G. Wang Y. Yao N. Zhong W. Ziarko

Table of Contents

Contributed Papers Propositional Logics from Rough Set Theory . . . . . . . . . . . . . . . . . . . . . . . . Mohua Banerjee and Md. Aquil Khan

1

Intuitionistic Rough Sets for Database Applications . . . . . . . . . . . . . . . . . . Theresa Beaubouef and Frederick E. Petry

26

An Experimental Comparison of Three Rough Set Approaches to Missing Attribute Values . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Jerzy W. Grzymala-Busse and Witold J. Grzymala-Busse

31

Pawlak’s Landscaping with Rough Sets . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Mihir K. Chakraborty

51

A Comparison of Pawlak’s and Skowron–Stepaniuk’s Approximation of Concepts . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Anna Gomoli´ nska

64

Data Preparation for Data Mining in Medical Data Sets . . . . . . . . . . . . . . Grzegorz Ilczuk and Alicja Wakulicz-Deja

83

A Wistech Paradigm for Intelligent Systems . . . . . . . . . . . . . . . . . . . . . . . . . Andrzej Jankowski and Andrzej Skowron

94

The Domain of Acoustics Seen from the Rough Sets Perspective . . . . . . . Bozena Kostek

133

Rule Evaluations, Attributes, and Rough Sets: Extension and a Case Study . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Jiye Li, Puntip Pattaraintakorn, and Nick Cercone

152

The Impact of Rough Set Research in China: In Commemoration of Professor Zdzislaw Pawlak . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Qing Liu and Hui Sun

172

A Four-Valued Logic for Rough Set-Like Approximate Reasoning . . . . . . Jan Maluszy´ nski, Andrzej Szalas, and Aida Vit´ oria

176

On Representation and Analysis of Crisp and Fuzzy Information Systems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Alicja Mieszkowicz-Rolka and Leszek Rolka

191

On Partial Covers, Reducts and Decision Rules with Weights . . . . . . . . . . Mikhail Ju. Moshkov, Marcin Piliszczuk, and Beata Zielosko

211

XII

Table of Contents

A Personal View on AI, Rough Set Theory and Professor Pawlak . . . . . . Toshinori Munakata

247

Formal Topology and Information Systems . . . . . . . . . . . . . . . . . . . . . . . . . . Piero Pagliani and Mihir K. Chakraborty

253

On Conjugate Information Systems: A Proposition on How to Learn Concepts in Humane Sciences by Means of Rough Set Theory . . . . . . . . . Maria Semeniuk–Polkowska

298

Discovering Association Rules in Incomplete Transactional Databases . . . Grzegorz Protaziuk and Henryk Rybinski

308

On Combined Classiﬁers, Rule Induction and Rough Sets . . . . . . . . . . . . . Jerzy Stefanowski

329

Approximation Spaces in Multi Relational Knowledge Discovery . . . . . . . Jaroslaw Stepaniuk

351

Finding Relevant Attributes in High Dimensional Data: A Distributed Computing Hybrid Data Mining Strategy . . . . . . . . . . . . . . . . . . . . . . . . . . . Julio J. Vald´es and Alan J. Barton

366

A Model PM for Preprocessing and Data Mining Proper Process . . . . . . . Anita Wasilewska, Ernestina Menasalvas, and Christelle Scharﬀ

397

Monographs Lattice Theory for Rough Sets . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Jouni J¨ arvinen

400

Author Index . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .

499

Propositional Logics from Rough Set Theory Mohua Banerjee and Md. Aquil Khan Department of Mathematics and Statistics, Indian Institute of Technology, Kanpur 208 016, India {mohua,mdaquil}@iitk.ac.in

Abstract. The article focusses on propositional logics with semantics based on rough sets. Many approaches to rough sets (including generalizations) have come to the fore since the inception of the theory, and resulted in diﬀerent “rough logics” as well. The essential idea behind these logics is, quite naturally, to interpret well-formed formulae as rough sets in (generalized) approximation spaces. The syntax, in most cases, consists of modal operators along with the standard Boolean connectives, in order to reﬂect the concepts of lower and upper approximations. Non-Boolean operators make appearances in some cases too. Information systems (“complete” and “incomplete”) have always been the “practical” source for approximation spaces. Characterization theorems have established that a rough set semantics based on these “induced” spaces, is no diﬀerent from the one mentioned above. We also outline some other logics related to rough sets, e.g. logics of information systems – which, in particular, feature expressions corresponding to attributes in their language. These systems address various issues, such as the temporal aspect of information, multiagent systems, rough relations. An attempt is made here to place this gamut of work, spread over the last 20 years, in one platform. We present the various relationships that emerge and indicate questions that surface.

1

Introduction

A “logic of rough sets” would, in the natural sense, represent a formal system, statements in the language of which would be interpreted as rough sets in some approximation space. Thus “models” in the semantics of such a system would be approximation spaces, equipped with a meaning function that assigns rough sets to well-formed formulae (wﬀs) of the language. Rough sets have been deﬁned in more than one way for a Pawlak approximation space (X, R) – [1] lists ﬁve deﬁnitions, all of which are equivalent to each other. One of these is most commonly used: (*) a rough set in (X, R), is the pair (A, A), for each A ⊆ X, where A, A denote the lower and upper approximations of A respectively. Another is a deﬁnition given by Pawlak in [2], and of interest to us in this paper:

The author acknowledges the support of the Department of Computer Science, University of Regina, Canada, during a visit to which the paper was ﬁnalized.

J.F. Peters et al. (Eds.): Transactions on Rough Sets VI, LNCS 4374, pp. 1–25, 2007. c Springer-Verlag Berlin Heidelberg 2007

2

M. Banerjee and Md.A. Khan

(**) A ⊆ X is a rough set in (X, R), provided the boundary of A, BnA = ∅. For generality’s sake, we could remove the restriction in (**) and consider deﬁnable sets (i.e. subsets with empty boundary) as special cases of rough sets. Thus, in the semantics based on approximation spaces, the meaning function deﬁning models, assigns to wﬀs either subsets of the domain, or pairs of subsets in accordance with (*) [3,4,5,6,7,8,9,10]. This is true even for semantics based on generalized approximation spaces, where diﬀerent relations (may be more than one in number, with operations on them) are considered [6,11]. The logics invariably involve modalities to express the concepts of lower and upper approximations – some are simply known normal modal logics, or have non-Boolean connectives (and no modalities) in the language, but there are translations into modal logics. We make a study of this group of systems in Section 2. It may be remarked that the “rough logic” proposed by Pawlak [3] (the ﬁrst system to be called so) makes an appearance here (cf. Section 2.6). The “practical” source of Pawlak approximation spaces are complete / deterministic information systems. These have the form S ≡ (U, A, V al, f ), where U is a set of objects, A a set of attributes, V al a set of values for the attributes, and f a function from U × A to V al. An equivalence relation RS is induced on U (thus giving the approximation space (U, RS )), as x RS y in U , if and only if f (x, a) = f (y, a), for all a ∈ A. The converse also holds: given any approximation space (U, R), one can deﬁne an information system S ≡ (U, A, V al, f ) such that the induced equivalence RS is just the relation R. So, in eﬀect, a semantics based on approximation spaces induced by complete information systems, is identical to the one discussed above. Generalized information systems, termed incomplete/nondeterministic, are those where f is a function from U × A to P(V al), and yields diﬀerent kinds of binary relations (e.g. similarity, inclusion – cf. Section 3.1) apart from equivalences, on U . Thus any information system (complete or incomplete) on a domain U , induces a relational system or a (generalized) approximation space on U , i.e. the (non-empty) set U together with a set of binary relations. This is called a standard structure on U [12,13,14]. For example, for the complete information system (U, A, V al, f ) above, (U, RS ) is a standard structure on U . In Section 3.1, (U, simS , inS ) is a standard structure for the incomplete information system S ≡ (U, A, V al, f ), with similarity and inclusion relations simS , inS . (Diﬀerent sets of relations can give diﬀerent standard structures on the same set U .) The induced relations in the standard structure may be characterized by a set of properties. As we know, equivalences are characterized by the properties of reﬂexivity, symmetry and transitivity. The similarity and inclusion relations considered in Section 3.1 are characterized by the properties (S1), (S2), (S4) − (S6) given there. By a general structure on U [12,13,14], one means any relational system comprising a non-empty set, along with binary relations that satisfy the set of properties characterizing the induced relations in the standard structure. Again, for the complete information system (U, A, V al, f ) above, any Pawlak approximation space (U, R) is a general structure. A general structure for S of

Propositional Logics from Rough Set Theory

3

Section 3.1, would be of the form (U, sim, in), where sim, in are binary relations on U satisfying (S1), (S2), (S4) − (S6). One ﬁnds logics with semantics deﬁned on incomplete information systems, for instance, in [15], or with semantics deﬁned on general structures [16]. However, Vakarelov [12,13,14,17] has established a series of characterization results, enabling an identiﬁcation of semantics based on general and standard structures (as in case of the Pawlak approximation space and complete information system above). In case of [15] too, we demonstrate here that the logic in question is equivalent to a normal modal logic with certain generalized approximation spaces deﬁning models. These systems are discussed in Section 3. In another line, there are “logics of information systems”, which accommodate in their language, expressions corresponding to objects and attributes [18,19,4,20]. Amongst these is a system that addresses the temporal aspect of information (cf. [4]), while [20] presents a logic for multiagent systems. There are also treatises on “rough relations” – a logic has been proposed [21] on the one hand, and on the other, we have the proposal of a logic programming language in “rough datalog” [22]. In Section 4, we brieﬂy sketch these and other approaches, such as rough mereology [23]. It will be seen that, some of the logics [4,16,20] have atomic propositions as (or built from) descriptors, the key feature of decision logic [2]. Decision logic is well-known, and not presented in this article. One should mention that a few of the logics described here, have also been used as a base to express various concepts involving rough sets. For instance, Yao and Lin [6] have deﬁned graded and probabilistic rough sets, using graded and probabilistic modal operators in the language of normal modal systems. Common and distributed knowledge operators have been interpreted in generalized approximation spaces by Wong [24]. In [25], another modal system (inspired by [3]) has been used to propose postulates for rough belief change. A comparative study of the presented logics is made in Section 5. The paper concludes by indicating possible future directions of investigation in Section 6.

2

Logics with Semantics Based on Approximation Spaces

In this section, we take a look at logics with approximation spaces deﬁning models. We ﬁnd six kinds of systems. For a logic L, “α is a theorem of L” shall be indicated by the notation L α. 2.1

Normal Modal Systems

The modal nature of the lower and upper approximations of rough sets was evident from the start. Hence, it is no surprise that normal modal systems were focussed upon, during investigations on logics for rough sets. In particular, in case of Pawlak rough sets, the two approximations considered as operators clearly obey all the S5 laws. The formal connection between the syntax of S5 and its semantics in terms of rough sets is given as follows [26]. According to the Kripke semantics for S5, a wﬀ α is interpreted by a function π as a subset in a non-empty domain U , the subset representing the extension

4

M. Banerjee and Md.A. Khan

of the formula – i.e. the collection of situations/objects/worlds where the wﬀ holds. Moreover, in an S5-model M ≡ (U, R, v) (say), the accessibility relation R is an equivalence on U . Further, if , ♦ denote the necessity and possibility operators respectively then for any wﬀ α, v(α) = v(α) and v(♦α) = v(α). A wﬀ α is true in M, if v(α) = U . Now it can easily be seen that all the S5 theorems involving and ♦ translate into valid properties of lower and upper approximations. Taking a cue from this connection, similar links have been pointed out (e.g. in [6,27]) between “rough sets” on generalized approximation spaces, and diﬀerent normal modal systems. The basic idea is to deﬁne generalized approximation operators corresponding to any binary relation R on the domain U – this has been done by many (e.g. for tolerance relations in [28] and others – cf. [29]). More explicitly, a map r : U → P(U ) is deﬁned as r(x) ≡ {y ∈ U : xRy}. Then the operators apr, apr : P(U ) → P(U ) are given by apr(A) ≡ {x : r(x) ⊆ A}, and apr(A) ≡ {x : r(x) ∩ A = ∅}. The rough set operators then satisfy various properties, depending upon the nature of R. Now let L denote a normal modal language, and M ≡ (U, R, v) be a model for L. v, as before, interprets a wﬀ as a subset in U . Then it is straightforward to observe that for any wﬀ α of L, v(α) = apr(v(α)), and dually, v(♦α) = apr(v(α)). By the above interpretation, the modal logics like KB, KT, K4, S5 etc. could be said to capture the properties of rough sets in generalized approximation spaces based on diﬀerent R (symmetric, reﬂexive, transitive, equivalence etc.). As remarked in the Introduction, this link has been made use of further. Considering graded and probabilistic modal operators on the above systems, graded and probabilistic rough sets have been deﬁned in [6]. Wong [24] has interpreted common and distributed knowledge operators (as deﬁned in logic of knowledge) in generalized approximation spaces with an indexed set of indiscernibility relations (corresponding to the knowledge operator of each agent). 2.2

DAL

[11] considers generalized approximation spaces containing a family of equivalence relations instead of just one. The logic DAL that is deﬁned in [11], has models based on these spaces. Further, the set of equivalence relations is assumed to be closed with respect to the operations of intersection and transitive closure of union of relations. The language of DAL, expectedly, includes a family of modal operators intended to correspond to the indiscernibility relations on the domains of the models. Formally, this is done by having a set R (say) of relational variables apart from the set P of propositional ones. There are binary operations ∩, , and a collection REL of relational expressions is built inductively out of the members of R with these operations. Apart from the classical Boolean connectives, a modal connective [R] is then introduced in the language for each R ∈ REL.

Propositional Logics from Rough Set Theory

5

A DAL-model is a structure U ≡ (U, {ρR }R∈REL , m), where, (i) for any R ∈ REL, ρR is an equivalence relation in the set U ; (ii) ρR∩S is the greatest equivalence relation in U included in both ρR and ρS ; (iii) ρRS is the least equivalence relation including both ρR and ρS ; and (iv) m is the meaning function from P ∪ R to P(U ) ∪ {ρR }R∈REL such that m(p) ⊆ U , for p ∈ P, and m(R) ≡ ρR , for R ∈ REL. For evaluating truth of wﬀs in DAL-models, one deﬁnes a function v that is determined by the meaning function m: v(p) ≡ m(p), for p ∈ P, v([R]α) ≡ {x ∈ U : y ∈ v(α), for all y such that x m(R) y}, the Boolean cases being deﬁned in the standard way. Deﬁnitions of truth and validity then are as usual: α is true in U, provided v(α) = U , and valid if it is true in all DAL-models. DAL has been axiomatized as follows. The connective is the dual of []. A1. All classical tautologies, A2. [R](α → β) → ([R]α → [R]β), A3. [R]α → α, A5. R α → [R] R α, A5. [R S]α → [R]α ∧ [S]α, A6. (([P ]α → [R]α) ∧ ([P ]α → [S]α)) → ([P ]α → [R S]α), A7. [R]α ∨ [S]α → [R ∧ S]α, A8. (([R]α → [P ]α) ∧ ([S]α → [P ]α)) → ([R ∧ S]α → [P ]α). The only rules of inference are Modus Ponens and Necessitation (corresponding to the connective [R] for each R ∈ REL). The axiomatization yields a completeness result with respect to the aforementioned semantics. Theorem 1. For any DAL-wﬀ α, DAL α, if and only if α is valid. 2.3

Pre-rough Logic

Following in the footsteps of Rasiowa, the algebra of rough sets was investigated in [7] in order to arrive at a logic for the theory. An algebraic structure called pre-rough algebra was proposed – this is a quasi Boolean algebra [30] along with a topological operator satisfying all the properties of an interior, and more. A corresponding logic P RL was framed, and observed to be sound and complete with respect to a semantics based on rough sets. The language of P RL has the primitive logical symbols ¬, , . , ♦ are duals of , , while ⇒ is deﬁned as: α ⇒ β ≡ (¬α β) (¬♦α ♦β), for any wﬀs α, β of P RL.

6

M. Banerjee and Md.A. Khan

As in the case of S5, a model for P RL is of the form M ≡ (U, R, v), where the departure from the S5-semantics lies in the deﬁnition of the meaning function v with respect to the connectives of conjunction and implication ⇒. For any α, β in P RL, S, T ⊆ U , v(α β) ≡ v(α) v(β), and v(α ⇒ β) ≡ v((¬α β) (¬♦α ♦β)), where c S T ≡ (S ∩ T ) ∪ (S ∩ T ∩ (S ∩ T ) ) (c denoting complementation). Deﬁnition of truth of a wﬀ α in M remains the same: this is if and only if v(α) = U . It may then be noticed that ⇒ reﬂects rough inclusion: a wﬀ α ⇒ β is true in (U, R, v) provided v(α) is roughly included in v(β). Further, / are operations that reduce to ordinary set intersection / union only when working on deﬁnable sets. α is valid (written |=RS α), if and only if α is true in every P RL-model. Following are the axiom schemes for P RL: 1. α ⇒ α 2a. ¬¬α ⇒ α 2b. α ⇒ ¬¬α 3. α β ⇒ α 4. α β ⇒ β α 5a. α (β γ) ⇒ (α β) (α γ) 5b. (α β) (α γ) ⇒ α (β γ) 6. α ⇒ α 7a. (α β) ⇒ (α) (β) 7b. (α) (β) ⇒ (α β) 8. α ⇒ α 9. ♦α ⇒ α 10a. (α β) ⇒ α β 10b. α β ⇒ (α β) Rules of inference : 1.

α 2. α⇒β α⇒β β⇒γ β α⇒γ modus ponens hypothetical syllogism

3.

α β⇒α

4.

α⇒β ¬β ⇒ ¬α

5.

α⇒β α⇒γ α⇒βγ

6.

α ⇒ β, β ⇒ α γ ⇒ δ, δ ⇒ γ (α ⇒ γ) ⇒ (β ⇒ δ)

7.

α⇒β α ⇒ β

8.

α α

9.

α ⇒ β ♦α ⇒ ♦β α⇒β

One can then prove, for any P RL-wﬀ α,

Propositional Logics from Rough Set Theory

7

Theorem 2. P RL α, if and only if |=RS α. We shall meet this logic and its semantics again in the coming sections. 2.4

3-Valued L ukasiewicz Logic L3

The connection of rough sets with 3-valuedness, also came up in the context of algebraic investigations. For example, in [31,32,33], an equivalence of 3-valued L ukasiewicz (Moisil) algebras with rough set structures was observed. In terms of logic, the way we can set up a formal link between the intensely studied L3 and a rough set semantics – in fact, the semantics just outlined in Section 2.3, is as follows. Let us recall Wajsberg’s axiomatization of L3 (cf. [34]). The logical symbols ¬, → are taken to be primitive. Axiom schemes: 1. 2. 3. 4.

α → (β → α). (α → β) → ((β → γ) → (α → γ)). ((α → ¬α) → α) → α. (¬α → ¬β) → (β → α).

The only rule of inference is Modus Ponens. L3 is known to be sound and complete with respect to the class of 3-valued L ukasiewicz (Moisil) algebras, as well as with respect to the semantics on 3 ≡ {0, 1/2, 1}, with L ukasiewicz negation and implication [34]. Now a logic L1 is said to be embeddable into a logic L2 , provided there is a translation of wﬀs of L1 into L2 , such that L1 α if and only if L2 α for any wﬀ α of L1 . We use the denotation L1 L2 . L1 L2 denotes existence of embeddings both ways. [31] establishes the following. There are translations ◦ from L3 into P RL and ∗ from P RL into L3 given by (¬α)◦ ≡ ¬α◦ , (α → β)◦ ≡ (♦¬α◦ β ◦ ) (♦β ◦ ¬α◦ ); (¬α)∗ ≡ ¬α∗ , (α β)∗ ≡ (α∗ → β ∗ ) → β ∗ , (α β)∗ ≡ ¬(¬α∗ ¬β ∗ ), (♦α)∗ ≡ ¬α∗ → α∗ . (One may notice that for any α, (α◦ )∗ and (α∗ )◦ are logically equivalent to α in the respective systems.) It is then shown that L3 P RL. Thus Theorem 3 (a) L3 α, if and only if |=RS α◦ , for an L3-wﬀ α and (b) L3 α∗ , if and only if |=RS α, for a P RL-wﬀ α.

8

2.5

M. Banerjee and Md.A. Khan

Logic for Regular Double Stone Algebras

Another line of algebraic investigation has resulted in linking rough set structures with the class of regular double Stone algebras [35]. A double Stone algebra (DSA) is a structure (L, , ,∗ ,+ , 0, 1) such that (L, , , 0, 1) is a bounded distributive lattice, y ≤ x∗ if and only if y x = 0, y ≥ x+ if and only if y x = 1 and x∗ x∗∗ = 1, x+ x++ = 0. The operations ∗ ,+ , as evident, are two kinds of complementation on the domain. The DSA is regular if, in addition to the above, for all x ∈ L, x x+ ≤ x x∗ holds. This is equivalent to requiring that x∗ = y ∗ , x+ = y + imply x = y, for all x, y ∈ L. Considering the deﬁnition (*) of rough sets (cf. Introduction), one ﬁnds that the collection RS of rough sets (X, X) over an approximation space (U, R) can be made into a regular DSA. The zero of the structure is the element (∅, ∅), while the unit is (U, U ). The operations , ,∗ ,+ are deﬁned as (X, X) (Y , Y ) ≡ (X ∪ Y , X ∪ Y ), (X, X) (Y , Y ) ≡ (X ∩ Y , X ∩ Y ), c c (X, X)∗ ≡ (X , X ), (X, X)+ ≡ (X c , X c ). For the converse, Comer shows that any regular DSA is isomorphic to a subalgebra of RS for some approximation space (U, R). untsch [8] as follows. Using these facts, a logic LD for rough sets is deﬁned by D¨ The language of LD has two unary connectives ∗ ,+ (for two kinds of negation), apart from the binary connectives ∨, ∧ and constant symbol . We write α∗ , α+ instead of ∗ α,+ α, just to keep parity with the algebraic notation used above. A model of LD is a pair (W, v), where W is a (non-empty) set and v is the meaning function assigning to propositional variables, pairs in P(W ) × P(W ) such that if v(p) = (A, B) then A ⊆ B. v(p) = (A, B) is to express that “p holds at all states of A and does not hold at any state outside B”. For , we have v() ≡ (W, W ). v is extended to the set of all wﬀs recursively: if v(α) = (A, B) and v(β) = (C, D) then v(α ∨ β) ≡ (A ∪ C, B ∪ D), v(α ∧ β) ≡ (A ∩ C, B ∩ D), v(α∗ ) ≡ (B c , B c ), v(α+ ) ≡ (Ac , Ac ). A wﬀ α is true in a model (W, v), provided v(α) = (W, W ). We would now like to make explicit, how v interprets the wﬀs of LD as rough sets over some approximation space. One refers to [8], and [35].

Propositional Logics from Rough Set Theory

9

Consider the range ran(v) of the map v in P(W ) × P(W ). It can be shown that it forms a regular DSA through the operations , ,∗ ,+ : v(α) v(β) ≡ v(α ∨ β), v(α) v(β) ≡ v(α ∧ β), v(α)∗ ≡ v(α∗ ), v(α)+ ≡ v(α+ ). v(∗ ) (or v(+ )) is the zero ((∅, ∅)) of the algebra, while v() = (W, W ) is the unit. In fact, the variety of regular DSA’s is just the one generated by regular DSA’s of the kind ran(v), where v ranges over all meaning functions for all models. Using the correspondence between classes of algebras and logic [36], [8] concludes, amongst other properties of LD , that Theorem 4. LD has a ﬁnitely complete and strongly sound Hilbert style axiom system. Through Comer’s representation result, ran(v) corresponding to any model (W, v) of LD , is isomorphic to a subcollection of RS for some approximation space (U, R). We can now say that v(α) for a wﬀ α, can be identiﬁed with a rough set over some (U, R) in precisely the following manner. Let U consist of all the join irreducible elements of ran(v), i.e. v(α) ∈ U , if and only if v(α) = (∅, ∅), and for all wﬀs β, γ, if v(α) = v(β) v(γ) then either v(α) = v(β) or v(α) = v(γ). An equivalence relation R on U can then be obtained, where R is given by: v(α) R v(β) if and only if v(α∗∗ ) = v(β ∗∗ ), i.e. if and only if B = D, where v(α) = (A, B) and v(β) = (C, D). Now deﬁne f : ran(v) → P(U ) such that for v(α) = (A, B), f (A, B) ≡ {v(β) = (C, D) ∈ U : C ⊆ A, D ⊆ B}. Finally, deﬁne the map g : ran(v) → P(U ) × P(U ) as: g(A, B) ≡ (f (A, A), f (B, B)), where v(α) = (A, B). (Note that (A, A), (B, B) ∈ U , as v(α++ ) = (A, A), and v(α∗∗ ) = (B, B).) It can then be shown that (a) g is injective, and (b) g preserves , ,∗ ,+ . Moreover, if v(α) = (A, B), g(v(α)) = ( f (A, B), f (A, B) ), a rough set in the approximation space (U, R). [8] does not present an explicit proof method for the logic LD – the only comment on the matter is vide Theorem 4. Recently, Dai [9] has presented a sequent calculus for a logic (denoted RDSL) with a semantics based on the regular DSAs formed by collections of rough sets of the kind RS over some approximation space (U, R) (deﬁned earlier in the section). The language of RDSL is the same as that of LD , except that the constant symbol ⊥ (dual for ) is included amongst the

10

M. Banerjee and Md.A. Khan

primitive symbols. Models are of the form (RS, v), where v, the meaning function, is a map from the set of propositional variables to RS. Thus v(p), for a propositional variable p, is a pair (X, X) in the approximation space (U, R). v is extended to the set of all wﬀs in the same way as for models of LD . We note that an RDSL-model (RS, v) may be identiﬁed with the LD -model (U, v). On the other hand, due to Comer’s representation result, given any LD -model (W, v), there is an isomorphism f from ran(v) to a subalgebra (S, say) of RS on some approximation space. One can thus ﬁnd an RDSLmodel (RS, v ) such that ran(v ) is S, i.e. v (p) ≡ f (v(p)), for every propositional variable p. So, in this sense, the classes of models of the two logics are identiﬁable. As in classical sequent calculus, for ﬁnite sequences of wﬀs Γ ≡ (p1 , p2 , . . . pm ) and Δ ≡ (q1 , q2 , . . . qn ) in RDSL, the sequent Γ ⇒ Δ is said to be valid in a model (RS, v) if and only if v(p1 ) . . . v(pm ) ≤ v(q1 ) . . . v(qn ). , are the operations in the regular DSA (RS, , ,∗ ,+ , < ∅, ∅ >, < U, U >). Γ ⇒ Δ is said to be valid (in notation, |=RDSA Γ ⇒ Δ) if and only if Γ ⇒ Δ is valid in every RDSL-model. The standard classical axiom p ⇒ p and rules for the connectives ∧, ∨ and constant symbols , ⊥ are considered to deﬁne derivability (RDSL ). In addition, the axioms and rules for the two negations ∗ ,+ are as follows. 1. p ⇒ p∗∗ . 2. p∗ ⇒ p∗∗∗ . 3. p ⇒ p++ . 4. p+ ⇒ p+++ . (R∗ ) Γ ⇒ Δ (R+ ) Γ ⇒ Δ Δ∗ ⇒ Γ ∗ Δ+ ⇒ Γ + Soundness and completeness are then proved, with respect to the semantics sketched. Theorem 5. RDSL Γ ⇒ Δ, if and only if |=RDSA Γ ⇒ Δ. 2.6

Logic for Rough Truth or of Rough Consequence

In [3], a logic Rl (the ﬁrst in literature to be called “rough logic”) was proposed, along with a very appealing notion of rough truth. The language of Rl consists of the standard Boolean connectives, and models M ≡ (U, R, v) are based on approximation spaces. v assigns subsets of the domain U to wﬀs in the usual manner. Five logical values of “truth”, “falsity”, “rough truth”, “rough falsity” and “rough inconsistency” are considered in this work, with truth and falsity representing the limit of our partial knowledge. As we know, a wﬀ α is true in M, if v(α) = U . α is said to be surely/possibly true on x ∈ U , if x ∈ v(α) (v(α)) respectively. α is roughly true in M, if it is possibly true on every x in U , i.e. v(α) = U , or in other words, v(α) is

Propositional Logics from Rough Set Theory

11

externally indiscernible [37] in (U, R). On the other hand, α is roughly false, when v(α) = ∅ (v(α) is internally indiscernible), and α is roughly inconsistent, if it is both roughly true and false (v(α) is totally indiscernible). Let us consider the modal system S5. Note that models of S5 and Rl are identical. We can then eﬀect a translation of the above concepts into S5. In(U, R, v), a wﬀ α can be termed roughly true if v(α) = v(♦α) = U , roughly false if v(α) = v(α) = ∅, and roughly inconsistent if both hold. In [10], a logic Lr having the same models as above was proposed, with the speciality that the syntax-semantics relationships are explored with rough truth replacing truth and rough validity replacing validity. The notion of consistency is replaced by one of rough consistency too. The consequence relation deﬁning the logic is also non-standard. These ideas were ﬁrst mooted in [5,26], and Lr is a modiﬁed version of the formal system discussed there. Lr has a normal modal language. A model M ≡ (U, R, v) is a rough model of Γ , if and only if for every γ ∈ Γ , v(♦γ) = U , i.e. γ is roughly true in M. α is a rough semantic consequence of Γ (denoted Γ |≈α) if and only if every rough model of Γ is a rough model of α. If Γ is empty, α is said to be roughly valid, written |≈α. There are two rules of inference: R1 . if S5

α R2 . ♦α β ♦β ♦α → ♦β ♦α ∧ ♦β

The consequence relation is deﬁned as follows. Let Γ be any set of wﬀs and α any wﬀ in Lr . α is a rough consequence of Γ (denoted Γ |∼α) if and only if there is a sequence α1 , ..., αn (≡ α) such that each αi (i = 1, ..., n) is either (i) a theorem of S5, or (ii) a member of Γ , or (iii) derived from some of α1 , ..., αi−1 by R1 or R2 . If Γ is empty, α is said to be a rough theorem, written |∼α. A kind of “rough Modus Ponens” is then derivable, in the form: if Γ |∼α, S5 α → β with S5 α ≈ α then β. Here ≈ reﬂects the notion of “rough equality”, α ≈ β ≡ (α ↔ β) ∧ (♦α ↔ ♦β). One also obtains soundness of Lr with respect to the above semantics: if Γ |∼α then Γ |≈α. It is clear that in the face of an incomplete description of a concept p, p and “not” p (in the classical sense) may not always represent conﬂicting situations. To accommodate this possibility, a set Γ of wﬀs is termed roughly consistent if and only if the set ♦Γ ≡ {♦γ : γ ∈ Γ } is S5-consistent. With the help of this notion, one obtains Theorem 6. (Completeness) (a) Γ is roughly consistent if and only if it has a rough model. (b) For any Lr -wﬀ α, if Γ |≈α then Γ |∼α. Thus, Lr appears as another system that is able to address rough sets and related notions. We shall remark on its relationship with other well-known systems in Section 5. It may be mentioned that Lr has been used as the base logic for a proposal of rough belief change in [25].

12

3

M. Banerjee and Md.A. Khan

Logics with Semantics Based on Information Systems

We now present logics, the models of which are deﬁned on approximation spaces induced by information systems. We ﬁnd one pioneering system N IL that has inspired the proposal of many others in the same line. The section also includes a logic by Nakamura, the models of which are directly deﬁned on information systems. 3.1

N IL

Recall that an incomplete information system is of the form S ≡ (U, A, V al, f ), where U is a set of objects, A a set of attributes, V al a set of values for the attributes, and f a function from U × A to P(V al). The logic N IL proposed by Orlowska and Pawlak [16] works on incomplete information systems, in which the function f satisﬁes an additional condition: () f (x, a) = ∅, for all x ∈ U, a ∈ A. One observes that, given S ≡ (U, A, V al, f ), two particular kinds of binary relations on the domain U are induced – these dictate the formulation of N IL. Let x, y ∈ U . Similarity (simS ): x simS y if and only if f (x, a) ∩ f (y, a) = ∅, for all a ∈ A. Inclusion (inS ): x inS y if and only if f (x, a) ⊆ f (y, a), for all a ∈ A. It can be shown that for every incomplete information system S ≡ (U, A, V al, f ) and x, y, z ∈ U , the following hold. (S1) x inS x. (S2) if x inS y and y inS z then x inS z. (S3) if x simS y for some y, then x simS x. (S4) if x simS y then y simS x. (S5) if x simS y, x inS u, y inS v then u simS v. Further, if the condition () is satisﬁed by f then sim satisﬁes (S6) x simS x. Thus a standard structure (cf. Introduction) corresponding to an incompleteinformation system S ≡ (U, A, V al, f ) with condition (), would be (U, simS , inS ). On the other hand, a general structure for S would be of the form (U, sim, in), where sim, in are binary relations on U satisfying (S1), (S2), (S4) − (S6). For brevity, we refer to these as standard and general N IL-structures respectively. N IL could be termed as a modal version of decision logic introduced by Pawlak [2], an association similar to that of rough logic [3] and S5 (cf. Section 2.6). The atomic propositions of N IL are the descriptors of decision logic – of the form (a, v), where a is an “attribute constant”, and v a constant representing “value of attribute”. Apart from the standard Boolean connectives ¬, ∨, the language contains modal connectives , 1 , 2 corresponding to sim, in and the inverse in−1 of in respectively. Wﬀs are built, as usual, out of the atomic propositions (descriptors) and the connectives. Note that there are no operations on the attribute or value constants.

Propositional Logics from Rough Set Theory

13

A N IL-model M ≡ (U, sim, in, m) consists of a general structure (U, sim, in) as above, along with a meaning function m from the set of all descriptors to the set P(U ). m is extended recursively to the set of all N IL-wﬀs in the usual manner. In particular, m(α) ≡ {x ∈ U : y ∈ m(α) for all y such that x sim y}. Similarly one deﬁnes m(1 α), and m(2 α). α is true in the model M, if m(α) = U. The following deductive system for N IL was proposed in [16]. Axiom schemes: A1. All classical tautologies, A2. 2 (α → β) → (2 α → 2 β), A3. 1 (α → β) → (1 α → 1 β), A4. (α → β) → (α → β), A5. α → 1 ¬2 ¬α, A6. α → 2 ¬1 ¬α, A7. 2 α → α, A8. 1 α → α, A9. α → α, A10. 2 α → 2 2 α, A11. 1 α → 1 1 α, A12. α → ¬¬α, A13. α → 2 1 α. Rules of inference:

(R1) α, α → β (R2) α β 2 α (R3)

α 1 α

(R4) α α

It has been proved that Theorem 7. For any N IL-wﬀ α, N IL α if and only if α is true in all N ILmodels. 3.2

Logics by Vakarelov

Vakarelov addresses the issue of completeness of various logics, the models of which are based on standard structures corresponding to some information system. For instance, in the case of N IL, the question would be about a completeness theorem with respect to the class of N IL-models deﬁned on standard N IL-structures (cf. Section 3.1). In [12], such a theorem is proved, via a key

14

M. Banerjee and Md.A. Khan

characterization result. In fact, this result set the ground for a series of similar observations when the binary relations involved are changed. Proposition 1. (Characterization) Let (U, sim, in) be a general N IL-structure. Then there exists an information system S ≡ (U, A, V al, f ) with f satisfying (), such that simS = sim and inS = in. In other words, the classes of N IL-models based on standard and general N ILstructures are identical. Hence one obtains the required completeness theorem. The condition (), viz. f (x, a) = ∅ for all x ∈ U, a ∈ A, is a restrictive one. However, it is observed by Vakarelov that even if this condition is dropped, a characterization result similar to Proposition 1 can be obtained. Instead of reﬂexivity of sim (cf. property (S6), Section 3.1), we now have just the condition of quasireﬂexivity – cf. property (S3): if x sim y for some y, then x sim x. The corresponding logic can be obtained from N IL by replacing the axiom A9 by ¬(p ∧ ¬p) → (α → α). Following this approach, one handles the cases of incomplete information systems inducing diﬀerent binary relations. For example, [14,13,17] consider these relations amongst others, for S ≡ (U, A, V al, f ): Indiscernibility (indS ): x indS y if and only if f (x, a) = f (y, a), for all a ∈ A, w Weak indiscernibility (indw S ): x indS y if and only if f (x, a) = f (y, a), for some a ∈ A, w Weak similarity (simw S ): x simS y if and only if f (x, a) ∩ f (y, a) = ∅, for some a ∈ A. Complementarity (com): x com y if and only if f (x, a) = (V ALa \ f (y, a)), for all a ∈ A, where V ala is the value set for the particular attribute a, and V al ≡ ∪{V ala : a ∈ A}. The characterization result for each has been obtained, the corresponding logical system is deﬁned and the completeness theorem with respect to models on the intended standard structures is proved. 3.3

Logic by Nakamura

[15] discusses a logic with models on incomplete information systems. We recall (cf. Introduction) that given a complete information system S ≡ (U, A, V al, f ), one can deﬁne the equivalence relation RS . The lower approximation of X(⊆ U ) under this relation is denoted as X S , and its upper approximation as X S . Nakamura deﬁnes a completation S0 of an incomplete information system S as a complete information system that can be constructed from S by selecting any one value from f (x, a)(⊆ V al), for each x ∈ U, a ∈ A. If f (x, a) = ∅, one selects a special symbol . The relationship of S0 and S is expressed as S0 ≥ S. Now the “lower” and “upper approximations” X, X of X ⊆ U in an incomplete information system S ≡ (U, A, V al, f ) are deﬁned as follows: (∗)

X ≡ ∩S0 ≥S X S0 , X ≡ ∪S0 ≥S X S0 .

Propositional Logics from Rough Set Theory

15

With this background, a logic IN CRL is proposed, having the standard Boolean connectives, and two modal operators [], (corresponding to “surely” and “possibly” respectively). An IN CRL-model is an incomplete information system S ≡ (U, A, V al, f ) along with a meaning function vS from the set of propositional variables of the language to P(U ). vS is extended as usual for the wﬀs involving Boolean connectives. For wﬀs with modal operators, one makes use of completations S0 of S and the preceding deﬁnitions of lower and upper approximations given in (∗). vS ([ ]α) ≡ ∩S0 ≥S vS (α)S

0

= vS (α),

vS ( α) ≡ ∪S0 ≥S vS (α)S0 = vS (α). Truth and validity of wﬀs are deﬁned again as for most of the previous systems. Nakamura points out relationships of IN CRL with the modal system KT B, in particular that all theorems of KT B are valid wﬀs of IN CRL. We shall take a further look at the two logics in Section 5.

4

Other Approaches

This section outlines a few proposals of logics related to rough sets, the models of which are based on structures that are even more generalized than the ones already presented. As we shall see, these logics have dimensions not accounted for in the systems presented so far. 4.1

Temporal Approach

Orlowska (cf. [4]), deﬁnes a logic LT with models on dynamic information systems, in order to deal with the temporal aspect of information. A set T of moments of time, and a suitable relation R on the set T are considered along with the set U of objects and A of attributes. Formally, a dynamic information system is a tuple S ≡ (U, A, V al, T, R, f ), where V al ≡ ∪{V ala : a ∈ A}, (V ala , as in Section 3.2, being the value set for the particular attribute a) and the information function f : U × T × A → V al satisﬁes the condition that f (x, t, a) ∈ V ALa , for any x ∈ U, t ∈ T, a ∈ A. In the language of LT , atomic statements are descriptors of decision logic, together with an object constant x – so these are triples (x, a, v), and are intended to express: “object x assumes value v for attribute a”. There are modal operators to reﬂect the relations R and R−1 . The truth of all statements of the language is evaluated in a model based on a dynamic information system, with respect to moments of time, i.e. members of the set T . An LT -model is a tuple M ≡ (S, m) where S is a dynamic information system, and m a meaning function which assigns objects, attributes and values from U, A, V al to the respective constants. The satisﬁability of a formula α in a model M at a moment t(∈ T ) of time is deﬁned inductively as follows: M, t |= (x, a, v) if and only if f (m(x), t, m(a)) = m(v).

16

M. Banerjee and Md.A. Khan

For the Boolean cases, we have the usual deﬁnitions. For the modal case, M, t |= [R]α if and only if for all t ∈ T , if (t, t ) ∈ R then M, t |= α. A wﬀ is true in M, provided it is satisﬁed in M at every t ∈ T . LT is complete with respect to this class of models, for the axioms of linear time temporal logic, and an axiom which says that the values of attributes are uniquely assigned to objects. 4.2

Multiagent Systems

[20] describes a logic, that takes into account a (ﬁnite) collection of agents and their knowledge bases. We denote the logic as LMA . The language of LMA has “agent constants” along with two special constants 0,1. Binary operations +,. are provided to build the set T of terms from these constants. Wﬀs of one kind are obtained from terms, and are of the form s ⇒ t, s, t ∈ T , where ⇒ is a binary relational symbol. s ⇒ t is to reﬂect that “the classiﬁcation ability of agent t is at least as good as that of agent s”. Furthermore, there are attribute as well as attribute-value constants. Descriptors formed by these constants constitute atomic propositions, and using connectives ∧, ¬ and modal operators It , t ∈ T (representing “partial knowledge” of each agent), give wﬀs of another kind. LMA -models are not approximation spaces, but what could be called “partition spaces” on information systems. Informally put, a model consists of an information system S ≡ (U, A, V al, f ), and a family of partitions {Et }t∈T on the domain U – each corresponding to the knowledge base of an agent. The family is shown to have a lattice structure, and the ordering involved gives the interpretation of the relational symbol ⇒. Wﬀs built out of descriptors are interpreted in the standard way, in the information system S. The partial knowledge operator It for a term t reﬂects the lower approximation operator with respect to the partition Et on U . An axiomatization of LMA is presented, to give soundness and completeness results. In the context of multiagent systems, it is worth mentioning the approach followed in [38], even though a formal logic based on it has not been deﬁned yet. Property systems (P -systems) are deﬁned as triples of the form (U, A, |=), where U is a set of objects, A a set of properties, and |= a “fulﬁlment” relation between U and A. For each P -system P, a collection P op of interior and closure operators satisfying speciﬁc properties are considered. These operators could be regarded as generalizations of lower and upper approximations. Now given a family {Pk }k∈K of P -systems (each for an agent, say) over some index set K and over the same set U of objects, one obtains a multiagent pre-topological approximation space as a structure (U, {Pkop }k∈K ). It is to be seen if such a generalized structure could form the basis of a semantics of some formal logical framework. 4.3

Rough Relations

Discussion about relations on approximation spaces, started from [39]. We ﬁnd two directions of work on this topic.

Propositional Logics from Rough Set Theory

17

Logic of Rough Relations: [40] considers another generalization of the notion of an approximation space – taking systems of the form AS ≡ (U, I, v), where U is a non-empty set of objects, I : U → P(U ) an uncertainty function, and v : P(U ) × P(U ) → [0, 1] is a rough inclusion function satisfying the following conditions: v(X, X) = 1 for any X ⊆ U , v(X, Y ) = 1 implies v(Z, Y ) ≥ v(Z, X) for any X, Y, Z ⊆ U, v(∅, X) = 1 for any X ⊆ U . For any subset X of U , we then have the lower and upper approximations: L(AS, X) ≡ {x ∈ U : v(I(x), X) = 1}, U (AS, X) ≡ {x ∈ U : v(I(x), X) > 0}. A ‘rough set’ in AS is the pair (L(AS, X), U (AS, X)). The above is motivated from the fact that any Pawlak approximation space (U, R) is an instance of a generalized space as just deﬁned. Indeed, we consider the function I that assigns to every object its equivalence class under R, and the inclusion function v as: v(S, R) ≡

card(S∩R) card(S)

1

if S = ∅ if S = ∅

For an approximation space AS ≡ (U, I, v) with U = U1 × U2 and v as in the special case above, [21] discusses relations R ⊆ U1 × U2 . The lower and upper approximation of R in AS are taken, and a rough relation is just a rough set in AS. A decidable multimodal logic is proposed – for reasoning about properties of rough relations. The modal operators correspond to a set of relations on the domain of the above generalized approximation spaces, as well as the lower and upper approximations of these relations. An axiomatization for the logic is given, and completeness is proved with respect to a Kripke-style semantics. Rough Datalog: Just as decision tables [2] are (complete) information systems with special attributes, viz. the decision attributes, [22] considers a decision system (U, A ∪ {d}) – but with a diﬀerence. Each attribute a in A is a partial map from U to a value set Va , and d, the decision attribute, is a partial map from U to {0, 1}. It is possible that for some x ∈ U , all attribute values (including the value of d) are undeﬁned. A ‘rough set’ X is taken to be a pair (X + , X − ), where X + is the set of elements of U that may belong to X, while X − contains those elements of U that may not belong to X. d indicates the information about membership of an object of U in X. Formally, let A ≡ {a1 , ..., an }, A(x) ≡ (a1 (x), ..., an (x)) for each x ∈ U , and A−1 (t) ≡ {x ∈ U : A(x) = t}, for t ∈ Va1 × ... × Van . (Note that for some x ∈ U , A(x) could be undeﬁned). Then X + ≡ {x ∈ U : A is deﬁned for x, and d(x ) = 1, for some x ∈ A−1 (A(x))}, and X − ≡ {x ∈ U : A is deﬁned for x, and d(x ) = 0, for some x ∈ A−1 (A(x))}.

18

M. Banerjee and Md.A. Khan

This deﬁnition implies that X + and X − may not be disjoint, allowing for the presence of conﬂicting (contradictory) decisions in the decision table. On the other hand, X + and X − may not cover U either, allowing for the possibility that there is no available information about membership in X. With these deﬁnitions, ‘rough relations’ are considered in [22]. Standard relational data base techniques, such as relational algebraic operations (e.g. union, complement, Cartesian product, projection) on crisp relations, are extended to the case of rough relations. A declarative language for deﬁning and querying these relations is introduced - pointing to a link of rough sets (as just deﬁned) with logic programming. 4.4

Logics with Attribute Expressions

As we have seen, LT and LMA (cf. Sections 4.1 and 4.2 respectively) have attribute expressions in the language that are interpreted in information systems. N IL (cf. Section 3.1), also has attribute constants in the language. But unlike the models of LT and LMA , the standard or general N IL-structures deﬁning N IL-models do not accommodate attributes, and the wﬀs (which are built using the attribute constants) point to collections of objects of the domain. A class of logics with attribute expressions are also deﬁned in [18,19]. Models are based on structures of the form (U, A, {ind(P )}P ⊆A ), where the “indiscernibility” relation ind(P ) for each subset P of the attribute set A, has to satisfy certain conditions. For the models of one of the logics, for example, the following conditions are stipulated for ind(P ): (U 1) ind(P ) is an equivalence relation on U , (U 2) ind(P ∪ Q) = ind(P ) ∩ ind(Q), (U 3) if P ⊆ Q then ind(Q) ⊆ ind(P ), and (U 4) ind(∅) = U × U . Other logics may be obtained by changing some of (U 1) − (U 4). The language of the logics has a set of variables each representing a set of attributes, as well as constants to represent all one element sets of attributes. Further, the language can express the result of (set-theoretic) operations on sets of attributes. The logics are multimodal – there is a modal operator to reﬂect the indiscernibility relation for each set of attributes as above. A usual Kripke-style semantics is given, and a number of valid wﬀs presented. However, as remarked in [19], we do not know of a complete axiomatization for such logics. 4.5

Rough Mereology

This is an approach inspired by the theory of mereology due to Le´sniewski (1916). Le´sniewski propounds a theory of sets that has containment as the primitive relation, rather than membership. Drawing from this classical theory, rough mereology has been proposed [23], providing a useful notion of rough containment, of “being a part, in a degree”. Formally, this can be deﬁned as a real binary function μ on the domain with values in [0,1], satisfying certain conditions (abstracted from the properties of

Propositional Logics from Rough Set Theory

19

classical containment). A given information system (U, A, V al, f ), a partition of A into, say A1 , ..., An and a set of weights {w1 , ..., wn }, generate μ(x, y), can n x, y ∈ U . It is assumed that wi ∈ [0, 1], i = 1, ..., n, and i=1 wi = 1. A pre-rough inclusion μo is ﬁrst deﬁned: μo (x, y) ≡ ni=1 wi .(|indi (x, y)|/|Ai |), where indi (x, y) ≡ {a ∈ Ai : f (x, a) = f (y, a)}. μo can then be extended to rough inclusion μ over P(U ) by using t-norms and t-conorms. Rough inclusion can be used, for instance, in specifying approximate decision rules. It may be remarked that predicate logics corresponding to rough inclusions have been proposed recently in [41].

5

Comparative Study

We now discuss some relationships between the logics presented in Sections 2 and 3. 5.1

Embeddings

Let us recall the notion of an embedding of logics – cf. Section 2.4. We consider the logics P RL, L3, LD , RDSL presented in Sections 2.3, 2.4 and 2.5 respectively, and point out interrelationships, as well as relations with other known logics. (1) L3 P RL: This has already been seen in Section 2.4. untsch and Pagliani, (2) L3 LD : As summarized in [1] and observed by D¨ regular double Stone algebras and 3-valued L ukasiewicz algebras are equivalent to each other via suitable transformations. Passing on to the respective logics, we would thus ﬁnd embeddings both ways, between LD and L3. (3) LD RDSL: We can deﬁne, in RDSL, that a wﬀ α is a theorem (valid), if and only if the sequent ⇒ α is derivable (valid). Using the formal argument made in Section 2.5 to show that the classes of models of the logics LD and RDSL are identiﬁable and Theorems 4, 5, one gets the result with the identity embedding. (4) L3 LSN : LSN denotes constructive logic with strong negation [30]. We note that semi-simple Nelson algebras are the algebraic counterparts for ukasiewicz LSN . The equivalence of semi-simple Nelson algebras and 3-valued L algebras through suitable translations has also been observed e.g. by Pagliani. Hence the stated embedding. (5) P RL S5: One observes [31] a translation of wﬀs of P RL into S5 that assigns the operations of negation ¬ and necessity in P RL those same operations of S5. Further, is translated in terms of the conjunction ∧ and disjunction ∨ of S5 as:

20

M. Banerjee and Md.A. Khan

(α β) ≡ (α ∧ β ) ∨ (α ∧ M β ∧ ¬M (α ∧ β )). Then it can be shown that P RL α if and only if α , for any wﬀ α of P RL. (6) S5 Lr : The logic Lr for rough truth is able to capture, as the class of its theorems, exactly the “♦-image” of the class of S5-theorems, i.e. S5 ♦α if and only if |∼α [5,10]. Note that the languages of Lr and S5 are the same. We translate α in S5 to α∗ ≡ Lα. Then α if and only if |∼α∗ . For the other direction, we consider the translation α◦ ≡ M α. (7) J Lr : In 1948, Ja´skowski proposed a “discussive” logic – he wanted a formalism to represent reasoning during a discourse. Each thesis, a discussive assertion of the system, is supposed either to reﬂect the opinion of a participant in the discourse, or to hold for a certain “admissible” meaning of the terms used in it. Formally, any thesis α is actually interpreted as “it is possible that α”, and the modal operator ♦ is used for the expression. The logic J (cf. [42]) is such a system. The J-consequence, deﬁned over S5, is such that: J α if and only if S5 ♦α. Because of the relationship between Lr and S5 noted in (6) above, we have J Lr with the identity embedding. In the whole process, one has obtained an alternative formulation of the paraconsistent logic J (proposed in a diﬀerent context altogether), and established a link between Pawlak’s and Ja´skowski’s ideas. 5.2

KT B and Nakamura’s Logic IN CRL

We refer to Section 3.3, and present a connection between IN CRL, and the normal modal system KT B. KT B, as we know, is sound and complete with respect to the class of reﬂexive and symmetric Kripke frames. Let S ≡ (U, A, V al, f ) be an incomplete information system, and let us consider the relation on U deﬁned as follows: x y if and only if there exists a completation S0 of S such that x RS0 y. Clearly is reﬂexive and symmetric, but not transitive. From the deﬁnitions of vS ([ ]α) and vS ( α), we see that x ∈ vS ([ ]α) if and only if, for all y ∈ U such that x y, y ∈ vS (α), and x ∈ vS ( α) if and only if, there exists y ∈ U such that x y and y ∈ vS (α). So all provable wﬀs of the modal logic KT B are valid in IN CRL. What about the converse – are all valid wﬀs of IN CRL provable in KT B? [15] makes a cryptic comment about this, we establish the converse here. KT B provides an axiomatization for IN CRL: We show that if α is not provable in KT B then it is not valid in IN CRL. It suﬃces then, to construct an incomplete information system S ≡ (U, A, {V ala }a∈A , f ) for any given KT Bframe (W, R), such that is identical with R. Let g be a function from R (⊆ W × W ) to some set C of constants, satisfying the following conditions: (i) g(x, y) = g(y, x), (ii) g(x, y) = g(t, z) implies that either x = t and y = z, or x = z and y = t.

Propositional Logics from Rough Set Theory

21

(g essentially assigns, upto symmetry, a unique constant from C to every pair in R.) Now consider U ≡ W, A ≡ {a}, where a is a new symbol. Further, deﬁne f (x, a) ≡ {g(x, y) : y ∈ U and (x, y) ∈ R}, so that V ala ⊆ C. We claim that xRy if and only if x y. Suppose xRy. Then g(x, y) ∈ f (x, a)∩ f (y, a) and hence x y. Conversely, if x y, there exists d ∈ f (x, a) ∩ f (y, a). Now d ∈ f (x, a) implies that d = g(x, z), for some z ∈ U such that (x, z) ∈ R, and d ∈ f (y, a) implies that d = g(y, t), for some t ∈ U such that (y, t) ∈ R. From the property of g, it follows that either x = y or x = t, whence by reﬂexivity and symmetry of R, we get xRy. The proof above, in fact, yields a characterization theorem, viz. given any reﬂexive, symmetric frame (W, R), there exists an incomplete information system S ≡ (U, A, {V ala }a∈A , f ) satisfying the condition () (cf. Section 3.1) such that R = = simS . 5.3

Normal Modal Systems and Vakarelov’s Logics

Vakarelov has proved the characterization theorem for incomplete information systems with respect to diﬀerent sets of relations [12,14,13,17]. As we have remarked in the Introduction, a special case would be obtained with respect to the indiscernibility relation on the Pawlak approximation space. One ﬁnds that if we restrict the logics presented in [14,13,17] to take a modal operator corresponding only to the indiscernibility relation, the resulting system would be just the modal logic S5. As noted at the end of Section 5.2, if an incomplete information system satisﬁes the condition (), then the similarity relation simS is the same as the relation . So it follows that if we restrict the logic N IL to take only the modality in the language then the corresponding logic will be just IN CRL, or, in other words, KT B. 5.4

DAL Again

Observing Vakarelov’s strain of work, it may be tempting to look for a kind of characterization result in the case of DAL (cf. Section 2.2) as well. Consider a general DAL-structure U ≡ (U, {Ri }i∈I ), where the family {Ri }i∈I of equivalence relations is closed under intersection and transitive closure of union. Can one ﬁnd an incomplete information system S ≡ (U, A, V al, f ) such that the standard structure for S is just U? Let us assume that the standard structure is obtained “naturally” from S, viz. that the equivalence relations in it are the ones induced by the subsets of A. As it turns out, this is a hard question. However, we can ﬁnd an information system, such that the standard structure obtained from it in the above manner cannot be a general DAL-structure. Suppose for some incomplete information system S ≡ (U, A, V al, f ), R and P are the equivalence relations induced by subsets R , P of A respectively – we denote this as ind(R ) = R and ind(P ) = P . For the equivalence relation

22

M. Banerjee and Md.A. Khan

R ∩ P , R ∪ P ⊆ A is such that ind(R ∪ P ) = R ∩ P . But in the case of R P , there may not be any Q ⊆ A such that ind(Q) = R P. Consider the following example [11]. Example 1. U ≡ {o1, o2, o3, o4, o5, o6, o7}, where each oi consists of circles and squares. Let A ≡ {number of circles (), number of squares ()}. The information function is given by the following table: ————————————————————– ————————————————————– o1 1 1 o2 1 2 o3 2 1 o4 2 2 o5 3 3 o6 3 4 o7 3 4 ————————————————————— Equivalence classes of indiscernibility relations ind() and ind() are: ind() : {o1, o2}, {o3, o4}, {o5, o6, o7}, ind() : {o1, o3}, {o2, o4}, {o5}, {o6, o7}. The transitive closure of these relations gives the following equivalence classes: ind() ind() : {o1, o2, o3, o4}, {o5, o6, o7}. Clearly there is no Q ⊆ A such that ind(Q) = ind() ind().

6

Summary and Questions

We have tried to present the various proposals of logics with semantics based on rough sets, including some generalizations. Two main approaches emerge, discussed in Sections 2 and 3. One of these considers logics, the models of which are approximation spaces, while the other considers approximation spaces, but those induced by information systems. However, it is found through characterization results, that both lines of study converge, in that the two semantics for a particular system are identical. This actually reﬂects on the apt description of the properties of the relations deﬁning the approximation spaces. The only exception is the logic DAL of the ﬁrst category. As remarked in Section 5.4, given a general DAL-structure U ≡ (U, {Ri }i∈I ), it does not seem easy to construct an information system “naturally” to obtain U back as its standard structure. In case of the logics with attributes as expressions (cf. Section 4.4), one encounters a problem even earlier. The models here are based on structures of the form (U, A, {ind(P )}P ⊆A ), and there does not appear easily a corresponding “general” structure of the kind U ≡ (U, {Ri }i∈I ), with appropriate closure conditions on {Ri }i∈I . These logics have not been axiomatized, though the language can express a lot about attributes – that few of the other systems are able to do.

Propositional Logics from Rough Set Theory

23

An interesting picture is obtained from the logics of Section 2, leaving out DAL and other systems with models based on generalized spaces. Most of the logics are embeddable into each other (cf. Section 5). We have LD L3 P RL S5 Lr J. (1) In one sense then, the embeddings in (1) establish that no ‘new’ logic surfaces with the kind of rough set semantics deﬁned. But in another sense, well-known systems have been imparted a rough set interpretation. It should be noted that though the embeddings are deﬁned with respect to theoremhood, the relationships would hold in some cases (e.g. L3 − P RL and Lr − J) if derivability of wﬀs from non-empty premise sets is considered [31,10]. One could attempt to settle the question for the rest. (1) indicates another interesting future line of work, viz. an investigation for logics and interrelations, that may result on replacing S5 by other non-modal systems (as in [6]). All the systems presented other than LT (cf. Section 4.1), deal with static information. The semantics of LT essentially gives rise to a family of approximation spaces on the same domain, the indiscernibility relations changing with moments of time. One could further enquire about the behaviour of rough sets in such a dynamic information system. As remarked in Section 4.2, another open direction relates to a study of logics that may be obtained from the generalized approach in [38]. Overall, one may say that it has been a remarkable journey in the exploration of logics, beginning with a deceptively simple proposal of “rough sets”. We have seen the introduction of novel concepts – e.g. of “rough truth”, “rough modus ponens”, “rough consistency”, “rough mereology”. The journey has, by no means, ended. Pawlak’s theory has just opened up the horizon before us, to reveal a number of yet unexplored directions in the study of “rough logics”.

References 1. Banerjee, M., Chakraborty, M.K.: Algebras from rough sets. In Pal, S.K., Polkowski, L., Skowron, A., eds.: Rough-neuro Computing: Techniques for Computing with Words. Springer Verlag, Berlin (2004) 157–184 2. Pawlak, Z.: Rough Sets. Theoretical Aspects of Reasoning about Data. Kluwer Academic Publishers, Dordrecht (1991) 3. Pawlak, Z.: Rough logic. Bull. Polish Acad. Sc. (Tech. Sc.) 35 (1987) 253–258 4. Orlowska, E.: Kripke semantics for knowledge representation logics. Studia Logica XLIX (1990) 255–272 5. Chakraborty, M.K., Banerjee, M.: Rough consequence. Bull. Polish Acad. Sc.(Math.) 41(4) (1993) 299–304 6. Yao, Y., Lin, T.Y.: Generalization of rough sets using modal logics. Intelligent Automation and Soft Computing 2 (1996) 103–120 7. Banerjee, M., Chakraborty, M.K.: Rough sets through algebraic logic. Fundamenta Informaticae 28(3,4) (1996) 211–221 8. D¨ untsch, I.: A logic for rough sets. Theoretical Computer Science 179 (1997) 427–436

24

M. Banerjee and Md.A. Khan

9. Dai, J.H.: Logic for rough sets with rough double Stone algebraic semantics. In Slezak, D., Wang, G., Szczuka, M.S., D¨ untsch, I., Yao, Y., eds.: Proc. RSFSDMGrC(1), Canada, LNCS 3641, Springer Verlag (2005, 141-148) 10. Banerjee, M.: Logic for rough truth. Fundamenta Informaticae 71(2-3) (2006) 139–151 11. Farinas Del Cerro, L., Orlowska, E.: DAL – a logic for data analysis. Theoretical Computer Science 36 (1997) 251–264 12. Vakarelov, D.: Abstract characterization of some knowledge representation systems and the logic N IL of nondeterministic information. In Jorrand, P., Sgurev, V., eds.: Artiﬁcial Intelligence II. North–Holland (1987) 255–260 13. Vakarelov, D.: A modal logic for similarity relations in Pawlak knowledge representation systems. Fundamenta Informaticae 15 (1991) 61–79 14. Vakarelov, D.: Modal logics for knowledge representation systems. Theoretical Computer Science 90 (1991) 433–456 15. Nakamura, A.: A rough logic based on incomplete information and its application. Int. J. Approximate Reasoning 15 (1996) 367–378 16. Orlowska, E., Pawlak, Z.: Representation of nondeterministic information. Theoretical Computer Science 29 (1984) 27–39 17. Vakarelov, D., Balbiani, P.: A modal logic for indiscernibilty and complementarity in information systems. Fundamenta Informaticae 50 (2002) 243–263 18. Orlowska, E.: Logic of nondeterministic information. Studia Logica 1 (1985) 91–100 19. Orlowska, E.: Logic of indiscernibility relations. In Goos, G., Hartmanis, J., eds.: Proc. Symposium on Computation Theory, Zabr´ ow, 1984, LNCS 208, Springer Verlag (1985, 177–186) 20. Rauszer, C.M.: Rough logic for multiagent systems. In Masuch, M., Polos, L., eds.: Knowledge Representation and Reasoning under Uncertainty, LNAI 808. SpringerVerlag (1994) 161–181 21. Stepaniuk, J.: Rough relations and logics. In Polkowski, L., Skowron, A., eds.: Rough Sets in Knowledge Discovery 1: Methodology and Applications. PhysicaVerlag (1998) 248–260 22. Maluszy´ nski, Vit´ oria, A.: Toward rough datalog: embedding rough sets in prolog. In Pal, S.K., Polkowski, L., Skowron, A., eds.: Rough-neuro Computing: Techniques for Computing with Words. Springer Verlag, Berlin (2004) 297–332 23. Polkowski, L., Skowron, A.: Rough mereology: a new paradigm for approximate reasoning. Int. J. Approximate Reasoning 15(4) (1997) 333–365 24. Wong, S.K.M.: A rough set model for reasoning about knowledge. In Orlowska, E., ed.: Incomplete Information: Rough Set Analysis. Studies in Fuzziness and Soft Computing, vol. 13. Physica-Verlag (1998) 276–285 25. Banerjee, M.: Rough belief change. Transactions of Rough Sets V LNCS 4100 (2006) 25–38 26. Banerjee, M., Chakraborty, M.K.: Rough consequence and rough algebra. In Ziarko, W.P., ed.: Rough Sets, Fuzzy Sets and Knowledge Discovery, Proc. Int. Workshop on Rough Sets and Knowledge Discovery (RSKD’93). Workshops in Computing, London, Springer Verlag (1994, 196–207) 27. Yao, Y.: Constructive and algebraic methods of the theory of rough sets. Information Sciences 109 (1998) 21–47 28. Pomykala, J.: Approximation, similarity and rough construction. preprint CT–93– 07, ILLC Prepublication Series, University of Amsterdam (1993)

Propositional Logics from Rough Set Theory

25

29. Komorowski, J., Pawlak, Z., Polkowski, L., Skowron, A.: Rough sets: a tutorial. In Pal, S.K., Skowron, A., eds.: Rough Fuzzy Hybridization: A New Trend in Decision-Making. Springer Verlag, Singapore (1999) 3–98 30. Rasiowa, H.: An Algebraic Approach to Non-classical Logics. North Holland, Amsterdam (1974) 31. Banerjee, M.: Rough sets and 3-valued L ukasiewicz logic. Fundamenta Informaticae 32 (1997) 213–220 32. Pagliani, P.: Rough set theory and logic-algebraic structures. In Orlowska, E., ed.: Incomplete Information: Rough Set Analysis. Studies in Fuzziness and Soft Computing, vol. 13. Physica-Verlag (1998) 109–190 33. Iturrioz, L.: Rough sets and three-valued structures. In Orlowska, E., ed.: Logic at Work: Essays Dedicated to the Memory of Helena Rasiowa. Studies in Fuzziness and Soft Computing, vol. 24. Physica-Verlag (1999) 596–603 34. Boicescu, V., Filipoiu, A., Georgescu, G., Rudeano, S.: L ukasiewicz-Moisil Algebras. North Holland, Amsterdam (1991) 35. Comer, S.: Perfect extensions of regular double Stone algebras. Algebra Universalis 34 (1995) 96–109 36. Andr´eka, H., N´emeti, I., Sain, I.: Abstract model theoretic approach to algebraic logic. CCSOM working paper, Department of Statistics and Methodology, University of Amsterdam (1992) 37. Pawlak, Z.: Rough sets. Int. J. Comp. Inf. Sci. 11(5) (1982) 341–356 38. Pagliani, P., Chakraborty, M.K.: Information quanta and approximation spaces I: non-classical approximation operators. In: Proc. 2005 IEEE Conf. on Granular Computing, IEEE Press (2005, 605–610) 39. Pawlak, Z.: Rough relations. ICS PAS Reports 435 (1981) 40. Skowron, A., Stepaniuk, J.: Tolerance approximation spaces. Fundamenta Informaticae 27 (1996) 245–253 41. Polkowski, L.: Rough mereological reasoning in rough set theory: recent results and problems. In Wang, G., Peters, J.F., Skowron, A., Yao, Y., eds.: Proc. Rough Sets and Knowledge Technology (RSKT 2006), China, 2006, LNAI 4062, Springer Verlag (2006, 79–92) 42. da Costa, N.C.A., Doria, F.A.: On Ja´skowski’s discussive logics. Studia Logica 54 (1995) 33–60

Intuitionistic Rough Sets for Database Applications Theresa Beaubouef1 and Frederick E. Petry2 1

2

1

Southeastern Louisiana University Dept. of Computer Science & Ind. Technology Hammond, LA 70402, USA [email protected] Center for Intelligent and Knowledge-Based Systems Tulane University New Orleans, LA 70118, USA [email protected]

Introduction

We introduce the intuitionistic rough set and intuitionistic rough relational and object oriented database models. The intuitionistic rough set database models draw beneﬁts from both the rough set and intuitionistic techniques, providing greater management of uncertainty for databases applications in a less than certain world. We provide the foundation for the integration of intuitionistic rough sets into modeling of uncertainty in databases. This builds upon some of our previous research [2,3] with integrating fuzzy and rough set techniques for uncertainty management in databases.

2

Intuitionistic Rough Sets

An intuitionistic set [1] (intuitionistic fuzzy set) is a generalization of the traditional fuzzy set. Let set X be ﬁxed. An intuitionistic set A is deﬁned by the following: A = {x, μA (x), νA (x) : x ∈ X} where μA (x) → [0, 1], and νA (x) → [0, 1]. The degree of membership of element x ∈ X to the set A is denoted by μA (x), and the degree of nonmembership of element x ∈ X to the set A is denoted by νA (x). A is a subset of X. For all x ∈ X, 0 ≤ μA (x)+νA (x) ≤ 1. A hesitation margin, πA (x) = 1−(μA (x)+νA (x)), expresses a degree of uncertainty about whether x belongs to X or not, or uncertainty about the membership degree. This hesitancy may cater toward membership or nonmembership. We next deﬁne the intuitionistic rough set, which incorporates the beneﬁcial properties of both rough set [5] and intuitionistic set techniques. Intuitionistic rough sets are generalizations of fuzzy rough sets that give more information about the uncertain, or boundary region. They follow the deﬁnitions for partitioning of the universe into equivalence classes as in rough sets, but instead J.F. Peters et al. (Eds.): Transactions on Rough Sets VI, LNCS 4374, pp. 26–30, 2007. c Springer-Verlag Berlin Heidelberg 2007

Intuitionistic Rough Sets for Database Applications

27

of having a simple boundary region, there are basically two boundaries formed from the membership and nonmembership functions. Let U be a universe, Y a rough set in U , deﬁned on a partitioning of U into equivalence classes. Deﬁnition 1. An intuitionistic rough set Y in U is Y, μY (x), νY (x), where μY (x) is a membership function which associates a grade of membership from the interval [0,1] with every element (equivalence class) of U , and νY (x) associates a degree of non membership from the interval [0,1] with every element (equivalence class) of U , where 0 ≤ μY (x) + νY (x) ≤ 1, where x denotes the equivalence class containing x. A hesitation margin is πY (x) = 1 − (μY (x) + νY (x)). Consider the following special cases μ, ν for some element of Y : 1, 0 denotes total membership. This correspond to elements found in RY . 0, 1 denotes elements that do not belong to Y . Same as U − RY . 0.5, 0.5 corresponds to traditional rough set boundary region. p, 1 − p corresponds to fuzzy rough set in that there is a single boundary. In this case we assume that any degree of membership has a corresponding complementary degree of non membership. p, 0 corresponds to fuzzy rough set. 0, q This case can not be modeled by fuzzy rough sets. It denotes things that are not a member of RY or RY . It falls somewhere in the region U − RY . p, q Intuitionistic set general case , has membership and nonmembership. Let Y denote the complement of Y . Then the intuitionistic set having μY (x), μY (x) is the same as fuzzy rough set. The last two cases above, 0, q and p, q, cannot be represented by fuzzy sets, rough sets, or fuzzy rough sets. These are the situations which show that intuitionistic rough sets provide greater uncertainty management than the others alone. Note, however, that with the intuitionistic set we do not lose the information about uncertainty provided by other set theories, since from the ﬁrst few cases we see that they are special cases of the intuitionistic rough set. Although there are several various way of combining rough and fuzzy sets, we focus on those fuzzy rough sets as deﬁned in [2,3] and used for fuzzy rough databases, since our intuitionistic rough relational database model follows from this. The intuitionistic rough relational database model will have an advantage over the rough and fuzzy rough database models in that the non membership uncertainty of intuitionistic set theory will also play a role, providing even greater uncertainty management than the original models.

3

Intuitionistic Rough Relational Database Model

The intuitionistic rough relational database, as in the ordinary relational database, represents data as a collection of relations containing tuples. Because a relation is considered a set having the tuples as its members, the tuples are

28

T. Beaubouef and F.E. Petry

unordered. In addition, there can be no duplicate tuples in a relation. A tuple ti takes the form (di1 , di2 , . . . , dim , diμ , diν ), where dij is a domain value of a particular domain set Dj and diμ ∈ Dμ , where Dμ is the interval [0,1], the domain for intuitionistic membership values, and Dv is the interval [0,1], the domain for intuitionistic nonmembership values. In the ordinary relational database, dij ∈ Dj . In the intuitionistic rough relational database, except for the intuitionistic membership and nonmembership values, however, dij ∈ Dj , and although dij is not restricted to be a singleton, dij = ∅. Let P (Di ) denote any non-null member of the powerset of Di . Deﬁnition 2. A intuitionistic rough relation R is a subset of the set cross product P (D1 ) × P (D2 ) × . . . × P (Dm ) × Dμ × Dnu . An intuitionistic rough tuple t is any member of R. If ti is some arbitrary tuple, then ti = (di1 , di2 , . . . , dim , diμ , diν ) where dij ∈ Dj and diμ ∈ Dμ , diν ∈ Dν . Let [dxy ] denote the equivalence class to which dxy belongs. When dxy is a set of values, the equivalence class is formed by taking the union of equivalence classes of members of the set; if dxy = {c1 , c2 , ..., cn }, then [dxy ] = [c1] × [c2] × . . .× [cn ]. Deﬁnition 3. Tuples ti = (di1 , di2 , . . . , din , diμ , diν ) and tk = (dk1 , dk2 , . . . , dkn , dkμ , dkν ) are redundant if [dij ] = [dkj ] for all j = 1, . . . , n. In [3], we deﬁned several operators for the rough relational algebra, and in [2] demonstrated the expressive power of the fuzzy rough versions of these operators in the fuzzy rough relational database model. In an extension of this work we do the same for the rough intuitionistic database.

4

Intuitionistic Rough Object-Oriented Database (IROODB) Model

We next develop the intuitionistic rough object-oriented database model. We follow the formal framework and type deﬁnitions for generalized object-oriented databases proposed by [4] and extended for rough sets in [3]. We extend this framework, however, to allow for intuitionistic rough set indiscernibility and approximation regions for the representation of uncertainty as we have previously done for relational databases [2,3]. The intuitionistic rough object database scheme is formally deﬁned by the following type system and constraints. type ], where T can be a literal type Tliteral , The type system, T S = [T, P, fimpl which can be a base type, a collection literal type, or a structured literal type. It also contains Tobject , which speciﬁes object types, Tref erence , the set of speciﬁcations for reference types, and a void type. In the type system, each domain domts ∈ Dts , the set of domains. This domain set, along with a set of operators Ots and a set of axioms Ats , capture the semantics of the type speciﬁcation. The type system is then deﬁned based on these type speciﬁcations, the set of all programs P , and the implementation function mapping each type speciﬁcation

Intuitionistic Rough Sets for Database Applications

29

for a domain onto a subset of ρ(P ) – the powerset of P that contains all the implementations for the type system: type : T → ρ(P ) giving ts → {p1 , p2 , . . . pn }. fimpl

We are particularly interested in object types, and specify a class t of object types as Class id(id1 : s1 ; . . . ; idn : sn ) or Class id : id1 , . . . , idn (id1 : s1 ; . . . ; idn : sn ) where id, an identiﬁer, names an object type, {idi : 1 ≤ i ≤ m} is a ﬁnite set of identiﬁers denoting parent types of t, and {idi : si : 1 ≤ i ≤ n} is the ﬁnite set of characteristics speciﬁed for object type t within its syntax. This set includes all the attributes, relationships and method signatures for the object type. The identiﬁer for a characteristic is idi and the speciﬁcation is si for each of the idi : si . See [4] for details of how rough set concepts are integrated in this OO model, and how changing the granularity of the partitioning aﬀects query results. In that paper the OO model is extended for fuzzy and rough set uncertainty. If we extend the rough OODB further to allow for intuitionistic types, the type speciﬁcations T can be generalized to a set Tˇ as in [4], so that the deﬁnitions of the domains are generalized to intuitionistic sets. For every ts ∈ T , having domain ts being domts , the type system ts ∈ T is generalized to ts ∈ Tˇ , where domain of ts is denoted by domts and is deﬁned as the set ρ(domts ) of intuitionistic sets on domts , and Ots is generalized to Ots , which contains the generalized version of the operators. type The generalized type system then is a triple GT S = [Tˇ, P, f impl ], where Tˇ type

is the generalized type system, P is the set of all programs, and f impl maps each ts ∈ Tˇ onto that subset of P that contains the implementation for ts. An type instance of this GTS is a generalized type t = [ts, f impl (ts)], ts ∈ Tˇ . A generalized object belonging to this class is deﬁned by o = [oid, N, t, type f impl (ts), ν], where ν draws values from the generalized domain that allows an object to contain intuitionistic membership and nonmembership values as part of the state of the object. Both intuitionistic and rough set uncertainty management can be used in this generalized OODB model. We extended a formal framework of object-oriented databases to allow for modeling of various types of imprecision, vagueness, and uncertainty that typically occur in spatial data. The model is based on a formal type system and speciﬁed constraints, thus preserving integrity of the database, while at the same time allowing an OODB to be generalized in such a way as to include both intuitionistic and rough set uncertainty, both well-developed methods of uncertainty management. Incorporation of intuitionistic and rough set uncertainty into the OODB model is essential for representing imprecision and uncertainty in spatial data entities and in their interrelationships.

5

Conclusion

We introduced the intuitionistic rough set, then discussed how the intuitionistic rough set generalizes each of traditional rough, fuzzy, fuzzy-rough, and

30

T. Beaubouef and F.E. Petry

intuitionistic sets. The intuitionistic rough relational database model was introduced. This model allows for both rough and intuitionistic modeling of uncertainty. Because real world applications involve uncertainty, this model can more accurately represent data and relationships than traditional relational databases. We have also introduced our model for intuitionistic object-oriented databases and shown the signiﬁcance of both rough sets and intuitionistic sets for uncertainty management.

References 1. Atanassov, K.: Intuitionistic Fuzzy Sets. Fuzzy Sets and Systems 20 (1986) 87–96 2. Beaubouef T., Petry F.: Fuzzy Rough Set Techniques for Uncertainty Processing in a Relational Database. International Journal of Intelligent Systems 15(5) (2000) 389–424 3. Beaubouef, T., Petry, F., Buckles, B.: Extension of the Relational Database and its Algebra with Rough Set Techniques. Computational Intelligence 11(2) (1995) 233–245 4. De Tr´e, G., De Caluwe, R. A Generalized Object-Oriented Database Model with Generalized Constraints. In Proc. of NAFIPS’99, New York (1999) 381–386. 5. Pawlak, Z.:Rough Sets. International Journal of Man-Machine Studies 21 (1984) 127–134

An Experimental Comparison of Three Rough Set Approaches to Missing Attribute Values Jerzy W. Grzymala-Busse1,2 and Witold J. Grzymala-Busse3 1

Department of Electrical Engineering and Computer Science University of Kansas, Lawrence, KS 66045, USA 2 Institute of Computer Science, Polish Academy of Sciences, 01–237 Warsaw, Poland 3 Touchnet Information Systems, Inc., Lenexa, KS 66219, USA

Abstract. In this paper we present results of experiments conducted to compare three types of missing attribute values: lost values, ”do not care” conditions and attribute-concept values. For our experiments we selected six well known data sets. For every data set we created 30 new data sets replacing speciﬁed values by three diﬀerent types of missing attribute values, starting from 10%, ending with 100%, with increment of 10%. For all concepts of every data set concept lower and upper approximations were computed. Error rates were evaluated using ten-fold cross validation. Overall, interpreting missing attribute values as lost provides the best result for most incomplete data sets. Keywords: missing attribute values, incomplete data sets, concept approximations, LERS data mining system, MLEM2 algorithm.

1

Introduction

Real-life data are frequently incomplete, i.e., values for some attributes are missing. Appropriate handling of missing attribute values is one of the most important tasks of data mining. In this paper we assume that missing attribute values have three diﬀerent interpretations. The ﬁrst possibility is that missing attribute values are lost. Such values are interpreted as originally speciﬁed, but currently unavailable since these values were incidentally erased, forgotten to be recorded, etc. A rough set approach to incomplete data sets in which all attribute values were lost was presented for the ﬁrst time in [12], where two algorithms for rule induction, modiﬁed to handle lost attribute values, were introduced. The next possibility are ”do not care” conditions. Such missing attribute values were irrelevant during collection of data. Simply, an expert decided that the attribute value was irrelevant for a classiﬁcation or diagnosis of the case. For example, a data set describing ﬂu patients may contain, among other attributes, an attribute Color of hair. Though some scrupulous patients may ﬁll in this value, other patients may assume that this attribute is irrelevant for the ﬂu diagnosis J.F. Peters et al. (Eds.): Transactions on Rough Sets VI, LNCS 4374, pp. 31–50, 2007. c Springer-Verlag Berlin Heidelberg 2007

32

J.W. Grzymala-Busse and W.J. Grzymala-Busse

and leave it unspeciﬁed. If we suspect that this attribute does matter, the best interpretation for missing attribute values is replacing them by all possible existing attribute values. A rough set approach to incomplete data sets in which all attribute values were ”do not care” conditions was presented for the ﬁrst time in [4], where a method for rule induction was introduced in which each missing attribute value was replaced by all values from the domain of the attribute. The third possibility is a missing attribute value interpreted as an attributeconcept value. It is a similar case to a ”do not care” condition, however, it is restricted to a speciﬁc concept. A concept (class) is a set of all cases classiﬁed (or diagnosed) the same way. Using this interpretation, we will replace a missing attribute value by all values of the same attribute typical for the concept to which the case belongs. Let us consider a patient, sick with ﬂu, from the ﬂu data set, with a missing attribute value for Color of hair. Other patients, sick with ﬂu, ﬁlled in values brown and grey for this attribute. On the other hand, healthy patients characterized the color of their hair as blond and brown. Using attribute-concept value interpretation, this missing attribute value is replaced by brown and grey. If we would use ”do not care” condition interpretation, the same missing attribute value should be replaced by blond, brown, and grey. This approach was introduced in [10]. In general, incomplete decision tables are described by characteristic relations, in a similar way as complete decision tables are described by indiscernibility relations [7,8,9]. In rough set theory, one of the basic notions is the idea of lower and upper approximations. For complete decision tables, once the indiscernibility relation is ﬁxed and the concept (a set of cases) is given, the lower and upper approximations are unique. For incomplete decision tables, for a given characteristic relation and concept, there are three important and diﬀerent possibilities to deﬁne lower and upper approximations, called singleton, subset, and concept approximations [7]. Singleton lower and upper approximations were studied in [14,15,19,21,22]. Note that similar deﬁnitions of lower and upper approximations, though not for incomplete decision tables, were studied in [16,24,25]. Note that some other rough-set approaches to missing attribute values were presented in [4,11,13,23] as well.

2

Blocks of Attribute-Value Pairs—Complete Data

We assume that the input data sets are presented in the form of a decision table. An example of a decision table is shown in Table 1. Rows of the decision table represent cases, while columns are labeled by variables. The set of all cases will be denoted by U . In Table 1, U = {1, 2, ..., 6}. Independent variables are called attributes and a dependent variable is called a decision and is denoted by d. The set of all attributes will be denoted by A. In Table 1, A = {Temperature, Headache, Cough}. Any decision table deﬁnes a function ρ that maps the direct product of U and A into the set of all values. For example, in

An Experimental Comparison of Three Rough Set Approaches

33

Table 1. A complete decision table Attributes

Decision

Case

Temperature

Headache

Cough

Flu

1 2 3 4 5 6

high very high high high normal normal

yes yes no yes yes no

yes no no yes no yes

yes yes no yes no no

Table 1, ρ(1, T emperature) = high. A decision table with completely speciﬁed function ρ will be called completely speciﬁed, or, for the sake of simplicity, complete. In practice, input data for data mining are frequently aﬀected by missing attribute values. In other words, the corresponding function ρ is incompletely speciﬁed (partial). A decision table with an incompletely speciﬁed function ρ will be called incomplete. Function ρ describing Table 1 is completely speciﬁed. An important tool to analyze complete decision tables is a block of the attribute-value pair. Let a be an attribute, i.e., a ∈ A and let v be a value of a for some case. For complete decision tables if t = (a, v) is an attribute-value pair then a block of t, denoted [t], is a set of all cases from U that for attribute a have value v. Rough set theory [17], [18] is based on the idea of an indiscernibility relation, deﬁned for complete decision tables. Let B be a nonempty subset of the set A of all attributes. The indiscernibility relation IN D(B) is a relation on U deﬁned for x, y ∈ U as follows (x , y) ∈ IND(B ) if and only if ρ(x , a) = ρ(y, a) for all a ∈ B . The indiscernibility relation IN D(B) is an equivalence relation. Equivalence classes of IN D(B) are called elementary sets of B and are denoted by [x]B . For example, for Table 1, elementary sets of IN D(A) are {1, 4}, {2}, {3}, {5}, {6}. Additionally, IND(B) = {(1, 1), (1, 4), (2, 2), (3, 3), (4, 1), (4, 4), (5, 5), (6, 6)}. The indiscernibility relation IN D(B) may be computed using the idea of blocks of attribute-value pairs. Let a be an attribute, i.e., a ∈ A and let v be a value of a for some case. For complete decision tables if t = (a, v) is an attribute-value pair then a block of t, denoted [t], is a set of all cases from U that for attribute a have value v. For Table 1, [(Temperature, high)] = {1, 3, 4}, [(Temperature, very high)] = {2}, [(Temperature, normal)] = {5, 6}, [(Headache, yes)] = {1, 2, 4, 5}, [(Headache, no)] = {3, 6},

34

J.W. Grzymala-Busse and W.J. Grzymala-Busse

[(Cough, yes)] = {1, 4, 6}, [(Cough, no)] = {2, 3, 5}. The indiscernibility relation IN D(B) is known when all elementary blocks of IND(B) are known. Such elementary blocks of B are intersections of the corresponding attribute-value pairs, i.e., for any case x ∈ U , [x]B = ∩{[(a, v)] | a ∈ B, ρ(x, a) = v}. We will illustrate the idea how to compute elementary sets of B for Table 1 and B = A. [1]A [2]A [3]A [5]A [6]A

= [4]A = {1, 3, 4} ∩ {1, 2, 4, 5} ∩ {1, 4, 6} = {1, 4}, = {2} ∩ {1, 2, 4, 5} ∩ {2, 3, 5} = {2}, = {1, 3, 4} ∩ {3, 6} ∩ {2, 3, 5} = {3}, = {5, 6} ∩ {1, 2, 4, 5} ∩ {2, 3, 5} = {5}, = {5, 6} ∩ {3, 6} ∩ {1, 4, 6} = {6},

For completely speciﬁed decision tables lower and upper approximations are deﬁned using the indiscernibility relation. Any ﬁnite union of elementary sets, associated with B, will be called a B-deﬁnable set. Let X be any subset of the set U of all cases. The set X is called a concept and is usually deﬁned as the set of all cases deﬁned by a speciﬁc value of the decision. In general, X is not a B-deﬁnable set. However, set X may be approximated by two B-deﬁnable sets, the ﬁrst one is called a B-lower approximation of X, denoted by BX and deﬁned as follows ∪{[x]B | x ∈ U, [x]B ⊆ X}, The second set is called a B-upper approximation of X, denoted by BX and deﬁned as follows ∪{[x]B | x ∈ U, [x]B ∩ X = ∅). Data set presented in Table 1 is consistent (the lower approximation is equal to the upper approximation for every concept), hence the certain rule set and the possible rule set are identical. Rules in the LERS format (every rule is equipped with three numbers, the total number of attribute-value pairs on the left-hand side of the rule, the total number of examples correctly classiﬁed by the rule during training, and the total number of training cases matching the left-hand side of the rule) [6] are: 2, 2, 2 (Temperature, high) & (Headache, yes) -> (Flu, yes) 1, 1, 1 (Temperature, very high) -> (Flu, yes) 1, 2, 2 (Temperature, normal) -> (Flu, no) 1, 2, 2 (Headache, no) -> (Flu, no)

An Experimental Comparison of Three Rough Set Approaches

35

Note that the above rules were induced by the MLEM2 (Modiﬁed Learning from Examples Module, version 2) option of the LERS (Learning from Examples based on Rough Sets) data mining system [2,5,6].

3

Blocks of Attribute-Value Pairs—Incomplete Data

For the rest of the paper we will assume that all decision values are speciﬁed, i.e., they are not missing. Additionally, we will assume that lost values will be denoted by ”?”, ”do not care” conditions by ”*”, and attribute-concept values by ”−”. Additionally, we will assume that for each case at least one attribute value is speciﬁed. Table 2 is Table 1 with eight attribute values missing. All of these missing attribute values are lost. Table 2. An incomplete decision table (all missing attribute values are lost values) Attributes

Decision

Case

Temperature

Headache

Cough

Flu

1 2 3 4 5 6

high ? ? high ? normal

? yes no ? yes no

yes ? ? yes no ?

yes yes no yes no no

For incomplete decision tables, a block of an attribute-value pair must be modiﬁed in the following way: – If for an attribute a there exists a case x such that ρ(x, a) = ?, i.e., the corresponding value is lost, then the case x should not be included in any blocks [(a, v)] for all values v of attribute a, – If for an attribute a there exists a case x such that the corresponding value is a ”do not care” condition, i.e., ρ(x, a) = ∗, then the case x should be included in blocks [(a, v)] for all speciﬁed values v of attribute a. – If for an attribute a there exists a case x such that the corresponding value is an attribute-concept value, i.e., ρ(x, a) = −, then the corresponding case x should be included in blocks [(a, v)] for all speciﬁed values v ∈ V (x, a) of attribute a, where V (x , a) = {ρ(y, a) | ρ(y, a) is speciﬁed , y ∈ U, ρ(y, d) = ρ(x, d)}.

36

J.W. Grzymala-Busse and W.J. Grzymala-Busse

Thus, for Table 2, [(Temperature, high)] = {1, 4}, [(Temperature, normal)] = {6}, [(Headache, yes)] = {2, 5}, [(Headache, no)] = {3, 6}, [(Cough, yes)] = {1, 4}, [(Cough, no)] = {5}. For incomplete data sets the idea of the elementary block is extended to a characteristic set. For a case x ∈ U the characteristic set KB (x) is deﬁned as the intersection of the sets K(x, a), for all a ∈ B, where the set K(x, a) is deﬁned in the following way: – If ρ(x, a) is speciﬁed, then K(x, a) is the block [(a, ρ(x, a)] of attribute a and its value ρ(x, a), – If ρ(x, a) = ? or ρ(x, a) = ∗ then the set K(x, a) = U , – If ρ(x, a) = −, then the corresponding set K(x, a) is equal to the union of all blocks of attribute-value pairs (a, v), where v ∈ V (x, a) if V (x, a) is nonempty. If V (x, a) is empty, K(x, a) = U . Thus, for Table 2 KA (1) = {1, 4} ∩ U ∩ {1, 4} = {1, 4}, KA (2) = U ∩ {2, 5} ∩ U = {2, 5}, KA (3) = U ∩ {3, 6} ∩ U = {3, 6}, KA (4) = {1, 4} ∩ U ∩ {1, 4} = {1, 4}, KA (5) = U ∩ {2, 5} ∩ {5} = {5}, KA (6) = {6} ∩ {3, 6} ∩ U = {6}, Characteristic set KB (x) may be interpreted as the set of cases that are indistinguishable from x using all attributes from B and using a given interpretation of missing attribute values. Thus, KA (x) is the set of all cases that cannot be distinguished from x using all attributes. In [24] KA (x) was called a successor neighborhood of x, see also [16,19,24,25]. Obviously, when a data set is complete, for given B ⊆ A, all characteristic sets KB (x) are identical with elementary blocks [x]B . The characteristic relation R(B) is a relation on U deﬁned for x, y ∈ U as follows (x , y) ∈ R(B ) if and only if y ∈ KB (x ). The characteristic relation R(B) is reﬂexive but—in general—does not need to be symmetric or transitive. Also, the characteristic relation R(B) is known if we know characteristic sets KB (x) for all x ∈ U . In our example, R(A) = {(1, 1), (1, 4), (2, 2), (2, 5), (3, 3), (3, 6), (4, 1), (4, 5), (5, 5), (6, 6)}. The most convenient way to deﬁne the characteristic relation is through the characteristic sets. For decision tables, in which all missing attribute values are lost, a special characteristic relation was deﬁned in [21], see also, e.g., [20,22]. For incompletely speciﬁed decision tables lower and upper approximations may be deﬁned in a few diﬀerent ways. First, the deﬁnition of deﬁnability should

An Experimental Comparison of Three Rough Set Approaches

37

Fig. 1. Bankruptcy data—certain rule sets

be modiﬁed. A union of some intersections of attribute-value pair blocks, in any such intersection all attributes should be diﬀerent and attributes are members of B, will be called B-locally deﬁnable sets. A union of characteristic sets KB (x), where x ∈ X ⊆ U will be called a B-globally deﬁnable set. Any set X that is B -globally deﬁnable is B -locally deﬁnable, the converse is not true. In this paper we quote three diﬀerent deﬁnitions of lower and upper approximations [7,8,9]. Let X be a concept, let B be a subset of the set A of all attributes, and let R(B) be the characteristic relation of the incomplete decision table with characteristic sets K(x), where x ∈ U . Our ﬁrst deﬁnition uses a similar idea as in the previous articles on incompletely speciﬁed decision tables [14,15,20,21,22], i.e., lower and upper approximations are sets of singletons from the universe U satisfying some properties. Thus, lower and upper approximations are deﬁned by constructing both sets from singletons. We will call these approximations singleton. Namely, a singleton B-lower approximation of X is deﬁned as follows: BX = {x ∈ U | KB (x) ⊆ X}. A singleton B-upper approximation of X is BX = {x ∈ U | KB (x) ∩ X = ∅}. In our example of the decision table presented in Table 2 let us say that B = A. Then the singleton A-lower and A-upper approximations of the two concepts: {1, 2, 4} and {3, 5, 6} are:

38

J.W. Grzymala-Busse and W.J. Grzymala-Busse

Fig. 2. Bankruptcy data—possible rule sets

A{1, 2, 4} = {1, 4}, A{3, 5, 6} = {3, 5, 6}, A{1, 2, 4} = {1, 2, 4}, A{3, 5, 6} = {2, 3, 5, 6}. Note that the set A{1, 2, 4} is not even A-locally deﬁnable. Hence, as it was previously argued in [7,8,9], singleton approximations should not be used for rule induction. Obviously, if a set is not B-locally deﬁnable then it cannot be expressed by rule sets using attributes from B. We may deﬁne lower and upper approximations for incomplete decision tables by using characteristic sets instead of elementary sets. There are two ways to do this. Using the ﬁrst way, a subset B-lower approximation of X is deﬁned as follows: BX = ∪{KB (x) | x ∈ U, KB (x) ⊆ X}. A subset B-upper approximation of X is BX = ∪{KB (x) | x ∈ U, KB (x) ∩ X = ∅}.

An Experimental Comparison of Three Rough Set Approaches

39

Fig. 3. Breast cancer (Slovenia) data—certain rule sets

Since any characteristic relation R(B) is reﬂexive, for any concept X, singleton B-lower and B-upper approximations of X are subsets of the subset B-lower and B-upper approximations of X, respectively [9]. For the same decision table, presented in Table 2, the subset A-lower and A-upper approximations are A{1, 2, 4} = {1, 4}, A{3, 5, 6} = {3, 5, 6}, A{1, 2, 4} = {1, 2, 4, 5}, A{3, 5, 6} = {2, 3, 5, 6}. The second possibility is to modify the subset deﬁnition of lower and upper approximation by replacing the universe U from the subset deﬁnition by a concept X. A concept B-lower approximation of the concept X is deﬁned as follows: BX = ∪{KB (x) | x ∈ X, KB (x) ⊆ X}. Obviously, the subset B-lower approximation of X is the same set as the concept B-lower approximation of X [7]. A concept B-upper approximation of the concept X is deﬁned as follows:

40

J.W. Grzymala-Busse and W.J. Grzymala-Busse

Fig. 4. Breast cancer (Slovenia) data—possible rule sets

BX = ∪{KB (x) | x ∈ X, KB (x) ∩ X = ∅} = ∪{KB (x) | x ∈ X}. The concept B-upper approximation of X is a subset of the subset B-upper approximation of X [7]. For the decision table presented in Table 2, the concept A-lower and A-upper approximations are A{1, 2, 4} = {1, 4}, A{3, 5, 6} = {3, 5, 6}, A{1, 2, 4} = {1, 2, 4, 5}, A{3, 5, 6} = {2, 3, 5, 6}. Note that for complete decision tables, all three deﬁnitions of lower approximations, singleton, subset and concept, coalesce to the same deﬁnition. Also, for complete decision tables, all three deﬁnitions of upper approximations coalesce to the same deﬁnition. This is not true for incomplete decision tables, as our example shows. For Table 2, certain rules [3], induced from the concept lower approximations are

An Experimental Comparison of Three Rough Set Approaches

41

Fig. 5. Breast cancer (Wisconsin) data—certain rule sets

1, 2, 2 (Temperature, high) -> (Flu, yes) 1, 2, 2 (Headache, no) -> (Flu, no) 1, 1, 1 (Cough, no) -> (Flu, no) and possible rules [3], induced from the concept upper approximations, are 1, 2, 2 (Temperature, high) -> (Flu, yes) 1, 2, 2 (Headache, yes) -> (Flu, yes) 1, 2, 2 (Headache, no) -> (Flu, no) 1, 1, 1 (Cough, no) -> (Flu, no) Table 3 shows a modiﬁcation of Table 2, where all lost values are replaced by ”do not care” conditions. For decision tables where all missing attribute values are ”do not care” conditions a special characteristic relation was deﬁned in [14], see also, e.g., [15]. Blocks of attribute-value pairs are [(Temperature, high)] = {1, 2, 3, 4, 5}, [(Temperature, normal)] = {2, 3, 5, 6},

42

J.W. Grzymala-Busse and W.J. Grzymala-Busse

Table 3. An incomplete decision table (all missing attribute values are lost values) Attributes

Decision

Case

Temperature

Headache

Cough

Flu

1 2 3 4 5 6

high * * high * normal

* yes no * yes no

yes * * yes no *

yes yes no yes no no

Fig. 6. Breast cancer (Wisconsin) data—possible rule sets

[(Headache, yes)] = {1, 2, 4, 5}, [(Headache, no)] = {1, 3, 4, 6}, [(Cough, yes)] = {1, 2, 3, 4, 6}, [(Cough, no)] = {2, 3, 5, 6}. Characteristic sets are KA (1) = {1, 2, 3, 4, 5} ∩ U ∩ {1, 2, 3, 4, 6} = {1, 2, 3, 4}, KA (2) = U ∩ {1, 2, 4, 5} ∩ U = {1, 2, 4, 5}, KA (3) = U ∩ {1, 3, 4, 6} ∩ U = {1, 3, 4, 6},

An Experimental Comparison of Three Rough Set Approaches

43

KA (4) = {1, 2, 3, 4, 5} ∩ U ∩ {1, 2, 3, 4, 6} = {1, 2, 3, 4}, KA (5) = U ∩ {1, 2, 4, 5} ∩ {2, 3, 5, 6} = {2, 5}, KA (6) = {2, 3, 5, 6} ∩ {1, 3, 4, 6} ∩ U = {3, 6}, For the decision table presented in Table 3, the concept A-lower and A-upper approximations are A{1, 2, 4} = ∅, A{3, 5, 6} = {3, 6}, A{1, 2, 4} = {1, 2, 3, 4, 5}, A{3, 5, 6} = U. In our example, the concept A-lower approximation of {1, 2, 4} is the empty set. With large percentage of missing attribute values interpreted as ”do not care” conditions, empty lower approximations cause large increases of error rates during ten-fold cross validation. For Table 3, the only certain rule, induced from the concept lower approximation, is

Fig. 7. Image segmentation data—certain rule sets

44

J.W. Grzymala-Busse and W.J. Grzymala-Busse

Fig. 8. Image segmentation data—possible rule sets

2, 2, 2 (Temperature, normal) & (Headache, no) -> (Flu, no) and possible rules, induced from the concept upper approximations, are 1, 3, 5 (Temperature, high) -> (Flu, yes) 1, 2, 5 (Temperature, high) -> (Flu, no) 1, 3, 4 (Temperature, normal) -> (Flu, no) Table 4 is another modiﬁcation of Table 2, where all lost values are replaced by attribute-concept values. Blocks of attribute-value pairs are [(Temperature, high)] = {1, 2, 4}, [(Temperature, normal)] = {3, 5, 6}, [(Headache, yes)] = {1, 2, 4, 5}, [(Headache, no)] = {3, 6}, [(Cough, yes)] = {1, 2, 4}, [(Cough, no)] = {3, 5, 6}. Characteristic sets are KA (1) = {1, 2, 4} ∩ {1, 2, 4, 5} ∩ {1, 2, 4} = {1, 2, 4}, KA (2) = {1, 2, 4} ∩ {1, 2, 4, 5} ∩ {1, 2, 4} = {1, 2, 4}, KA (3) = {3, 5, 6} ∩ {3, 6} ∩ {3, 5, 6} = {3, 6},

An Experimental Comparison of Three Rough Set Approaches

45

Table 4. An incomplete decision table (all missing attribute values are lost values) Attributes

Decision

Case

Temperature

Headache

Cough

Flu

1 2 3 4 5 6

high – – high – normal

– yes no – yes no

yes – – yes no –

yes yes no yes no no

Fig. 9. Iris data—certain rule sets

KA (4) = {1, 2, 4} ∩ {1, 2, 4, 5} ∩ {1, 2, 4} = {1, 2, 4}, KA (5) = {3, 5, 6} ∩ {1, 2, 4, 5} ∩ {3, 5, 6} = {5}, KA (6) = {3, 5, 6} ∩ {3, 6} ∩ {3, 5, 6} = {3, 6}, For the decision table presented in Table 4, the concept A-lower and A-upper approximations are A{1, 2, 4} = {1, 2, 4}, A{3, 5, 6} = {3, 5, 6},

46

J.W. Grzymala-Busse and W.J. Grzymala-Busse

Fig. 10. Iris data—possible rule sets

A{1, 2, 4} = {1, 2, 4}, A{3, 5, 6} = {3, 5, 6}. For Table 4, certain rules, induced from the concept lower approximations, are identical with possible rules, induced from concept upper approximations 1, 3, 3 (Temperature, high) -> (Flu, yes) 1, 3, 3 (Temperature, normal) -> (Flu, no)

4

Experiments

For our experiments six typical data sets were used, see Table 5. These data sets were complete (all attribute values were completely speciﬁed), with the exception of breast cancer (Slovenia) data set, which originally contained 11 cases (out of 286) with missing attribute values. These 11 cases were removed. In two data sets: bankruptcy and iris all attributes were numerical. These data sets were processed as numerical (i.e., discretization was done during rule induction by MLEM2). The image segmentation data set was converted into symbolic using a discretization method based on agglomerative cluster analysis (this method was described, e.g., in [1]).

An Experimental Comparison of Three Rough Set Approaches

47

Table 5. Data sets used for experiments Data set cases Bankruptcy Breast cancer (Slovenia) Breast cancer (Wisconsin) Image segmentation Iris Lymphography

66 277 625 210 150 148

Number of attributes

concepts

5 9 9 19 4 18

2 2 9 7 3 4

Fig. 11. Lymphography data—certain rule sets

To each data set we conducted a series of three experiments, adding incrementally (with 10% increment) missing attribute values of three diﬀerent types. Thus, we started each series of experiments with no missing attribute values, then we added 10% of missing attribute values of given type, then we added additional 10% of missing attribute values of the same type, etc., until reaching a level of 100% missing attribute values. For each data set and a speciﬁc type of missing attribute values ten additional data sets were created. Furthermore, for each data set with some percentage of missing attribute values, experiments were conducted separately for certain and possible rule sets, using concept lower and upper approximations, respectively. Ten-fold cross

48

J.W. Grzymala-Busse and W.J. Grzymala-Busse

Fig. 12. Lymphography data—possible rule sets

validation was used to compute an error rate. Rule sets were induced by the MLEM2 option of the LERS data mining system [2,5,6]. Results of our experiments are presented in Figures 1–12. In all 12 ﬁgures, lost values, ”do not care” conditions, and attribute-concept values denote percentage of error rate for experiments with missing attribute values interpreted as lost values, ”do not care” conditions, and attribute-concept values, respectively.

5

Conclusions

During all series of experiments the error rate was aﬀected by large variance. Moreover, for some data sets (e.g., breast cancer (Wisconsin)), adding a small amount of missing attribute values resulted in a decreased error rate. Most likely, in these data sets, attributes aﬀected by missing attribute values were not important. In eﬀect, the induced rule sets were more general and better. It is clear that inducing certain rule sets while using a ”do not care” condition approach to missing attribute values was the worst approach. This was caused by the fact that lower approximations of concepts, with large number of missing attribute values, were empty. Another surprising conclusion is that for some data sets (breast cancer (Slovenia) and breast cancer (Wisconsin)) adding a large number of missing attribute values does not aﬀect the error rate seriously—the error rate was almost the same for larger and larger number of missing attribute values.

An Experimental Comparison of Three Rough Set Approaches

49

Overall, it seems that the interpretation of missing attribute values as lost is the best approach among our three types of missing attribute value interpretations. Taking into account a large variance, the diﬀerence between error rates for certain and possible rule sets is negligible.

References 1. Chmielewski, M.R. and Grzymala-Busse, J.W.: Global discretization of continuous attributes as preprocessing for machine learning. Int. Journal of Approximate Reasoning 15 (1996) 319–331. 2. Chan, C.C. and Grzymala-Busse, J.W.: On the attribute redundancy and the learning programs ID3, PRISM, and LEM2. Department of Computer Science, University of Kansas, TR-91-14, December 1991, 20 pp. 3. Grzymala-Busse, J.W.: Knowledge acquisition under uncertainty—A rough set approach. Journal of Intelligent & Robotic Systems 1 (1988), 3–16. 4. Grzymala-Busse, J.W.: On the unknown attribute values in learning from examples. Proc. of the ISMIS-91, 6th International Symposium on Methodologies for Intelligent Systems, Charlotte, North Carolina, October 16–19, 1991. Lecture Notes in Artiﬁcial Intelligence, vol. 542, Springer-Verlag, Berlin, Heidelberg, New York (1991) 368–377. 5. Grzymala-Busse, J.W.: LERS—A system for learning from examples based on rough sets. In Intelligent Decision Support. Handbook of Applications and Advances of the Rough Sets Theory, ed. by R. Slowinski, Kluwer Academic Publishers, Dordrecht, Boston, London (1992) 3–18. 6. Grzymala-Busse., J.W.: MLEM2: A new algorithm for rule induction from imperfect data. Proceedings of the 9th International Conference on Information Processing and Management of Uncertainty in Knowledge-Based Systems, IPMU 2002, July 1–5, Annecy, France, 243–250. 7. Grzymala-Busse, J.W.: Rough set strategies to data with missing attribute values. Workshop Notes, Foundations and New Directions of Data Mining, the 3-rd International Conference on Data Mining, Melbourne, FL, USA, November 19–22, 2003, 56–63. 8. Grzymala-Busse, J.W.: Data with missing attribute values: Generalization of idiscernibility relation and rule induction. Transactions on Rough Sets, Lecture Notes in Computer Science Journal Subline, Springer-Verlag, vol. 1 (2004) 78–95. 9. Grzymala-Busse, J.W.: Characteristic relations for incomplete data: A generalization of the indiscernibility relation. Proc. of the RSCTC’2004, the Fourth International Conference on Rough Sets and Current Trends in Computing, Uppsala, Sweden, June 1–5, 2004. Lecture Notes in Artiﬁcial Intelligence 3066, SpringerVerlag 2004, 244–253. 10. Grzymala-Busse, J.W.: Three approaches to missing attribute valuesA rough set perspective. Proceedings of the Workshop on Foundation of Data Mining, associated with the Fourth IEEE International Conference on Data Mining, Brighton, UK, November 1–4, 2004, 55–62. 11. Grzymala-Busse, J.W. and Hu, M.: A comparison of several approaches to missing attribute values in data mining. Proceedings of the Second International Conference on Rough Sets and Current Trends in Computing RSCTC’2000, Banﬀ, Canada, October 16–19, 2000, 340–347.

50

J.W. Grzymala-Busse and W.J. Grzymala-Busse

12. Grzymala-Busse, J.W. and Wang A.Y.: Modiﬁed algorithms LEM1 and LEM2 for rule induction from data with missing attribute values. Proc. of the Fifth International Workshop on Rough Sets and Soft Computing (RSSC’97) at the Third Joint Conference on Information Sciences (JCIS’97), Research Triangle Park, NC, March 2–5, 1997, 69–72. 13. Hong, T.P., Tseng L.H. and Chien, B.C.: Learning coverage rules from incomplete data based on rough sets. Proc. of the IEEE International Conference on Systems, Man and Cybernetics, Hague, the Netherlands, October 10–13, 2004, 3226–3231. 14. Kryszkiewicz, M.: Rough set approach to incomplete information systems. Proc. of the Second Annual Joint Conference on Information Sciences, Wrightsville Beach, NC, September 28–October 1, 1995, 194–197. 15. Kryszkiewicz, M.: Rules in incomplete information systems. Information Sciences 113 (1999) 271–292. and knowledge base systems. Fourth International Symposium on Methodologies of Intelligent Systems (Poster Sessions), Charlotte, North Carolina, October 12–14, 1989, 75–86. Tucson, Arizona, December 4–8, 1989, 286–293. 16. Lin, T.Y.: Topological and fuzzy rough sets. In Intelligent Decision Support. Handbook of Applications and Advances of the Rough Sets Theory, ed. by R. Slowinski, Kluwer Academic Publishers, Dordrecht, Boston, London (1992) 287–304. 17. Pawlak, Z.: Rough Sets. International Journal of Computer and Information Sciences 11 (1982) 341–356. 18. Pawlak, Z.: Rough Sets. Theoretical Aspects of Reasoning about Data. Kluwer Academic Publishers, Dordrecht, Boston, London (1991). 19. Slowinski, R. and Vanderpooten, D.: A generalized deﬁnition of rough approximations based on similarity. IEEE Transactions on Knowledge and Data Engineering 12 (2000) 331–336. 20. Stefanowski, J.: Algorithms of Decision Rule Induction in Data Mining. Poznan University of Technology Press, Poznan, Poland (2001). 21. Stefanowski, J. and Tsoukias, A.: On the extension of rough sets under incomplete information. Proc. of the 7th International Workshop on New Directions in Rough Sets, Data Mining, and Granular-Soft Computing, RSFDGrC’1999, Ube, Yamaguchi, Japan, November 8–10, 1999, 73–81. 22. Stefanowski, J. and Tsoukias, A.: Incomplete information tables and rough classiﬁcation. Computational Intelligence 17 (2001) 545–566. 23. Wang, G.: Extension of rough set under incomplete information systems. Proc. of the IEEE International Conference on Fuzzy Systems (FUZZ IEEE’2002), vol. 2, Honolulu, HI, May 12–17, 2002, 1098–1103. 24. Yao, Y.Y.: Relational interpretations of neighborhood operators and rough set approximation operators. Information Sciences 111 (1998) 239–259. 25. Yao, Y.Y. and Lin, T.Y.: Generalization of rough sets using modal logics. Intelligent Automation and Soft Computing 2 (1996) 103–119.

Pawlak’s Landscaping with Rough Sets Mihir K. Chakraborty Department of Pure Mathematics, University of Calcutta 35 Ballygunge Circular Road, Kolkata 700019, India [email protected]

Abstract. This paper reviews, rather non-technically, Pawlak’s approach to vagueness through rough sets and looks for a foundation of rough sets in an early work of Obtulowicz. An extension of Obtulowicz’s proposal is suggested that in turn, hints at a uniﬁed approach to rough sets and fuzzy sets.

1

Introduction

The concluding decades of the past century have added several outstanding, signiﬁcant and elegant contributions to human knowledge of which Rough Set Theory is one. Zdzislaw Pawlak, a Professor of Computer Science from Poland, ﬁrst proposed this theory in 1982 through his publication entitled ‘Rough Sets’ [20]. Surprisingly, this is again a contribution to humanity from one belonging to the ﬁeld of computer science – during the same period, the same community gifted several other elegant creations, like Fuzzy Set Theory by Lotﬁ Zadeh in 1965. It is also interesting to note that both the theories address basically the same issue, viz. ‘vagueness’ and this fact is not merely a coincidence. ‘Vagueness’ had been an outstanding issue. Great minds of the antiquity, both of the East and the West delved into the notion exhibited in various forms. (Theseuses’ ship [43,17], the Sorites [43,13], or the tetra-lemma (Catuskoti) [35,32], for example). Following Enlightenment, with the rise of modern rationality, embodied in the methods of physical sciences, more speciﬁcally physics, ‘vagueness’ had been gradually pushed aside to the fringes like the indigenous population in Australia or in America and other places. Use of imprecise terms were not only marginalized, but virtually banished from all serious discourses as expressed by the rationalist, humanist Bertrand Russell in the lines (in, Our Knowledge of External World as a Field of Scientiﬁc Method in Philosophy) (cf. Hao Wang [42]): “The study of logic becomes the central study in philosophy: it gives the method of research in philosophy, just as mathematics gives the method in physics;”

I would like to express my sincere thanks to Smita Sirker, Dept. of Philosophy, Jadavpur University, for kindly reading the ﬁrst draft and making valuable comments, particularly on the main philosophy of this article viz. the relationship between the existence of an object in a concept and indiscernibility.

J.F. Peters et al. (Eds.): Transactions on Rough Sets VI, LNCS 4374, pp. 51–63, 2007. c Springer-Verlag Berlin Heidelberg 2007

52

M.K. Chakraborty

and further from his article ‘Vagueness’ (Australian Journal of Philosophy)[37]: “Logical words, like the rest, when used by human beings, share the vagueness of all other words. There is however, less vagueness about logical words than about the words of daily life.” However, in spite of the great tide of modern rationality all over the world, a sense of understanding that there exists an essential role of vagueness in human knowledge system as well as life, was not totally wiped out from Eastern thoughts. Ironically, the most advanced technology – computer science, the most recent gift of modernity, has ushered the study of vagueness spectacularly from the disrespectful margin straightaway to centre-stage. That ‘vagueness’ in general is diﬀerent from ‘probability’ has somewhat been accepted nowadays after the long, ﬁerce debates that took place during the years immediately following the advent of fuzzy set theory in 1965. So Pawlak did not have to ﬁght that battle. Yet he had to utter this warning which is an excellent distinctive criterion viz. “Vagueness is the property of sets... whereas uncertainty is the property of an element”[28]. Uncertainty leads to probabilistic studies. It is often said of course, that vagueness is uncertainty too but not of probabilistic kind. However, right from the beginning Pawlak wanted to point at the distinction between Rough Set theory and Fuzzy Set theory. In the introduction to his short communication [23] he declares “we compare this concept with that of the fuzzy set and we show that these two concepts are diﬀerent.” Diﬀerent in what sense? Early Pawlak (during the 80s) was ﬁrm in his belief that Rough Set is properly addressing vagueness since it talks about ‘boundaries’ of a set and the property ‘rough’ is ascribed to a set. On the other hand, although the qualiﬁer ‘fuzzy’ has been ascribed to sets too, in reality the theory deals with degree of membership of an object in a ‘set’ and hence is dealing with some kind of uncertainty of belongingness of objects. So according to the above quoted norm, fuzzy set theory is not addressing vagueness proper. However, in later Pawlak, perhaps a change in opinion is observed as reﬂected in the following categorical remark “Both fuzzy and rough set theory represent two diﬀerent approaches to vagueness. Fuzzy set theory addresses gradualness of knowledge, expressed by the fuzzy membership - whereas rough set theory addresses granularity of knowledge expressed by indiscernibility relation” [22]. We shall discuss the role of indiscernibility to some length in the foundations of fuzzy set theory as well as rough set theory and thus in vagueness. But it needs to be mentioned that the relationship between the two theories had quite naturally been a favourite topic of study in those turbulent decades. For example, Wygralak in the 1985 BUSEFAL Conference presented a paper [44] in which he established that basic operations on rough sets (i.e. union and intersection) can be expressed as some special operations as their membership functions. Pawlak already talked about the distinction and the irreducibility of rough sets to fuzzy sets. This point needs a little clariﬁcation since rough sets approximations are not distributive with respect to all set theoretic operations. Pawlak in [23] checked for the natural candidate for representation of rough sets by the 3-valued membership function

Pawlak’s Landscaping with Rough Sets

53

⎧ ⎨ 1 if x belongs to the lower approximation X(x) = .5 if x belongs to the boundary region ⎩ 0 otherwise. With such functions representing rough sets X and Y , the membership function of X ∪ Y can not be represented by max(X(x), Y (x)) as is done in fuzzy set theory, since it has to coincide with the function ⎧ ⎨ 1 if x belongs to the lower approximation of X ∪ Y X ∪ Y (x) = .5 if x belongs to the boundary region ⎩ 0 otherwise, but which fails. Similarly, X ∩ Y can not be represented by the function min(X(x), Y (x)). However if the membership function is modiﬁed as below the desired result is obtained. Wygralak proposed to deﬁne and of two rough sets X, Y , by min(1, X(x) + Y (x)) if X(x) = Y (x) = .5 and [x] ⊆ X ∪ Y X Y (x) = max(X(x), Y (x)) otherwise. and XY (x) =

max(0, X(x) + Y (x) − 1) if X(x) = Y (x) = .5 and [x] ∩ X ∩ Y = ∅ min(X(x), Y (x)) otherwise.

X Y is roughly equal to X ∪ Y and X Y is roughly equal to X ∩ Y . In spite of this claim of Wygralak, one may doubt about the acceptability or otherwise of such functions as operators for conjunction and disjunction because they might miss some important properties - but it requires detailed investigations to make a ﬁnal comment. In Moscow Conference, 1988, Dubois and Prade argue that “fuzzy sets and rough sets aim to diﬀerent purposes and that it is more natural to try to combine the two models of uncertainty (vagueness for fuzzy sets and coarseness for rough sets) in order to get a more accurate account of imperfect information.” They proposed interesting mathematical constructs known as rough-fuzzy sets and fuzzy-rough sets [9]. In the opinion of the present author, there is an essential indiscernibility underlying in all kinds of vagueness - indiscernibility giving rise to both granularity and gradualness. But indiscernibles may be of various types though one can probe into some essential features of this elusive notion. These investigations had also been a favourite topic in the 80s and 90s [1,34,41,12,14,10].

2

Indiscernibilities

One major diﬀerence in the approaches to indiscernibility lies in assuming it as a ‘yes’/‘no’ type crisp concept or a graded concept. In the ﬁrst approach, objects

54

M.K. Chakraborty

x and y are either indiscernible or not - this approach is adopted in Rough Set theory by Pawlak. Several generalizations, of course, have taken place. The general feature of these is that starting from some knowledge or data, diﬀerent components (or clusters) are computed and rough sets are construed out of these components. We shall, however, restrict this study to Pawlak-rough sets only. It would be interesting to investigate if the present ‘uniﬁed’ approach can be extended to generalized rough sets as well. The second approach, as pursued in Fuzzy Set theory, presumes that x and y may be indiscernible to a degree - that is, it is a graded notion. Another very important diﬀerence rests in assuming or not assuming or weakening the transitivity property of indiscernibility. In the ﬁrst case we get standard equivalence relation on the universe of discourse, as is the base of Pawlak-rough sets. The second case generates tolerance relation (reﬂexive-symmetric) as base – rough set theory on this base is also pursued. In the third case, we have fuzzy transitivity, viz. Ind(x, y) & Ind(y, z) ≤ Ind(x, z), where Ind(x, y) represents the indiscernibility degree between x and y, and & is an algebraic operation (a t-norm, perhaps) on a suitable truth set. This graded relation without being reduced to tolerance, relaxes the notion of hard transitivity and elegantly, takes care of the gradualness aspect by using an interactive conjunction operator as follows. Let x1 , x2 , x3 , x4 , ... be a sequence of objects such that Ind(xi , xi+1 ) = .5, for all i. Now let us take the product (×) as the operator for &. Since Ind(x1 , x2 ) & Ind(x2 , x3 ) ≤ Ind(x1 , x3 ), we get Ind(x1 , x3 ) ≥ .5 × .5 = .25. If the least value .25 is taken then Ind(x1 , x4 ) may be taken as Ind(x1 , x3 ) & Ind(x3 , x4 ) = .25 × .5 = .125. Thus indiscernibility degree gradually diminishes. It means that the indiscernibility between x1 and x4 is less than that between x1 and x3 , and this is further less than the indiscernibility between x1 and x2 – a feature quite intuitively acceptable. Symmetry is naturally expected of indiscernibility. In the fuzzy case it means that Ind(x, y) = Ind(y, x). We shall discuss about reﬂexivity property later. Before that let us examine the relationship between indiscernibility relation and a concept. A concept induces an indiscernibility relation in a universe of discourse. For example, the concept A gives rise to IndA given by 1 if x, y ∈ A or x, y ∈ Ac IndA (x, y) = 0 otherwise. Concepts A, B, and C similarly give rise to the relation 1 if x ∈ X if and only if y ∈ X, X = A, B, C IndA,B,C (x, y) = 0 otherwise. An instance of this latter case is depicted in the following diagram with ﬁfteen objects, x1 to x15 .

Pawlak’s Landscaping with Rough Sets

x1

x2

x14

x4

x3 x15

x5

x13 x9 x7 x12

55

x10

x6

x8

x11

We could also show IndA,B,C by constructing the following approximation space or information system. x1 x2 x3 x4 x5 x6 x7 x8 x9 x10 x11 x12 x13 x14 x15

A Y Y Y N Y N N N Y Y N N Y Y N

B N Y Y Y Y Y N N N Y N N N N Y

C N N N N Y Y Y Y Y Y N N N N N

Y stands for ‘yes, belongs to’ and N for ‘no, does not belong to’. So IndA,B,C (x, y) = 1, if and only if the rows corresponding to x and y are identical. Thus a set of concepts generates a partition of the universe and hence equivalently gives an approximation space. Can we retrieve A (in the ﬁrst example) or A, B, C (in the second example) from the indiscernibility relations IndA or IndA,B,C ? The two values 0 and 1 of the indiscernibility relation are not suﬃcient for this purpose. Only from the

56

M.K. Chakraborty

information about indiscernibility one cannot separate A from its complement Ac since IndA (x, x) = 1 whether x ∈ A or x ∈ Ac . So, let us take a third value 2 and stipulate. 2 if x ∈ A Ind∗A (x, x) = 1 if x ∈ Ac and Ind∗A (x, y)

=

1 if x = y, x, y ∈ A or x, y ∈ Ac 0 if x = y, x ∈ A and y ∈ Ac , or x ∈ Ac and y ∈ A.

It is obvious that IndA (x, y) = 1 if and only if Ind∗A (x, y) ≥ 1 and IndA (x, y) = 0 if and only if Ind∗A (x, y) = 0. The two functions IndA and Ind∗A coincide on (x, y), x = y, but while the ﬁrst one cannot make a distinction between A and its complement, Ind∗A can do this. Similarly, in order to retrieve A ∪ B ∪ C in the second case, we need the deﬁnition 2 if x ∈ A ∪ B ∪ C Ind∗A,B,C (x, x) = 1 if x ∈ A ∪ B ∪ C and

⎧ ⎨ 1 if x = y and the rows of x and y in the information Ind∗A,B,C (x, y) = system coincide ⎩ 0 otherwise. One point to be observed here is that although while Ind∗X (x, y) ≥ 1 implies that x and y are indiscernible in terms of the basic concepts, Ind∗X (x, x), i.e. the indiscernibility of x with itself is not of the same category for all x. Also to be noted that this representation may be applied to any subset (not only A∪B ∪C) of U obtained by unions of intersections of A, B and C, in other words unions of the indiscernibility classes or blocks in U determined by them. The procedure, however, is to deﬁne Ind∗A,B,C (x, x), so that only speciﬁed unions of blocks are obtained in the backward process. We shall summarize and axiomatize such properties of Ind∗X later. At this stage we only raise an issue. Could we not think that degree of belongingness of an object to a concept is determined by and is the same as the degree of indiscernibility of the object with itself relative to the same concept? Looking from this angle, indiscernibility becomes more a primitive notion and plays a key role in the process of learning as well as of categorization. A category is created depending on the similarity of its members. Based on the degree of indiscernibility of an object, its belongingness to a set/category/class is decided. x belongs to A to the extent 2 = Ind∗A (x, x), means x is within A, x belongs to A to the extent 1 = Ind∗A (x, x) means x is in Ac . That we are interested in A and not in its complement is also represented in the assignment of a lower degree of indiscernibility to elements of the complement with themselves. The idea shall

Pawlak’s Landscaping with Rough Sets

57

play a key role in what follows and is a basis for enquiries into vagueness. Even in the case of crisp concepts, we have noticed, while retrieving A from Ind∗A the elements of A have grade 2, and elements of Ac have grade 1, and a third value 0 is also needed to construct the partition. If, however, the concepts are vague, that is, admit borderline cases, it is necessary to introduce a fourth value and the indiscernibility generated may be deﬁned as above with certain additional conditions. These ideas were published long back in Obtulowicz’s paper [18] based on a ﬁrm category-theoretic basis proposed by Higgs [12], but probably escaped attention of the researchers in this ﬁeld. We re-present from his work the main deﬁnitions and representation theorems below with little notational changes to ﬁt in the present context. Let U be a universe and L(4) be the ordered set {0 ≤ 1 ≤ 2 ≤ 3} which is a complete distributive lattice (or complete Heyting algebra). Let Ind∗ : U × U → L(4) be an indiscernibility relation that satisﬁes conditions. H1 : Ind∗ (x, y) = Ind∗ (y, x) (Symmetry) H2 : Ind∗ (x, y) ∧ Ind∗ (y, z) ≤ Ind∗ (x, z) (Transitivity) and the following roughness conditions R1 R2 R3 R4

: : : :

1 ≤ Ind∗ (x, x) for x ∈ U if 2 ≤ Ind∗ (x, y), then x = y if Ind∗ (x, y) = 1, then Ind∗ (x, x) = Ind∗ (y, y) if Ind∗ (x, x) = 2, then there exists y such that Ind∗ (x, y) = 1

The signiﬁcance of roughness conditions shall be clear from Proposition 2 below. The following two propositions establish that (U, Ind∗ ) is a representation of any Pawlak-rough set in U . Proposition 1. Let (U, Ind∗ ) be given. Then the relation R deﬁned by xRy if and only if Ind∗ (x, y) ≥ 1 is an equivalence relation, and the pair (I, B) deﬁned by I = {x : Ind∗ (x, x) = 3}, B = {x : Ind∗ (x, x) = 2} constitute the interior and boundary of a rough set in (U, R). Proposition 2. Let (U, R) be an approximation space in which (I, B), the interior and boundary pair determines a rough set. Consider the mapping Ind∗ : U × U → L(4) given by ⎧ ⎨ 3 if x ∈ I Ind∗ (x, x) = 2 if x ∈ B ⎩ 1 if x ∈ U \ (I ∪ B) and ∗

Ind (x, y) =

1 if x = y, xRy holds 0 if x = y, xRy does not hold.

Then Ind∗ satisﬁes the conditions H1 , H2 , R1 , R2 , R3 , R4 . The following important feature is also observed in the two constructions by Propositions 1 and 2.

58

M.K. Chakraborty P rop1

P rop2

P rop1

(U, Ind∗ ) −→ (U, R, I, B) −→ (U, Ind∗ ) −→ (U, R, I, B). Thus an indiscernibility satisfying the roughness conditions gives rise to a rough set in an approximation space and vice versa. One can see that the above representation does not amount to the 3-valued semantics where an object falls under a concept with three grades viz. 3 (when it is in the deﬁnite region), 2 (when it is in the boundary), and 1 (outside the boundary). We have noticed diﬃculties with such representation (by Pawlak or Wygralak) in the beginning of this article. The membership function of Obtulowicz also says that Ind∗ (x, x) & Ind∗ (x, y) ≤ Ind∗ (y, y), i.e. the degree of belongingness of x to a concept & the degree of indiscernibility of x with y ≤ the degree of belongingness of y to the concept. This criterion is the so-called ‘saturatedness’ condition that has been elaborately discussed in [5]. In Leibniz’s terms, this is the version of the doctrine of Identity of Indiscernibles, viz. if an object x has a property P and an object y is indiscernible from x, then y has the property P . In the fuzzy context, a more general condition is taken viz. α(x) & Ind(x, y) ≤ α(y), where α(x) denotes the degree of belongingness of x to the fuzzy set α. The only addition here is the conceptual indulgence to the assumption that belongingness degree of x to a fuzzy set is the same as Ind∗ (x, x). It is also signiﬁcant to notice that the fuzzy set theoretic operators ‘max’ and ‘min’ are now applicable to obtain the union and intersection. A summary of what has been said so far is the following: – The underlying indiscernibility relation for any vague concept in U is a relation Ind∗ satisfying the conditions H1 , H2 , R1 , R2 , R3 , R4 . – Such a relation, which is a particular kind of fuzzy equivalence relation determines uniquely a rough set in the approximation space (U, R) where R is virtually the underlying indiscernibility and conditions R1 , R2 , R3 , R4 determine the interior and boundary of the rough set. – Conversely, any rough set in (U, R) given by the interior and boundary can be generated by an indiscernibility relations satisfying H1 , H2 , R1 , R2 , R3 , R4 . – This representation is one-to-one. – In the special case when the boundary region is empty, the condition R4 is dropped. We need a three-element complete Heyting algebra and Propositions 1 and 2 may be written accordingly. Earlier examples with the concept A, and concepts A, B, C are instances of such representation. One interesting extension of Obtulowicz’s representation of rough sets suggests itself, but has never been taken up. Formally, one can add more categories other than 0, 1, 2 and 3. For instance, let us take one more viz. 4. All the roughness conditions remain the same except that the condition R4 shall now be read as R4 : for each x, Ind(x, x) = 2 or 3 implies that there exists y such that Ind(x, y) = 1.

Pawlak’s Landscaping with Rough Sets

59

The rough set is now extended to (U, R, I, B1 , B2 ), where there are two layers in the boundary, elements of I have grade 4, while those of B1 and B2 have grades 3 and 2 respectively. The representation theorems now take the following shapes. Proposition 3. Let Ind∗ : U × U → L(5)(≡ {0 ≤ 1 ≤ 2 ≤ 3 ≤ 4} be an indiscernibility relation satisfying conditions H1 , H2 , R1 , R2 , R3 , R4 . Then R deﬁned by xRy if and only if Ind∗ (x, y) ≥ 1 is an equivalence relation, and the triple (I, B1 , B2 ) deﬁned by I = {x : Ind∗ (x, x) = 4}, B1 = {x : Ind∗ (x, x) = 3}, B2 = {x : Ind∗ (x, x) = 2} constitute the interior and the two layers of boundaries of a rough set in (U, R). Proposition 4. Let (U, R) be an approximation space in which (I, B1 , B2 ) determines a rough set. Then the mapping Ind∗ : U × U → L(5) given by ⎧ 4 if x ∈ I ⎪ ⎪ ⎨ 3 if x ∈ B1 ∗ Ind (x, x) = 2 if x ∈ B2 ⎪ ⎪ ⎩ 1 if x ∈ U \ (I ∪ B1 ∪ B2 ) and ∗

Ind (x, y) =

1 if x = y, xRy holds 0 if x = y, xRy does not hold.

Then Ind∗ satisﬁes the conditions H1 , H2 , R1 , R2 , R3 , R4 . Layers of the boundary may be enhanced arbitrarily, but ﬁnitely. The interpretation of these layers is zones of objects of gradually weaker possibilities of falling under the concept that render gradualness along with granularity. One can also see the possibility of interpreting membership values under rough membership function [22] as values of the lattice (after normalization, of course). Elements of a block with lower rough membership may be placed in the weaker layer of the boundary. Element of beyond-possible zone should be given the value 1 instead of 0 which should be retained to denote the discernibility of x, y, x = y and to determine the partition of the Universe. In ﬁnite situations there should not be any diﬃculty. The deﬁnitions of union intersection and complementation may be suitably deﬁned by using max, min and reversal of grades. There shall be a departure in this approach from that of Pawlak. ‘Roughness’ should no longer be considered as an adjective to an ordinary set in an approximation space but as a pair (I, B) of unions of blocks of the space, I being called the interior and B the boundary. This approach is equivalent to that taken by us in [2]. It is also in conformity with the philosophy of rough sets viz. “we ‘see’ elements of the universe through available informations” and hence “some elements may be ‘seen’ identical” and “this is to mean that if we see a set through informations, only, the approximations (lower and upper) can be observed.” And further “a vague property determines not only a single set of elements falling under the property but a family of sets which can be identiﬁed with this property up to indiscernibility.” All the above lines within quotation marks are from Pawlak’s writings glued together. So the present approach, though a departure, draws the support from his own feelings too.

60

3

M.K. Chakraborty

Conclusions

Like fuzzy sets, rough sets also have wide usages for example, in artiﬁcial intelligence, cognitive sciences, knowledge discovery from database, machine learning, expert system, inductive reasoning and pattern recognition. As in the case of any successful theory, it has to look back at one point of time. “The rough set theory has reached such a state that some kind of summary of its theoretical foundation is a must” [28] - this realization of Pawlak is quite justiﬁable. As it appears from the preceding discussions, the approach of Obtulowicz may serve as quite a reasonable foundation. This approach also suggests a kind of uniﬁcation of fuzzy sets and rough sets. Starting with some sort of indiscernibility in the Universe which is at the base of any vague concept and which arises out of data (concrete or subjective), objects of the Universe are categorized. In such a categorization, some tokens (not necessarily numbers) with varying algebraic structures play a role. A mathematical entity emerges representing the vague concept. In this representation, the following philosophy is adopted: “The degree of existence of an object in a concept is the degree to which the object is indiscernible with itself relative to the underlying the concept.” The mathematical entity is sometimes a fuzzy set and sometimes a rough set. Divergence occurs because of the nature of the indiscernibility (crisp or fuzzy) and the choice of categories (how many?) and their structures. Incorporation of layers in the boundary as proposed in the paper for the ﬁrst time brings rough sets closer to fuzzy sets. We think that this approach could help in erasing Pawlak’s persistent feeling of a sort of ‘supremacy’ of classical set theory over fuzzy set theory or rough set theory. This feeling is expressed in statements like “fuzzy set involves more advanced mathematical concepts real numbers and functions - whereas in classical set theory the notion of set is used as a fundamental notion of whole mathematics and is used to derive any other mathematical concept e.g. numbers and functions. Consequently, fuzzy set theory cannot replace classical set theory, because, in fact, the theory is needed to deﬁne fuzzy sets.” [21] Again, “in a manner similar to fuzzy set theory, rough set theory is not an alternative to classical set theory but it is embedded in it.”[22] Pawlak’s concern about the foundational problems of classical set theory and interest in the alternatives like, multisets, multi-fuzzy sets, Blizard sets, general sets, Mereology (Lesniewski), Alternative set theory (Vopenka), Penumbral set theory (Apostoli and Kanda) are well known [21,25]. In the proposal of Obtulowicz, what would, in fact, be needed at the beginning is a Universe, a collection of tokens with some structures including order (in particular the numbers) and the notion of indiscernibility which needs only an understanding of ‘pair’ and ‘correspondence’. These may constitute a nice, intuitively acceptable beginning. If one casts an oblique eye, we can take refuge to categorical foundation (Higgs [12], Banerjee and Chakraborty [3], and others). At this point we would like to draw the attention of readers to a paper of Goguen published in 1974. He claims “Ideally, we would like a foundation for fuzzy sets which justiﬁes the intuitive identiﬁcation of fuzzy sets with (inexact) concepts, and in which the familiar set operations are uniquely and inevitably determined. These

Pawlak’s Landscaping with Rough Sets

61

desires are clearly though not explicitly expressed in Zadeh (1965), and they are all satisﬁed by the system given in this paper.” He used as paradigm Lawvere’s “relatively unknown axiomatization of sets in the language of category theory”. In one of our papers [3] Higgs’ and Goguen’s categories are discussed and compared. But category-theoretic approach is still not popular among practitioners. If toiled and fuzzy as well as rough sets are found soundly based on categorical grounds, the gains appear to be enormous - reinstatement of ‘vagueness’ within the discourses of mathematics, logic and thus by sciences. A recent work by Skowron [39] once again brings to focus the importance of the study of vagueness and the role of rough set theoretic methods in such studies. He introduces two operations viz. inductive extension and granulation of approximation spaces and emphasizes on “important consequences of the paper for research on approximation of vague concepts and reasoning about them in the framework of adoptive learning.” He thinks that “this (adoptive learning) requires developing a new approach to vague concepts going beyond the traditional rough or fuzzy approaches.” This paper extends the notion of approximation space by incorporating rough inclusion or graded inclusion. It would be an interesting project to investigate if Obtulowicz’s proposal may be extended to this generalized context also. We engage not only into crisp (two-valued) talks about vagueness, but into multi-valued talks too (theory of graded consequence [6,7]) or we also talk ‘roughly’ (theory of rough consequence [5]). The underlying motivation to deﬁne graded consequence or rough consequence is to allow room for vagueness in the metalogical concepts like consistency, consequence, tautologihood, completeness, etc. This latter notion viz. rough consequence has its origin in Pawlak’s insightful work on rough truth [24] where he states “the rough (approximate) truth and falsity represent our partial knowledge about the world and with the increase of our knowledge the roughly true (or false) formulas tend to be more true (or false) and approach the truth and falsity closer and closer.” One is bound to recall Zadeh when he claims that the notion of truth itself is fuzzy. It would not be out of place to mention that the ﬁrst recorded works on rough logics are by Orlowska and Pawlak [19] and Rasiowa and Skowron [36]. With the advent of graded and rough consequences, the scenario of mathematics should change, in that there may be some mathematical predicates, truth of sentences relative to which may be partial and derivations involving which may not be of full strength. Professor Pawlak was an artist. His favourite example of vagueness was a ‘beautiful’ painting [22]. He was fond of landscaping. His work on rough sets is also a beautiful landscape-installation - to which we oﬀer this humble bouquet of ours that might develop roots striving to become an integral part of this scenario.

References 1. M. Banerjee. A Categorial Approach to the Algebra and Logic of the Indiscernible. Ph.D Thesis, University of Calcutta, 1993. 2. M. Banerjee and M.K. Chakraborty. Rough sets through algebraic logic. Fundamenta Informaticae, 28(3,4): 211–221, 1996.

62

M.K. Chakraborty

3. M. Banerjee and M.K. Chakraborty. Foundations of vagueness: a category-theoretic approach. Electronic Notes in Theoretical Computer Science, 82(4), 2003. 4. M.K. Chakraborty and M. Banerjee. Rough consequence. Bulletin of the Polish Academy of Sciences (Mathematics), 41(4):299–304, 1993. 5. M.K. Chakraborty and E. Orlowska. Substitutivity principles in some theories of uncertainty. Fundamenta Informaticae, 32:107–120, 1997. 6. M.K. Chakraborty and S. Basu. Graded consequence and some metalogical notions generalized. Fundamenta Informaticae, 32:299–311, 1997. 7. M.K. Chakraborty and S. Basu. Approximate reasoning methods in vagueness: graded and rough consequences. ICS Research Report, 29, Warsaw University of Technology, 1995. 8. S. Demri and E. Orlowska. (eds.) Incomplete Information: Structure, Inference, Complexity. Monographs in Theoretical Computer Science, Springer-Verlag, Heidelberg, 2002. 9. D. Dubois and H. Prade. Rough fuzzy sets and fuzzy rough sets. In Proc. International Conference on Fuzzy Sets in Informatics, Moscow. 1988, 20–23. 10. M. Eytan. Fuzzy sets: a topos-logical point of view. Fuzzy Sets and Systems, 5:47–67, 1981. 11. J. Goguen. Concept representation in natural and artiﬁcial languages: axioms, extensions and applications for fuzzy sets. International Journal for Man-Machine Studies, 6:513–561, 1975. 12. D. Higgs. A categorical approach to Boolean-valued set theory. Preprint, 1973. 13. D. Hyde. From heaps and gaps to heaps of gluts. Mind, 106:440–460, 1997. 14. J. Jacas. On the generators of T-indistinguishability operator. Stochastica, XIII: 49–63, 1988. 15. R. Keefe. Theories of Vagueness, Cambridge Studies in Philosophy, Cambridge, UK, 2000. 16. R. Keefe and P. Smith. (eds.) Vagueness: A Reader, MIT Press, Massachusetts, MA, 1997. 17. F. Keikeben. http://members.aol.com/kiekeben/theseus.html, 2000. 18. A. Obtulowicz. Rough sets and Heyting algebra valued sets. Bulletin of the Polish Academy of Sciences (Mathematics), 13(9-10):667–671, 1987. 19. E. Orlowska and Z. Pawlak. Representation of non-deterministic information. Theoretical Computer Science, 29:27–39, 1984. 20. Z. Pawlak. Rough sets. International Journal of Computer and Information Sciences, 11:341–356, 1982. 21. Z. Pawlak. Some issues on rough sets. Transactions of Rough sets I, 1–58, 1998. 22. Z. Pawlak. A treatise on rough sets. Transactions on Rough sets IV, 1–17, 2005. 23. Z. Pawlak. Rough sets and fuzzy sets. Fuzzy Sets and Systems 17:99–102, 1985. 24. Z. Pawlak. Rough logic. Bulletin of the Polish Academy of Sciences(Technical Sciences), 35(5-6):253–258, 1987. 25. Z. Pawlak. Hard and soft sets. ICS Research Report, 10/94, Warsaw University of Technology, 1994. 26. Z. Pawlak. Vagueness – a rough set view. LNCS 1261, Springer, 1997, 106–117. 27. Z. Pawlak. Vagueness and uncertainty: a rough set perspective. Computational Intelligence: An International Journal, 11:217–232, 1995. 28. Z. Pawlak. Rough sets, present state and further prospects. ICS Research Report, 15/19, Warsaw University of Technology, 1995. 29. Z. Pawlak and A. Skowron. Rudiments of rough sets. Information Sciences, to appear.

Pawlak’s Landscaping with Rough Sets

63

30. Z. Pawlak and A. Skowron. Rough sets: some extensions. Information Sciences, to appear. 31. Z. Pawlak and A. Skowron. Rough sets and Boolean reasoning, Information Sciences, to appear. 32. G. Priest and R. Routley. First historical introduction: a preliminary history of paraconsistent and dialethic approaches. In Priest, Routley and Normann, editors, Paraconsistent Logic, Essays on the Inconsistent. Philosophia Verlag, MunchenHamden-Wien, 1989, 1–75. 33. L. Polkowski. Rough Sets: Mathematical Foundations. Advances in Soft Computing, Physica Verlag, Hiedelberg, 2002. 34. A. Pultr, Fuzziness and fuzzy equality. In H.J. Skala, S. Termini and E. Trillas, editors, Aspects of Vagueness. D. Reidel Publishing Co., Dordrecht, Holland, 1984, 119–135. 35. P.T. Raju. The principle of four-coloured negation in Indian philosophy. Review of Metaphysics 7:694-713, 1953. 36. H. Rasiowa and A. Skowron. Rough concepts logic in computation theory. In A. Skowron, editor, LNCS 208, Springer, 1985, 288–297. 37. B. Russell. Vagueness. Australian Journal of Philosophy, 1:84–92, 1923. 38. A. Skowron. The relationship between the rough set theory and evidence theory. Bulletin of the Polish Academy of Sciences (Technical Sciences), 37(1-2):87–90, 1989. 39. A. Skowron. Rough sets and vague concepts. Fundamenta Informaticae, 64: 417–431, 2005. 40. A. Skowron and J.W. Grzymala-Busse. From rough set theory to evidence theory. In Yager, Fedrizzi and Kacprzyk, editors, Advances in the Dempster-Shafer Theory of Evidence. John Wiley & Sons, New York, 1994, 193–236. 41. E. Trillas and L. Valverde. An inquiry into indistinguishability operators. In H.J. Skala, S. Termini and E. Trillas, editors, Aspects of Vagueness. D. Reidel Publishing Co., Dordrecht, Holland, 1984, 231–256. 42. H. Wang. Beyond Analytic Philosophy. MIT Press, Cambridge, 1986. 43. D. Wiggins. Sameness and Substance. Oxford Blackwell, 1980, 92–94. 44. M. Wygralak. Some remarks on rough and fuzzy sets. BUSEFAL 21, 1985, 43–49. 45. L.A. Zadeh. Fuzzy Sets. Information and Control, 8:338–353, 1965.

A Comparison of Pawlak’s and Skowron–Stepaniuk’s Approximation of Concepts Anna Gomoli´ nska University of Bialystok, Department of Mathematics, Akademicka 2, 15267 Bialystok, Poland [email protected]

Abstract. In this article, we compare mappings of Pawlak’s lower and upper approximations of concepts with those proposed by Skowron and Stepaniuk. It is known that both approaches coincide for the standard rough inclusion, so we consider the case of an arbitrary rough inclusion function. Even if the approximation space investigated is based on an arbitrary non-empty binary relation, the lower approximation mappings are equal in both approaches. Nevertheless, the upper approximation mappings are diﬀerent in general.

In view of many generalizations and extensions of rough set theory some kind of uniﬁcation of the basic theory seems to be badly needed. (Z. Pawlak [1], p. 10)

1

Introduction

Nowadays, the Pawlak rough approximation of concepts [2,3,4,5] has become a classical research topic. Lower and upper rough approximations have been investigated by many researchers in the rough set community, to mention [6,7,8] by way of example. Skowron and Stepaniuk’s proposal regarding rough approximation of concepts [9,10] is well-known, yet less popular among researchers as a subject of study. A possible reason may be the fact that both approaches coincide for the standard rough inclusion, whereas this very function is the most known among rough inclusions. The aim of this paper is to study and to compare both Pawlak’s and Skowron– Stepaniuk’s approaches to approximation of concepts in the rough-set framework. To this end, we relax the usual assumptions about the approximation space considered. We start with an approximation space, understood as a structure M = (U, , κ), where U (the universe) is a non-empty set of objects, is a non-empty binary relation on U , and κ is a mapping on the set of pairs of sets of objects called a rough inclusion function. Step by step, we consider spaces

The research was supported by the grant 3T11C00226 from the Ministry of Science of the Republic of Poland.

J.F. Peters et al. (Eds.): Transactions on Rough Sets VI, LNCS 4374, pp. 64–82, 2007. c Springer-Verlag Berlin Heidelberg 2007

A Comparison of Pawlak’s and Skowron–Stepaniuk’s Approximation

65

based on serial relations, reﬂexive relations, transitive relations, symmetric (and in particular, tolerance) relations, and equivalence relations. By deﬁnition, κ is assumed to satisfy two postulates only which are in accordance with the axioms of rough mereology. However, we also investigate cases, where κ fulﬁlls additional conditions. In the paper, we examine two pairs of approximation mappings in line with Pawlak’s approach and two pairs of approximation mappings in line with Skowron–Stepaniuk’s proposal. Each pair consists of a lower approximation mapping and an upper approximation mapping. Both in the Pawlak case as well as in the Skowron–Stepaniuk case, approximation mappings constituting one pair are viewed as basic, whereas mappings forming the remaining pair are “deﬁnable” versions of the basic mappings. As regarding the results, the basic lower approximation mappings are equal in both approaches, and similarly for their “deﬁnable” variants. Basic upper approximation mappings (and similarly for their “deﬁnable” counterparts) are diﬀerent in general and may be compared only under special conditions put on κ. Therefore, we mainly try to compare these mappings indirectly via their properties. Apart from that, we aim at a uniform presentation of facts about lower and upper approximation mappings. Some of the facts are new, others are only recalled. We try to answer such questions as: What are the results of application of a given approximation mapping to the empty set and to the whole universe? What are the relationships among various forms of approximation? For instance, how is the lower approximation related to the upper one? Moreover, how are approximations of a concept related to the concept itself? Are the approximation mappings under investigation monotone, and if it is the case, what are the consequences? Last but not least, what a mapping may be obtained by various compositions of the approximation mappings? As we shall see, the mappings investigated can lack some essential properties attributed to an approximation mapping if the relation , underlying a given approximation space M , is not reﬂexive. For example, a lower approximation of a concept may not be included in that concept. However, slightly abusing the terminology, we shall use the names ‘lower approximation’ and ‘upper approximation’ for the sake of uniformity. Basic terminology and notation is introduced in Sect. 2. In Sect. 3, we present the notion of a rough approximation space and the mappings of Pawlak’s as well as Skowron–Stepaniuk’s lower and upper rough approximation of concepts. In Sect. 4, properties of these mappings are studied in the case of an approximation space based on an arbitrary non-empty binary relation . In Sect. 5, we examine the approximation mappings for special cases of approximation spaces, where and/or its converse relation −1 are serial and where is, in turn, a reﬂexive relation, a transitive relation, a symmetric relation, and an equivalence relation. The results are summarized brieﬂy in the last section.

2

Preliminaries

Let X, Y be any sets. Throughout the paper, the power set of X, the cardinality of X, the Cartesian product X × X, the identity mapping on X, and the set

66

A. Gomoli´ nska

of all mappings f : X → Y will be denoted by ℘X, #X, X 2 , idX , and Y X , respectively. Consider (partially) ordered sets (X, ≤) and (Y, ). A mapping f : X → Y is referred to as monotone, written f ∈ MON, if for any x, y ∈ X, x ≤ y implies f x f y. The operation of composition of relations will be denoted by ◦. In the case of mappings, the composition of f : X → Y with g : Y → Z is a mapping g ◦ f : X → Z such that for any x ∈ X, (g ◦ f )x = g(f x). For any sets X, Y , deﬁne a relation on (℘Y )℘X and operations , on ((℘Y )℘X )2 such that for any mappings f, g : ℘X → ℘Y and any Z ⊆ X, def

f g ⇔ ∀Z ⊆ X.f Z ⊆ gZ, def

def

(f g)Z = f Z ∩ gZ & (f g)Z = f Z ∪ gZ.

(1)

By assumption, ◦ will take the precedence of the operations just deﬁned, whereas the logical connectives of conjunction and disjunction will take the precedence of implication and double implication. Proposition 1. For any mappings f, g, h : ℘X → ℘Y , we have: (a) f f (b) f g & g f ⇒ f = g (c) f g & g h ⇒ f h (d) f g ⇒ f ◦ h g ◦ h (e) h ∈ MON & f g ⇒ h ◦ f h ◦ g (f ) f g h ⇔ f h & g h (g) f g h ⇔ f g & f h The proof is easy and, hence, omitted. Let us only note that is a partial ordering on (℘Y )℘X in virtue of (a)–(c).

3

Rough Approximation Spaces

The notion of a rough approximation space was obtained by Prof. Pawlak in the early 80’s of the 20th century as one of the results of investigations on approximation of vague concepts in information systems [2,3,4,5]. This basic notion was next reﬁned and generalized in several directions (see, e.g., [11,12,13,14,15]), yet we shall only focus upon the extension proposed by Skowron and Stepaniuk in [9,10], and elaborated in a series of research articles [16,17,18]. In [19,20]1 , Polkowski and Skowron introduced and characterized axiomatically the formal notion of a rough inclusion. Although this notion is unnecessary when discussing Pawlak’s classical approach, it is fundamental for Skowron– Stepaniuk’s one. Consider a non-empty set U of entities called objects. In com1

See also more recent papers.

A Comparison of Pawlak’s and Skowron–Stepaniuk’s Approximation

67

pliance with rough mereology, by a rough inclusion function (RIF for short) upon U we understand any mapping κ : (℘U )2 → [0, 1] satisfying rif 1 and rif 2 below: def

rif 1 (κ) ⇔ ∀X, Y.(κ(X, Y ) = 1 ⇔ X ⊆ Y ) def

rif 2 (κ) ⇔ ∀X, Y, Z.(Y ⊆ Z ⇒ κ(X, Y ) ≤ κ(X, Z)) RIFs are intended as mappings measuring the degrees of inclusion of sets of objects in sets of objects. Apart from the above postulates, one may consider other conditions, for instance, def

rif 3 (κ) ⇔ ∀X = ∅.κ(X, ∅) = 0, def

rif 4 (κ) ⇔ ∀X = ∅.∀Y.(κ(X, Y ) = 0 ⇒ X ∩ Y = ∅), def

rif 4∗ (κ) ⇔ ∀X = ∅.∀Y.(X ∩ Y = ∅ ⇒ κ(X, Y ) = 0), def

rif 5 (κ) ⇔ ∀X = ∅.∀Y.(κ(X, Y ) = 0 ⇔ X ∩ Y = ∅). One can easily see that rif 4∗ (κ) implies rif 3 (κ), whereas rif 5 (κ) if and only if rif 4 (κ) and rif 4∗ (κ). The most famous RIF is the standard one, deﬁned for the ukasiewicz [21] and ﬁnite universe and denoted by κ£ here, which goes back to L is based on the frequency count. κ£ is given by #(X∩Y ) if X = ∅ £ #X κ (X, Y ) = (2) 1 otherwise and fulﬁlls not only rif 1 , rif 2 but also rif 5 and some other conditions. By a rough approximation space we mean a triple M = (U, , κ), where U — the universe of M — is a non-empty set of objects as earlier, is a non-empty binary relation on U , and κ is a RIF upon U . Objects will be denoted by u with sub/superscripts if needed. Sets of objects of U are viewed as concepts of M . With every object u, there are associated two basic concepts: the image and the co-image of {u}, → {u} and ← {u}, deﬁned along the standard lines and called elementary granules of information 2 drawn to u. It is worth recalling that → {u} = −1← {u} and ← {u} = −1→ {u}. Let us note that induces mappings Γ , Γ∗ : U → ℘U , called uncertainty mappings in line with Skowron–Stepaniuk’s approach, such that for every object u ∈ U, (3) Γ u = ← {u} & Γ∗ u = → {u}. Thus, elementary granules of information are simply values of Γ, Γ ∗ . Clearly, Γ∗ = Γ−1 , and u ∈ Γ∗ u if and only if (u, u ) ∈ , i.e., if and only if u ∈ Γ u . Moreover, Γ∗ = Γ if is symmetric. On the other hand, every mapping Γ : U → ℘U induces a relation Γ on U such that for any objects u, u ∈ U , (u, u ) ∈ Γ ⇔ u ∈ Γ u . 2

(4)

The term ‘information granule’ was proposed by Zadeh [22] to denote a clump of objects drawn together on the basis of indiscernibility, similarity or functionality.

68

A. Gomoli´ nska

Since ΓΓ = Γ and Γ = , structures (U, , κ) and (U, Γ , κ) are interdeﬁnable, and similarly for (U, Γ, κ) and (U, Γ , κ). In the classical Pawlak approach, a rough approximation space is a pair (U, ), where U is a ﬁnite non-empty set and is an equivalence relation understood as a relation of indiscernibility of objects. Then, elementary granules of information are equivalence classes, i.e. sets of objects indiscernible from one another. Clearly, a natural augmentation of (U, ) with a RIF κ, results in an approximation space (U, , κ) in line with our approach. Keeping with the recent state-of-art, one can say that Skowron–Stepaniuk’s approximation spaces, introduced in [9,10], are of the form N = (U, Γ$ , κ$ ), where Γ$ is an uncertainty mapping such that for every object u, u ∈ Γ$ u, κ$ is a RIF, and $ is a list of tuning parameters to obtain a satisfactory quality of approximation of concepts. For the latter, such spaces are called parameterized approximation spaces as well. Henceforth, the parameters $ will be dropped for simplicity. One can easily see that N is based on a reﬂexive relation3 , e.g. Γ$ . Due to our earlier observations on interdeﬁnability of approximation spaces and the corresponding structures based on uncertainty mappings, and slightly abusing the original terminology, we shall think of Skowron–Stepaniuk’s approximation spaces as structures of the form (U, , κ), where is a reﬂexive relation on U . In the sequel, a concept X is referred to as -deﬁnable (resp., −1 -deﬁnable) if it is a set-theoretical union of elementary granules of the form Γ u (resp., Γ∗ u). Henceforth, references to will be omitted whenever possible. For instance, we shall write Γ and Γ ∗ instead of Γ and Γ∗ , respectively. Where is symmetric, both forms of deﬁnability coincide, so we may simply speak of deﬁnable or undeﬁnable concepts. The main idea underlying the Pawlak rough approximation of concepts is that even if a concept is not deﬁnable in a given space, it can be approximated from the inside and the outside by deﬁnable concepts. In this way, the Pawlak lower and upper rough approximation mappings, low∪ , upp∪ ∈ (℘U )℘U , respectively, are obtained such that for any concept X, def def low∪ X = {Γ u | Γ u ⊆ X} & upp∪ X = {Γ u | Γ u ∩ X = ∅}. (5) The lower approximation of X, low∪ X, is the largest -deﬁnable concept included in X, whereas the upper approximation of X, upp∪ X, is the least deﬁnable concept containing X provided that is serial. The diﬀerence bnd∪ X = upp∪ X − low∪ X def

(6)

is called the boundary region of X. When this region is empty, X is referred to as exact ; otherwise it is rough. In Pawlak’s approximation spaces, it turns out that a concept is exact if and only if it is deﬁnable. Apart from low∪ , upp∪ , we shall also refer to the mappings low, upp ∈ (℘U )℘U given below as the Pawlak lower and upper rough approximation mappings, respectively: 3

Primarily, parameterized approximation spaces were based on reﬂexive and symmetric (i.e., tolerance) relations.

A Comparison of Pawlak’s and Skowron–Stepaniuk’s Approximation def

def

lowX = {u | Γ u ⊆ X} & uppX = {u | Γ u ∩ X = ∅} ∪

69

(7)

∪

As a matter of fact, low = low and upp = upp for the Pawlak approximation spaces. In a general case, however, low, upp diﬀer from their -deﬁnable versions low∪ , upp∪ , respectively. The lower approximation of X, lowX, may be viewed as the set consisting of all objects u which surely belong to X since their elementary granules Γ u are included in X. On the other hand, the upper approximation of X, uppX, may be perceived as the set consisting of all objects u which possibly belong to X since their elementary granules Γ u overlap with X. The Skowron–Stepaniuk lower and upper rough approximation mappings, lowS, uppS ∈ (℘U )℘U , respectively, are deﬁned by the following conditions, for any concept X, def

def

lowS X = {u | κ(Γ u, X) = 1} & uppS X = {u | κ(Γ u, X) > 0}.

(8)

That is, the lower approximation of X, lowS X, consists of all objects u that their elementary granules Γ u are included in X to the highest degree 1. On the other hand, the upper approximation of X, uppS X, is the set of all objects u that their elementary granules Γ u are included in X to some positive degree. The boundary region of X is deﬁned as the set def

bndS X = {u | 0 < κ(Γ u, X) < 1}.

(9)

Mappings lowS∪ , uppS∪ , being -deﬁnable versions of lowS , uppS , are also referred to as the Skowron–Stepaniuk lower and upper rough approximation mappings, respectively. They are given by the following equalities: def lowS∪ X = {Γ u | κ(Γ u, X) = 1} def {Γ u | κ(Γ u, X) > 0} (10) uppS∪ X = Obviously, we can repeat the construction of approximation mappings for −1 what can be useful if is not symmetric. As a result, mappings low∗ , upp∗ , lowS∗ , and uppS∗ may be derived (as well as their −1 -deﬁnable versions which will not be presented here), where for any concept X, low∗ X = {u | Γ ∗ u ⊆ X} & upp∗ X = {u | Γ ∗ u ∩ X = ∅}, def

def

lowS∗ X = {u | κ(Γ ∗ u, X) = 1} & uppS∗ X = {u | κ(Γ ∗ u, X) > 0}. (11) def

def

The mapping upp∗ isparticularly important for our purposes. It turns out that for any concept X, {Γ u | u ∈ X} = upp∗ X. Mappings low, upp, lowS , and uppS will be viewed as basic. Thus, for every basic mapping f , f ∪ = upp∗ ◦ f.

(12)

Example 1. For the sake of illustration of the approximation mappings, consider a set U = {3, . . . , 10} consisting of 8 objects denoted by 3, . . . , 10, and a binary

70

A. Gomoli´ nska

relation on U generating the uncertainty mappings4 Γ, Γ ∗ shown in Table 1. is reﬂexive, so it is a similarity relation (but even not a tolerance relation). Let κ be any RIF as earlier and κ1 , κ2 be such RIFs that for any concepts X, Y where X = ∅, κ1 (X, Y ) = 0 ⇔ X ∩ upp∗ Y = ∅, κ2 (X, Y ) = 0 ⇔ uppX ∩ upp∗ Y = ∅.

(13)

Both κ1 and κ2 satisfy rif 4 since κ2 (X, Y ) = 0 implies κ1 (X, Y ) = 0, and the latter implies X ∩ Y = ∅. Indeed, X ⊆ uppX and Y ⊆ upp∗ Y due to reﬂexivity of . It is easy to see that κ1 (X, Y ) > 0 if and only if there exist u ∈ X and u ∈ Y such that (u, u ) ∈ . Furthermore, κ2 (X, Y ) > 0 if and only if there exist u ∈ X and u ∈ Y such that (u, u ) ∈ ◦ . Let uppS1 and uppS2 denote the Skowron–Stepaniuk upper approximation mappings based on κ1 and κ2 , respectively, i.e., for any concept X and i = 1, 2, uppSi X = {u | κi (Γ u, X) > 0}.

(14)

That is, u ∈ uppS1 X if and only if Γ u ∩ upp∗ X = ∅, and u ∈ uppS2 X if and only if (upp ◦ Γ )u ∩ upp∗ X = ∅. Values of upp ◦ Γ are given in Table 1. One can show that (15) upp uppS1 uppS2 , yet the converse inclusions may not hold in general. To see this, consider X = {3, 4}. Note that lowX = {4}, low∗ X = ∅, uppX = {3, 4, 5}, upp∗ X = {3, 4, 6}, uppS1 X = {3, 4, 5, 6}, and uppS2 X = {3, 4, 5, 6, 10}. Table 1. Values of Γ , Γ ∗ , and upp ◦ Γ u 3 4 5 6 7 8 9 10

4

Γu {3,4,6} {3,4} {3,4,5,6} {6,10} {7,8,9} {7,8} {8,9} {9,10}

Γ ∗u {3,4,5} {3,4,5} {5} {3,5,6} {7,8} {7,8,9} {7,9,10} {6,10}

upp(Γ u) {3,4,5,6} {3,4,5} {3,4,5,6} {3,5,6,10} {7,8,9,10} {7,8,9} {7,8,9,10} {6,7,9,10}

Properties of Approximation Mappings

We ﬁrst investigate properties of the basic mappings low, upp, lowS , and uppS , where is an arbitrary non-empty relation on U . Henceforth, f will denote low, upp or uppS . 4

We drop references to for simplicity.

A Comparison of Pawlak’s and Skowron–Stepaniuk’s Approximation

71

Proposition 2. Let X, Y be any concepts. Then, we have: (a) low = lowS uppS (b) rif 4 (κ) ⇒ upp uppS (c) upp∅ = ∅ & lowU = uppS U = U (d) {u | Γ u = ∅} ⊆ lowX = U − upp(U − X) (e) X ⊆ Y ⇒ f X ⊆ f Y (f ) f (X ∩ Y ) ⊆ f X ∩ f Y ⊆ f X ∪ f Y ⊆ f (X ∪ Y ) Proof. We prove (b), (d), and (e) only. To this end, consider any concepts X, Y and any object u. For (b) assume that rif 4 (κ) holds. Consider u ∈ uppX. By the deﬁnition of upp, (b1) Γ u ∩ X = ∅. Hence, Γ u = ∅. As a consequence, κ(Γ u, X) > 0 by the assumption and (b1). By the deﬁnition of uppS , u ∈ uppS X. Thus, uppX ⊆ uppS X. Immediately, upp uppS by the deﬁnition of . In case (d), ﬁrst suppose that Γ u = ∅. Hence, κ(Γ u, X) = 1 in virtue of rif 1 (κ), i.e., u ∈ lowX by the deﬁnition of low. In the sequel, u ∈ lowX if and only if (by the deﬁnition of low) Γ u ⊆ X if and only if Γ u ∩ (U − X) = ∅ if and only if (by the deﬁnition of upp) u ∈ upp(U − X) if and only if u ∈ U − upp(U − X). For (e) assume (e1) X ⊆ Y . First, let f = low and suppose that u ∈ lowX. By the deﬁnition of low, Γ u ⊆ X. Hence, Γ u ⊆ Y by (e1). Again by the deﬁnition of low, u ∈ lowY . Thus, lowX ⊆ lowY . Next, let f = upp and u ∈ uppX. By the deﬁnition of upp, Γ u ∩ X = ∅. Hence, Γ u ∩ Y = ∅ by (e1). By the deﬁnition of upp, u ∈ uppY . Thus, uppX ⊆ uppY . Finally, where f = uppS , assume that u ∈ uppS X. By the deﬁnition of uppS , (e2) κ(Γ u, X) > 0. In virtue of (e1) and rif 2 (κ), κ(Γ u, X) ≤ κ(Γ u, Y ). Hence, κ(Γ u, Y ) > 0 by (e2). By the deﬁnition of uppS , u ∈ uppS Y . Thus, uppS X ⊆ uppS Y .

Let us comment upon the properties. In virtue of property (a), Pawlak’s and Skowron–Stepaniuk’s approaches coincide as regarding the lower approximation. In the latter approach, the lower approximation of a concept is always included in the upper one. The Pawlak upper approximation is, in general, incomparable with the Skowron–Stepaniuk upper approximation unless some additional assumptions like rif 4 (κ) are made. Due to (b), the Pawlak upper approximation of a concept is included in the Skowron–Stepaniuk one if rif 4 (κ) is assumed. As we shall see later on, rif 4∗ (κ) will guarantee that the Skowron–Stepaniuk upper approximation of a concept is included in the Pawlak one provided that −1 is serial (see Proposition 6g). By (c), the Pawlak upper approximation of the empty set is empty as well. On the other hand, both the lower approximation of the universe as well as the Skowron–Stepaniuk upper approximation of the universe are equal to the whole universe. According to (d), the Pawlak lower and upper approximations are dual to each other. Moreover, all objects u with empty elementary granules Γ u belong to the lower approximation of any concept. By (e), the lower and upper approximation mappings are monotone. Property (f), being

72

A. Gomoli´ nska

a direct consequence of (e), may be strengthened for the Pawlak approximation mappings as follows: low(X ∩ Y ) = lowX ∩ lowY & upp(X ∪ Y ) = uppX ∪ uppY

(16)

We show the 1st property. “⊆” holds by Proposition 2f. For “⊇” assume that u ∈ lowX ∩ lowY . Hence, u ∈ lowX and u ∈ lowY . By the deﬁnition of low, Γ u ⊆ X and Γ u ⊆ Y . Thus, Γ u ⊆ X ∩ Y . By the deﬁnition of low, u ∈ low(X ∩ Y ). Below, we collect several facts about compositions of the approximation mappings examined. Proposition 3. In cases (d)–(f ), assume rif 4 (κ). The following may be obtained: (a) low ◦ low uppS ◦ low uppS ◦ uppS (b) low ◦ low low ◦ uppS uppS ◦ uppS (c) upp ◦ low upp ◦ uppS & low ◦ upp uppS ◦ upp (d) upp ◦ low uppS ◦ low & low ◦ upp low ◦ uppS (e) upp ◦ upp uppS ◦ upp uppS ◦ uppS (f ) upp ◦ upp upp ◦ uppS uppS ◦ uppS Proof. We prove (a), (e) only. In case (a), low ◦ low uppS ◦ low by Proposition 2a and Proposition 1d. Next, uppS ◦ low uppS ◦ uppS by Proposition 2a, monotonicity of uppS , and Proposition 1e. In case (e), assume rif 4 (κ). By Proposition 2b, upp uppS . Hence, upp◦upp uppS ◦upp by Proposition 1d, whereas uppS ◦upp uppS ◦uppS by monotonicity of uppS and Proposition 1e.

Clearly, properties analogous to Proposition 2, Proposition 3, and (16) hold for the ∗-versions of the basic mappings5 . Now, we can formulate several properties of the -deﬁnable versions of low, lowS , upp, and uppS . Proposition 4. For any concepts X, Y , we can prove that: (a) low∪ = lowS∪ uppS∪ upp∪ id℘U (b) rif 4 (κ) ⇒ upp∪ uppS∪ (c) upp∪ ∅ = ∅ (d) X ⊆ Y ⇒ f ∪ X ⊆ f ∪ Y (e) f ∪ (X ∩ Y ) ⊆ f ∪ X ∩ f ∪ Y ⊆ f ∪ X ∪ f ∪ Y ⊆ f ∪ (X ∪ Y ) Proof. We prove (a) only. To this end, let X be any concept and u be any object. First, low∪ = lowS∪ directly by (12) and Proposition 2a. In virtue of 5

To safe space, we do not formulate them explicitly. When referring to them, we shall attach ∗ to the name of a property as a superscript. For instance, we shall refer to ∀X, Y.(X ⊆ Y ⇒ f ∗ X ⊆ f ∗ Y ), being the counterpart of Proposition 2e, as Proposition 2e∗ .

A Comparison of Pawlak’s and Skowron–Stepaniuk’s Approximation

73

Proposition 1g, it remains to show that (a1) low∪ uppS∪ , (a2) low∪ upp∪ , and (a3) low∪ id℘U . (a1) holds by Proposition 2a and Proposition 2e∗ for f = upp∗ . For (a2) it suﬃces to prove low∪ X ⊆ upp∪ X. To this end, assume u ∈ low∪ X. By (12), u ∈ upp∗ (lowX), i.e., there is u such that (a4) u ∈ Γ u and u ∈ lowX. Hence, Γ u ⊆ X by the deﬁnition of low. Note that Γ u ∩ X = ∅ since Γ u = ∅ in virtue of (a4). Hence, u ∈ uppX by the deﬁnition of upp. As a consequence, u ∈ upp∗ (uppX) = upp∪ X due to (a4) and (12). For (a3) we prove that low∪ X ⊆ X. To this end, let u ∈ low∪ X. By arguments as earlier,

there is u such that u ∈ Γ u ⊆ X. Immediately, u ∈ X. In the case of the -deﬁnable versions of approximation mappings, the Pawlak lower approximation and the Skowron–Stepaniuk lower approximation are equal. The novelty is that the lower approximation of a concept is included not only in the Skowron–Stepaniuk upper approximation of that concept but also in the concept itself and in the Pawlak upper approximation of that concept. Obviously, (b) is a counterpart of Proposition 2b, whereas (c) corresponds to the 1st part of Proposition 2c. The remaining two properties are counterparts of Proposition 2e and Proposition 2f, respectively. For example, (d) says that the mappings of lower and upper approximations are monotone. Additionally, we can derive the counterpart of (16) for the Pawlak upper approximation: upp∪ (X ∪ Y ) = upp∪ X ∪ upp∪ Y.

(17)

Several observations upon compositions of the -deﬁnable versions of the basic approximation mappings are presented below. Proposition 5. In cases (e)–(g), assume rif 4 (κ). The following can be derived: (a) low∪ ◦ low∪ = low∪ upp∪ ◦ low∪ low∪ ◦ upp∪ = upp∪ upp∪ ◦ upp∪ (b) low∪ ◦ low∪ uppS∪ ◦ low∪ low∪ ◦ uppS∪ = uppS∪ uppS∪ ◦ uppS∪ (c) (upp∪ ◦ low∪ ) uppS∪ upp∪ ◦ uppS∪ (d) (uppS∪ ◦ low∪ ) upp∪ uppS∪ ◦ upp∪ (e) upp∪ upp∪ ◦ uppS∪ & upp∪ ◦ low∪ uppS∪ ◦ low∪ (f ) upp∪ ◦ upp∪ uppS∪ ◦ upp∪ uppS∪ ◦ uppS∪ (g) upp∪ ◦ upp∪ upp∪ ◦ uppS∪ uppS∪ ◦ uppS∪ Proof. We prove (a), (b) only. In case (a), ﬁrst note that low∪ ◦ low∪ low∪ , low∪ ◦ low∪ upp∪ ◦ low∪ , low∪ ◦ upp∪ upp∪ ◦ upp∪ , and low∪ ◦ upp∪ upp∪ by Proposition 4a and Proposition 1d. Subsequently, upp∪ ◦ low∪ upp∪ holds by Proposition 4a, monotonicity of upp∪ , and Proposition 1e. Now, we prove that low∪ low∪ ◦ low∪ . To this end, it suﬃces to show low low ◦ low∪ . Then, by monotonicity of upp∗ and Proposition 1e, we obtain upp∗ ◦ low upp∗ ◦ (low ◦ low∪ ) = (upp∗ ◦ low) ◦ low∪ which ﬁnally results in low∪ low∪ ◦low∪ by (12). Thus, consider a concept X and an object u such that u ∈ lowX. Hence, (a1) Γ u ⊆ X by the deﬁnition of low. Then, for every u ∈ Γ u, u ∈ low∪ X by (a1) and the deﬁnition of low∪ . In other words, Γ u ⊆ low∪ X.

74

A. Gomoli´ nska

By the deﬁnition of low, u ∈ low(low∪ X) = (low ◦ low∪ )X. In this way, we have proved that lowX ⊆ (low ◦ low∪ )X. Hence, immediately, low low ◦ low∪ by the deﬁnition of . Finally, we show that upp∪ low∪ ◦ upp∪ . As in the preceding case, it suﬃces to prove upp low ◦ upp∪ and, then, to apply Proposition 2e∗ , Proposition 1e, and (12). Thus, consider a concept X and an object u such that u ∈ uppX. Immediately, (a2) Γ u ∩ X = ∅ by the deﬁnition of upp. As earlier, for every u ∈ Γ u, u ∈ upp∪ X by (a2) and the deﬁnition of upp∪ . As a consequence, Γ u ⊆ upp∪ X. Hence, u ∈ low(upp∪ X) = (low ◦ upp∪ )X by the deﬁnition of low. We have shown that uppX ⊆ (low ◦ upp∪ )X. Finally, upp low ◦ upp∪ by the deﬁnition of . In case (b), note that low∪ ◦ low∪ uppS∪ ◦ low∪ , low∪ ◦ uppS∪ uppS∪ ◦ uppS∪ , and low∪ ◦uppS∪ uppS∪ follow from Proposition 4a and Proposition 1d. Moreover, uppS∪ ◦ low∪ uppS∪ by Proposition 4a, monotonicity of uppS∪ , and Proposition 1e. It remains to show that uppS∪ low∪ ◦uppS∪ . It suﬃces to prove uppS low◦uppS∪ and, then, to apply Proposition 2e∗ , Proposition 1e, and (12). To this end, consider a concept X and an object u such that u ∈ uppS X. Hence, (b1) κ(Γ u, X) > 0 by the deﬁnition of uppS . Then, for every u ∈ Γ u, u ∈ uppS∪ X by (b1) and the deﬁnition of uppS∪ . Thus, Γ u ⊆ uppS∪ X. Hence, u ∈ low(uppS∪ X) = (low ◦ uppS∪ )X by the deﬁnition of low. That is, we have proved uppS X ⊆ (low ◦ uppS∪ )X. Immediately, uppS low ◦ uppS∪ by the deﬁnition of .

In comparison to Proposition 3, more and stronger relationships may be noted than in the basic case. It is due to Proposition 4a and to the fact that if an object u belongs to the -deﬁnable lower or upper approximation of a concept (in either of the senses considered), then its elementary granule Γ ∗ u is non-empty. By way of example, the composition of the lower approximation mapping with itself equals to the lower approximation mapping, whereas the compositions of the upper approximation mappings with the lower approximation mapping equal to the former mappings in virtue of (a), (b). Such results cannot be obtained for the basic approximation mappings without extra assumptions about .

5

Properties of Approximation Mappings II

In this section, we present and discuss properties of approximation mappings for special cases of approximation spaces. In detail, we consider approximation spaces which, in turn, are based on serial relations, reﬂexive relations, transitive relations, symmetric relations (and, in particular, tolerance relations), and — last but not least — equivalence relations6 . 6

A technical remark can be handy here. Except for a few cases, in a given subsection, we only present these properties which can be derived for the kind of approximation space investigated, yet were not obtained under weaker assumptions. For instance, when discussing approximation spaces based on reﬂexive relations, we do not recall the properties obtained for spaces based on serial relations and, the more, for arbitrary spaces.

A Comparison of Pawlak’s and Skowron–Stepaniuk’s Approximation

5.1

75

The Case of Serial Relations

In this section, we discuss two independent cases, viz., the case where −1 is serial and the case where is serial. Let SER(U ) denote the set of all serial relations on U . First, let −1 ∈ SER(U ). That is, for every u ∈ U , there exists u ∈ U such that (u , u) ∈ (in other words, ← {u} = ∅ or Γ u = ∅). Proposition 6. Let X be any concept. Then, we have: (a) uppU = U & low∅ = low∪ ∅ = ∅ (b) rif 3 (κ) ⇒ uppS ∅ = uppS∪ ∅ = ∅ (c) low upp (d) low ◦ low upp ◦ low upp ◦ upp (e) low ◦ low low ◦ upp upp ◦ upp (f ) uppS ◦ low uppS ◦ upp & low ◦ uppS upp ◦ uppS (g) rif 4∗ (κ) ⇒ uppS upp & uppS∪ upp∪ (h) rif 5 (κ) ⇒ uppS = upp & uppS∪ = upp∪ Proof. We prove (a), (b) only. Consider any object u. For (a) note that Γ u ∩ U = ∅ by seriality of −1 . Hence, u ∈ uppU by the deﬁnition of upp. Thus, U ⊆ uppU and, ﬁnally, uppU = U . Moreover, it can never be Γ u ⊆ ∅. Immediately, low∅ = ∅ by the deﬁnition of low. Hence, low∪ ∅ = (upp∗ ◦ low)∅ = upp∗ (low∅) = upp∗ ∅ = ∅ by (12) and Proposition 2c∗ . For (b) assume that κ satisﬁes rif 3 . By seriality of −1 , Γ u = ∅. Then, (b1) κ(Γ u, ∅) = 0 in virtue of rif 3 (κ). Hence, (b2) uppS ∅ = ∅ by the deﬁnition of uppS . Next, uppS∪ ∅ = (upp∗ ◦ uppS )∅ = upp∗ (uppS ∅) = upp∗ ∅ = ∅ by (b2), (12), and Proposition 2c∗ .

Let us brieﬂy comment upon the results. Seriality of −1 guarantees that elementary granules of the form Γ u are non-empty. Thanks to that, the Pawlak upper approximation of the universe is the universe itself and the lower approximations of the empty set are empty by (a). Moreover, in virtue of (b), if rif 3 is satisﬁed by κ, then the Skowron–Stepaniuk upper approximations of the empty set are empty. By (c), the lower approximation of a concept is included in the Pawlak upper approximation of that concept. The next three properties, being consequences of (c), augment Proposition 3 by new facts on compositions of approximation mappings. If rif 4∗ is satisﬁed by κ, then the Skowron–Stepaniuk upper approximations of a concept are included in the corresponding Pawlak upper approximations of that concept by (g). Moreover, if rif 5 (κ) holds, then the Pawlak upper approximations and the Skowron–Stepaniuk upper approximations coincide due to (h). We now consider the case, where ∈ SER(U ). Then, for every u ∈ U , there exists u ∈ U such that (u, u ) ∈ (in other words, → {u} = ∅ or Γ ∗ u = ∅). First, observe that properties analogous to Proposition 6 can be obtained for the ∗-versions of the basic approximation mappings. In particular, upp∗ U = U . Furthermore, we can prove the following properties.

76

A. Gomoli´ nska

Proposition 7. Let X be any concept. Then, it holds: (a) uppS∪ U = U & (−1 ∈ SER(U ) ⇒ upp∪ U = U ) (b) upp∪ X ∪ low∪ (U − X) = U (c) id℘U upp∪ & (rif 4 (κ) ⇒ id℘U uppS∪ ) (d) uppS∪ uppS∪ ◦ upp∪ Proof. We show (a), (b) only. To this end, consider any concept X. For the 1st part of (a) note that uppS∪ U = (upp∗ ◦ uppS )U = upp∗ (uppS U ) = upp∗ U = U by (12), seriality of , Proposition 2c, and Proposition 6a∗ . Now, assume additionally that −1 is serial. Hence, upp∪ U = (upp∗ ◦ upp)U = upp∗ (uppU ) = upp∗ U = U by (12), Proposition 6a, seriality of , and Proposition 6a∗ . In case (b), upp∪ X ∪ low∪ (U − X) = (upp∗ ◦ upp)X ∪ (upp∗ ◦ low)(U − X) = upp∗ (uppX) ∪ upp∗ (low(U − X)) = upp∗ (uppX ∪ low(U − X)) = upp∗ U = U

by (12), (16∗ ), Proposition 2d, seriality of , and Proposition 6a∗ . Most of the properties above strongly depend on Proposition 6a∗ (a counterpart of Proposition 6a) which is a consequence of seriality of . By (a), the Skowron– Stepaniuk upper approximation of the universe is the whole universe, whereas the Pawlak upper approximation of the universe is the universe if both , −1 are serial. According to (b), every object belongs to the Pawlak upper approximation of a concept and/or to the lower approximation of the complement of that concept. (c) states that every concept is included in its Pawlak upper approximation, and similarly for the Skowron–Stepaniuk upper approximation if κ satisﬁes rif 4 . Finally, the Skowron–Stepaniuk upper approximation of a concept is included in the Skowron–Stepaniuk upper approximation of the Pawlak upper approximation of that concept due to (d). 5.2

The Case of Reﬂexive Relations

Assume that is reﬂexive. Then, for every u ∈ U , (u, u) ∈ . Immediately, (u, u) ∈ −1 , so −1 is reﬂexive as well. Thus, u ∈ Γ u ∩ Γ ∗ u. Clearly, every reﬂexive relation is serial as well. In the context of approximation spaces, reﬂexive relations are referred to as similarity relations. The set of all reﬂexive relations on U will be denoted by RF(U ). Proposition 8. We can prove that: (a) low id℘U upp & (rif 4 (κ) ⇒ id℘U uppS ) (b) f f ∪ (c) low ◦ low low upp ◦ low upp upp ◦ upp (d) low low ◦ upp upp (e) uppS ◦ low low ◦ uppS uppS uppS ◦ upp upp ◦ uppS

A Comparison of Pawlak’s and Skowron–Stepaniuk’s Approximation

77

Proof. We prove (e) only. First, uppS ◦ low uppS , uppS uppS ◦ upp by (a), monotonicity of uppS , and Proposition 1e. Next, low ◦ uppS uppS , uppS upp ◦ uppS by (a) and Proposition 1d.

If is reﬂexive, then Proposition 6c may be strengthened to the property (a) above. In detail, the lower approximation of a concept is included in that concept and, on the other hand, every concept is included in its Pawlak upper approximation. The same holds for the Skowron–Stepaniuk approximation if rif 4 (κ) is satisﬁed. By (b), the lower (resp., upper) approximation of a concept is included in the -deﬁnable version of the lower (upper) approximation of that concept. Moreover, the list of properties of compositions of approximation mappings can be extended with several new dependencies given by (c)–(e). 5.3

The Case of Transitive Relations

Now, suppose that is transitive, i.e., for every u, u , u ∈ U , (u, u ) ∈ and (u , u ) ∈ imply (u, u ) ∈ . Either both , −1 are transitive or both of them are not transitive. We denote the set of all transitive relations on U by TR(U ). Proposition 9. The following dependencies can be proved: (a) low∪ low (b) low low ◦ low & upp ◦ upp upp (c) ∈ RF(U ) ⇒ low ◦ low = low = low∪ & upp ◦ upp = upp Proof. We prove (a), (b) only. Consider any concept X and any object u. In case (a), we show that low∪ X ⊆ lowX which results in low∪ low by the deﬁnition of . To this end, assume that u ∈ low∪ X. By the deﬁnition of low∪ , there is u such that u ∈ Γ u , i.e., (a1) (u, u ) ∈ , and (a2) u ∈ lowX. Next, (a3) Γ u ⊆ X by (a2) and the deﬁnition of low. Consider any u ∈ Γ u (i.e., (u , u) ∈ ). Hence, (u , u ) ∈ in virtue of (a1) and transitivity of . In other words, u ∈ Γ u . Hence, immediately, u ∈ X by (a3). Thus, Γ u ⊆ X. Hence, u ∈ lowX by the deﬁnition of low. For the 1st part of (b) it suﬃces to prove that lowX ⊆ (low ◦ low)X, i.e., lowX ⊆ low(lowX), and to apply the deﬁnition of . To this end, assume that u ∈ lowX. By the deﬁnition of low, (b1) Γ u ⊆ X. We need to prove that u ∈ low(lowX), i.e., Γ u ⊆ lowX in virtue of the deﬁnition of low. Thus, consider any u ∈ Γ u (i.e., (b2) (u , u) ∈ ). It remains to show that u ∈ lowX, i.e., Γ u ⊆ X. Let u ∈ Γ u , i.e., (u , u ) ∈ . Hence, (u , u) ∈ by (b2) and transitivity of . That is, u ∈ Γ u. In virtue of (b1), u ∈ X which ends the proof of this part of (b). For the remaining part of (b) it suﬃces to show that (upp ◦ upp)X ⊆ uppX, i.e., upp(uppX) ⊆ uppX, and to apply the deﬁnition of . To this end, let u ∈ upp(uppX). By the deﬁnition of upp, Γ u ∩ uppX = ∅. Hence, there is (b3) u ∈ Γ u such that u ∈ uppX. By the deﬁnition of upp, Γ u ∩ X = ∅. Hence, there is (b4) u ∈ Γ u such that (b5) u ∈ X. Note that (u , u) ∈ and (u , u ) ∈ by (b3) and (b4), respectively. Hence, by transitivity of ,

78

A. Gomoli´ nska

(u , u) ∈ as well. That is, u ∈ Γ u. As a consequence, Γ u ∩ X = ∅ by (b5). By the deﬁnition of upp, u ∈ uppX.

Some comments can be useful. Due to transitivity of , the -deﬁnable version of the lower approximation of a concept is included in the lower approximation of that concept by (a). Moreover, the lower approximation of a concept is included in the lower approximation of the lower approximation of that concept, whereas the Pawlak upper approximation of the same form of the upper approximation of a concept is included in the Pawlak upper approximation of that concept by (b). Unfortunately, a similar result does not seem to hold for the Skowron–Stepaniuk upper approximation. In the sequel, assuming reﬂexivity and transitivity of , the number of diﬀerent compositions of approximation mappings may substantially be reduced thanks to (c) since, then, both versions of the lower approximation are equal to the composition of the lower approximation with itself, and the Pawlak upper approximation is equal to the composition of the very form of the upper approximation with itself. 5.4

The Case of Symmetric Relations

In this section, we examine the case, where is symmetric. Then, for every u, u ∈ U , (u, u ) ∈ implies (u , u) ∈ . Immediately, = −1 and Γ = Γ ∗ . Relations which are both reﬂexive and symmetric are called tolerance relations. Obviously, every tolerance relation is also a similarity relation7 . The sets of all symmetric relations and all tolerance relations on U will be denoted by SYM(U ) and TL(U ), respectively. Proposition 10. The following properties hold: (a) f = f ∗ & f ∪ = upp ◦ f (b) upp ◦ low low ◦ upp (c) ∈ TR(U ) ⇒ upp ◦ low low & upp∪ upp low ◦ upp & upp∪ ◦ f ∪ f ∪ Proof. We show (b), (c) only. Consider any concept X and any object u. In case (b), assume that u ∈ upp(lowX). By the deﬁnition of upp, Γ u∩lowX = ∅. Hence, there is u such that (b1) u ∈ Γ u and u ∈ lowX. From the latter, (b2) Γ u ⊆ X by the deﬁnition of low. In virtue of (b1), (u , u) ∈ . Hence, by symmetry of , (u, u ) ∈ . That is, u ∈ Γ u . Hence, (b3) u ∈ X due to (b2). We need to show that u ∈ low(uppX), i.e., Γ u ⊆ uppX by the deﬁnition of low. To this end, consider u ∈ Γ u (i.e., (u , u) ∈ ). By symmetry of , (u, u ) ∈ as well, i.e., u ∈ Γ u . Hence, Γ u ∩ X = ∅ by (b3). By the deﬁnition of upp, u ∈ uppX as required. Thus, we have proved that upp(lowX) ⊆ low(uppX), i.e., (upp ◦ low)X ⊆ (low ◦ upp)X. Immediately, upp ◦ low low ◦ upp by the deﬁnition of . 7

Tolerance relations will not be a subject to a separate study in this article. To list their properties, it suﬃces to merge the proposition below with the facts presented in Sect. 5.2.

A Comparison of Pawlak’s and Skowron–Stepaniuk’s Approximation

79

In case (c), assume additionally that is transitive. First, upp ◦ low low by the assumption, the 2nd part of (a), and Proposition 9a. Next, upp∪ upp by the assumption, the 2nd part of (a), and Proposition 9b. In the sequel, we show that uppX ⊆ (low ◦ upp)X, i.e., uppX ⊆ low(uppX). Then, upp low ◦ upp by the deﬁnition of . To this end, suppose that u ∈ uppX. By the deﬁnition of upp, Γ u ∩ X = ∅. Hence, there is u such that (c1) u ∈ X and u ∈ Γ u (i.e., (c2) (u , u) ∈ ). Consider any u ∈ Γ u (i.e., (u , u) ∈ ). By symmetry of , (u, u ) ∈ . Hence, (u , u ) ∈ by (c2) and transitivity of . In other words, u ∈ Γ u . Hence, Γ u ∩ X = ∅ in virtue of (c1). By the deﬁnition of upp, u ∈ uppX. Thus, we have proved that Γ u ⊆ uppX, i.e., u ∈ low(uppX) by the deﬁnition of low. For the remaining part of (c) note that upp∪ ◦ f ∪ = (upp ◦ upp) ◦ (upp ◦ f ) upp ◦ (upp ◦ f ) = (upp ◦ upp) ◦ f upp ◦ f = f ∪ by (a), transitivity of , Proposition 9b, and Proposition 1d.

Thus, whenever symmetry of is assumed, the ∗-versions of the basic approximation mappings coincide with the very mappings in virtue of (a). As a consequence, the -deﬁnable versions of the basic mappings are compositions of these mappings with the Pawlak upper approximation mapping. Due to (b), the Pawlak upper approximation of the lower approximation of a concept is included in the lower approximation of the Pawlak upper approximation of that concept. In virtue of (c), if is both symmetric and transitive, then — among others — the Pawlak upper approximation of the lower approximation of a concept is included in the lower approximation of that concept, the -deﬁnable version of the Pawlak upper approximation of a concept is included in the Pawlak upper approximation of that concept, and the latter one is included in the lower approximation of the Pawlak upper approximation of the very concept. 5.5

The Case of Equivalence Relations

Finally, we consider the case, where is an equivalence relation. In the context of Pawlak’s information systems and approximation spaces, equivalence relations on the set of objects are understood as indiscernibility relations. By deﬁnition, an equivalence relation is simultaneously reﬂexive, symmetric, and transitive, i.e., it is a transitive tolerance relation. Let EQ(U ) denote the set of all equivalence relations on U . Thus, EQ(U ) = TL(U ) ∩ TR(U ). Note that EQ(U ) = SER(U ) ∩ SYM(U ) ∩ TR(U ) as well. In the sequel, Γ u is called an equivalence class of u and it may be denoted by [u] (or simply, by [u] if is understood) along the standard lines. Note that ∀u, u ∈ U.(u ∈ Γ u ⇔ Γ u = Γ u). Proposition 11. It holds that: (a) f ∪ = f (b) low = low ◦ low = upp ◦ low low ◦ upp = upp ◦ upp = upp (c) low uppS ◦ low low ◦ uppS = uppS = upp ◦ uppS uppS ◦ upp

(18)

80

A. Gomoli´ nska

Proof. We show (a) only. Note that “” holds by reﬂexivity of and Proposition 8b. In the sequel, low∪ low by transitivity of and Proposition 9a. Moreover, upp∪ upp by symmetry and transitivity of , and Proposition 10c. Now, consider any concept X and any object u. To prove uppS∪ uppS , it sufﬁces to show that uppS∪ X ⊆ uppS X. Thus, suppose that u ∈ uppS∪ X. Hence, there is u such that (a1) u ∈ Γ u and u ∈ uppS X by the deﬁnition of uppS∪ . It follows from the latter that κ(Γ u , X) > 0 by the deﬁnition of uppS . By (18) and (a1), κ(Γ u, X) > 0. Finally, u ∈ uppS X by the deﬁnition of uppS .

Thus, whenever is an equivalence relation, the -deﬁnable versions of the basic approximation mappings coincide with the mappings by (a), respectively. As regarding the reduction in the number of diﬀerent compositions of the mappings considered, both the composition of the lower approximation mapping with itself and the composition of the lower approximation mapping with the Pawlak upper approximation mapping are equal (and hence, may be reduced) to the lower approximation mapping. Furthermore, both the composition of the Pawlak upper approximation mapping with the lower approximation mapping and the composition of the Pawlak upper approximation mapping with itself are equal to the Pawlak upper approximation mapping thanks to (b). Moreover by (c), both the composition of the Skowron–Stepaniuk upper approximation mapping with the lower approximation mapping and the composition of the Skowron–Stepaniuk upper approximation mapping with the Pawlak upper approximation mapping are equal to the Skowron–Stepaniuk upper approximation mapping. Finally, the lower approximation of a concept is included in the Skowron–Stepaniuk upper approximation of the lower approximation of that concept, the latter set is included in the Skowron–Stepaniuk upper approximation of the concept, and the very upper approximation of a concept is included in the Skowron–Stepaniuk upper approximation of the Pawlak upper approximation of that concept. Last but not least, observe that the upper approximation mappings uppS1 , uppS2 investigated in Example 1 are equal to upp if is an equivalence relation. Indeed, upp ◦ Γ = Γ then, and for any object u and any concept X, u ∈ uppS2 X if and only if Γ u ∩ upp∗ X = ∅ if and only if Γ u ∩ uppX = ∅ if and only if u ∈ upp(uppX) if and only if u ∈ uppX.

6

Summary

In this article, we studied and compared Pawlak’s rough approximation of concepts with Skowron–Stepaniuk’s approach within a general framework of approximation spaces of the form (U, , κ), where U is a non-empty set of objects, is a non-empty binary relation on U , and κ is a RIF satisfying rif 1 and rif 2 . The lower approximation mappings are the same in both approaches unlike the upper approximation ones8 . The latter mappings cannot be compared directly without additional assumptions made about κ. For the sake of illustration, we 8

The fact that Pawlak’s and Skowron–Stepaniuk’s upper approximations coincide for the standard RIF was known earlier.

A Comparison of Pawlak’s and Skowron–Stepaniuk’s Approximation

81

considered two special cases of κ and two corresponding Skowron–Stepaniuk upper approximation mappings in Example 1. In general, these two kinds of mappings are diﬀerent from each other and from Pawlak’s upper approximation mapping. However — as turned out — the three cases coincide if is an equivalence relation. In the paper presented, we compared Pawlak’s and Skowron–Stepaniuk’s upper approximation mappings indirectly by investigation of their properties. It is diﬃcult to say which mappings are totally better. While Pawlak’s approximation seems to be more suitable in some aspects, it is the Skowron–Stepaniuk approach which seems to provide more interesting results in other cases. As a side-eﬀect, we have obtained a fairly exhaustive list of basic mathematical properties of the mappings investigated, to be used in the future research and applications.

References 1. Pawlak, Z.: Rough set elements. In Polkowski, L., Skowron, A., eds.: Rough Sets in Knowledge Discovery 1. Volume 18 of Studies in Fuzziness and Soft Computing. Physica-Verlag, Heidelberg (1998) 10–30 2. Pawlak, Z.: Information systems – Theoretical foundations. Information Systems 6 (1981) 205–218 3. Pawlak, Z.: Rough sets. Computer and Information Sciences 11 (1982) 341–356 4. Pawlak, Z.: Information Systems. Theoretical Foundations (in Polish). Wydawnictwo Naukowo-Techniczne, Warsaw (1983) 5. Pawlak, Z.: Rough Sets. Theoretical Aspects of Reasoning About Data. Kluwer, Dordrecht (1991) 6. Pomykala, J.A.: Approximation operations in approximation space. Bull. Polish Acad. Sci. Math. 35 (1987) 653–662 7. Wybraniec-Skardowska, U.: On a generalization of approximation space. Bull. Polish Acad. Sci. Math. 37 (1989) 51–62 ˙ 8. Zakowski, W.: Approximations in the space (U, Π). Demonstratio Mathematica 16 (1983) 761–769 9. Skowron, A., Stepaniuk, J.: Generalized approximation spaces. In: Proc. 3rd Int. Workshop Rough Sets and Soft Computing, San Jose, USA, 1994, November 10-12. (1994) 156–163 10. Skowron, A., Stepaniuk, J.: Tolerance approximation spaces. Fundamenta Informaticae 27 (1996) 245–253 11. Gomoli´ nska, A.: Variable-precision compatibility spaces. Electronical Notices in Theoretical Computer Science 82 (2003) 1–12 http://www.elsevier.nl/locate/entcs/volume82.html. 12. Slowi´ nski, R., Vanderpooten, D.: Similarity relation as a basis for rough approximations. In Wang, P.P., ed.: Advances in Machine Intelligence and Soft Computing. Volume 4. Duke University Press (1997) 17–33 13. Yao, Y.Y., Wong, S.K.M., Lin, T.Y.: A review of rough set models. In Lin, T.Y., Cercone, N., eds.: Rough Sets and Data Mining: Analysis of Imprecise Data. Kluwer, Boston London Dordrecht (1997) 47–75 14. Ziarko, W.: Variable precision rough set model. J. Computer and System Sciences 46 (1993) 39–59 15. Ziarko, W.: Probabilistic decision tables in the variable precision rough set model. J. Comput. Intelligence 17 (2001) 593–603

82

A. Gomoli´ nska

16. Peters, J.F.: Approximation space for intelligent system design patterns. Engineering Applications of Artiﬁcial Intelligence 17 (2004) 1–8 17. Skowron, A.: Approximation spaces in rough neurocomputing. In Inuiguchi, M., Hirano, S., Tsumoto, S., eds.: Rough Set Theory and Granular Computing. Volume 125 of Studies in Fuzziness and Soft Computing. Springer-Verlag, Berlin Heidelberg (2003) 13–22 18. Skowron, A., Swiniarski, R., Synak, P.: Approximation spaces and information granulation. Transactions on Rough Sets III: Journal Subline to Lecture Notes in Computer Science 3400 (2005) 175–189 19. Polkowski, L., Skowron, A.: Rough mereology. Lecture Notes in Artiﬁcial Intelligence 869 (1994) 85–94 20. Polkowski, L., Skowron, A.: Rough mereology: A new paradigm for approximate reasoning. Int. J. Approximated Reasoning 15 (1996) 333–365 21. L ukasiewicz, J.: Die logischen Grundlagen der Wahrscheinlichkeitsrechnung. In Borkowski, L., ed.: Jan L ukasiewicz – Selected Works. North Holland, Polish Scientiﬁc Publ., Amsterdam London, Warsaw (1970) 16–63 First published in Krak´ ow in 1913. 22. Zadeh, L.A.: Outline of a new approach to the analysis of complex system and decision processes. IEEE Trans. on Systems, Man, and Cybernetics 3 (1973) 28–44

Data Preparation for Data Mining in Medical Data Sets Grzegorz Ilczuk1 and Alicja Wakulicz-Deja2 1

Siemens AG Medical Solutions, Allee am Roethelheimpark 2, 91052 Erlangen, Germany [email protected] 2 Institut of Informatics University of Silesia, Bedzinska 39, 41-200 Sosnowiec, Poland [email protected]

Abstract. Data preparation is a very important but also a time consuming part of a Data Mining process. In this paper we describe a hierarchical method of text classiﬁcation based on regular expressions. We use the presented method in our data mining system during a pre-processing stage to transform Latin free-text medical reports into a decision table. Such decision tables are used as an input for rough sets based rule induction subsystem. In this study we also compare accuracy and scalability of our method with a standard approach based on dictionary phrases. Keywords: rough sets, data preparation, regular expression.

1

Introduction

Preparation of data takes about 60% of a time needed for the whole Data Mining process and it is also deﬁned by Pyle as the most important part of a Data Exploration Process which leads to success [1]. This estimation is also valid in case of our Data Exploration system, where the entry stage of data processing is a key element for the further analysis. The mentioned Data Exploration system will be used in medicine (especially in cardiology) as a complete solution suitable for improving medical care and clinical work ﬂow through revealing new patterns and relations among data. Functional blocks of the system are: – Import subsystem-responsible for importing data from medical information systems into our storage subsystem – Data recognition subsystem-during this stage we use algorithms and methods described in this paper to transform the raw data to a form suitable for further Data Exploration – Data preprocessing-based on the statistical analysis of the transformed information noise and redundant data are removed [9] – Feature selection-this stage utilizes a few attribute reduction methods such as CFS (Correlation-based Feature Selection), Quickreduct and conjunction of these methods to select an optimal set of attributes for a further analysis J.F. Peters et al. (Eds.): Transactions on Rough Sets VI, LNCS 4374, pp. 83–93, 2007. c Springer-Verlag Berlin Heidelberg 2007

84

G. Ilczuk and A. Wakulicz-Deja

– Rule induction subsystem based on Rough Set Theory [6,7,11]. Early research on this area was described in [8,12] – Visualization of the knowledge discovery in a form easily understandable by humans for validating and extending of the collected knowledge [13] Most medical information useful for Data Mining is still written in form of freetext Latin reports. These reports are mostly used to extend a lapidary diagnosis written with statistical ICD-10 codes. There are some challenges to solve during analyzing such reports such as: diﬀerent descriptions for the same disease, non-standard abbreviations, misspelled words and a ﬂoating structure of such reports. In our ﬁrst solution of these problems we had used a phrase dictionary to map information from a report to an attribute. The main disadvantage of this approach was a lack of scalability and a diﬃcult maintenance. These facts leaded us to develop a diﬀerent approach. The proposed method and results achieved with it are presented in this study. Described in this paper technique is used in our Data Exploration system as a preprocessing step, which prepares data for rule induction. For generation of decision rules an own implementation of MLEM2 algorithm is used.

2

Rough Sets: Basic Notions and Medical Appliance

Developed by Pawlak and presented in 1982 Rough Sets theory is a mathematical approach to handle imprecision and uncertainty [4]. The main goal of rough set analysis is to synthesize approximation of concepts from the acquired data. Some basic deﬁnitions are presented below. Information system [4] is a pair A = (U, A) where U is a non-empty, ﬁnite set called the universe and A is a non-empty, ﬁnite set of attributes, i.e. a : U → Va for a ∈ A, where Va is called the value set of attribute a. Elements of U are called objects. The special case of information systems called decision system is deﬁned as A = (U, A ∪ {d}), where d ∈ A is a distinguished attribute called decision and elements of A are called conditions. A decision rule is deﬁned as r = (ai1 = v1 ) ∧ . . . ∧ (aim = vm ) ⇒ (d = k) where 1 ≤ i1 < . . . < im ≤ |A|, vi ∈ Vai . We say an object matches a rule if its attributes satisfy all atomic formulas (aij = vj ) of the rule. A rule is called minimal consistent with A when any decision rule r created from r by removing one of atomic formula of r is not consistent with A. In our Data Exploration system we use a modiﬁed version of LEM2 algoritm - MLEM2 to generate decision rules. LEM2 (Learning from Examples Module, version 2) algorithm was ﬁrstly presented in [14,15] and then implemented in [16]. LEM2 induces a rule set by exploring the space of blocks of attributevalue pairs to generate a local covering. Afterwards the found local covering is converted into the rule set. Following deﬁnitions must be quoted prior to deﬁne a local covering [11]. For a variable (attribute or decision) x and its value v, a block [(x, v)] of a variable-value pair (x, v) is the set of all cases for which variable x has value v.

Data Preparation for Data Mining in Medical Data Sets

85

Let B be a nonempty lower or upper approximation of a concept represented by a decision-value pair (d, w). Set B depends on a set T of attribute-value pairs (a, v) if and only if ∅ = [T ] =

[(a, v)] ⊆ B.

(1)

(a,v)∈T

Set T is a minimal complex of B if and only if B depends on T and no proper subset T , of T exists such that B depends on T , . Let T be a nonempty collection of nonempty sets of attribute-value pairs. Then T is a local covering of B if and only if the following conditions are satisﬁed: – each member T of T is a minimal complex of B, – T ∈T [T ] = B, and – T is minimal, i.e., T has the smallest possible number of members. Modiﬁed LEM2 (MLEM2) proposed by Grzymala-Busse in [11] in compare to LEM2 allows inducing rules from data containing numerical attributes without a need of a separate discretization step. Our implementation of MLEM2 algorithm induces decision rules from both lower approximation (certain rules) and upper approximation (possible rules). This technique allows us reasoning from ”real” data, which contains uncertain, noisy and redundant information. Decision rules are used in our system to present the extracted knowledge from medical data. This approach in medical domain has several advantages over other data mining techniques: – – – – –

Decision rules are easy to understand and verify Decision rules can be easily validated with existing knowledge Gathered decision rules can be modiﬁed and extended with a new knowledge If decision rules are used for classiﬁcation it is easy to explain the choice Simple structure of decision rules allows several ways of visualization

These advantages lead to a rapid growth of interest in appliance of rough set theory in medical domain. Many interesting case studies reported a successful appliance of rough set software systems. Some of them were: – Treatment of duodental ulcer by HSV described by Slowinski in [21,23] – Multistage analysis in progressive encephalopathy presented by Paszek in [20,22] – Preterm birth prediction researched by Grzymala-Busse [19] – Analysis of medical databases (headaches, CVD) [24] – Acute abdominal pain in childhood (MET system applied in Children’s Hospital of Eastern Ontario) [18] – Cardiac Tests analysis [25] More successful studies, not only from medical domain, are described in [17].

86

3

G. Ilczuk and A. Wakulicz-Deja

Regular Expression

The origins of, belonging to automata and formal language theory, regular expressions lie in 1940, when McCulloch and Pitts described a nervous system as a neurons in a small automata [3]. These models were then described by Kleene and Kozen using regular expression (regular set) notation [2]. Regular expressions consist of constants and operators that denote sets of strings and operations over these sets, respectively. Given a ﬁnite alphabet Σ the following constants are deﬁned: – empty set ∅ denoting the set ∅ – empty string denoting the set {} – literal character α in Σ denoting the set {α} Following operations are deﬁned: – concatenation RS denoting the set {αβ|α in R and β in S}. For example {”ab”, ”c”}{”d”, ”ef”} = {”abd”, ”abef”, ”cd”, ”cef”}. – alternation R|S denoting the set union of R and S. – Kleene star R∗ denoting the smallest superset of R that contains and is closed under string concatenation. This is the set of all strings that can be made by concatenating zero or more strings in R. For example, {”ab”,”c”}∗ = {, ”ab”, ”c”, ”abab”, ”abc”, ”cab”, ”cc”, ”ababab”, ...}. To avoid brackets it is assumed that the Kleene star has the highest priority, then concatenation and then set union. If there is no ambiguity then brackets may be omitted. For example, (ab)c is written as abc and a|(b(c∗)) can be written as a|bc∗.

4

Methods

In our research of analyzing medical data we would like to extend and complement information collected from clinical information systems in form of ICD-10 codes with additional information stored in free-text descriptions. A typical example of a such description is shown below: Status post implantationem pacemakeri VVI (1981, 1997) ppt. diss. A-V gr. III.Exhaustio pacemakeri. Morbus ischaemicus cordis. Insufficientia coronaria chronica CCS I. Myocardiopathia ischaemica in stadio comp. circulatoriae. Fibrillatio atriorum continua. Pacemaker dependent. A method suitable for our needs should therefore fulﬁll the following requirements: – it shall recognize misspelled and abbreviated words – it shall interpretable whole sentences

Data Preparation for Data Mining in Medical Data Sets

87

– it shall provide a back tracing so that an expert can always validate an assigned mapping – all mappings must done based 100% on information from an original text – it shall be easily maintainable and extendible With these requirements in mind we have developed a method which bases on a ﬁxed number of user deﬁned records each containing following three attributes: a level value (shown at ﬁgure 1 as ’LEVEL’), a mask coded using regular expressions for searching a phrase of text (’FIND TEXT’) and a string of text which will be used for replacing if the searched phrase is found (’REPLACE FOUND TEXT’). Deﬁned records are sorted incrementally based on their level value, so that, when the algorithm starts a group of records having the lowest value can be ﬁrstly selected. During the next step for each record from the group the algorithm tries to replace a found text with a speciﬁed string. When all records are processed then a next group of records with a next higher level value is selected and the process of searching/replacing text repeats. This algorithm ends when the last record from the group with the highest level value is processed. A simpliﬁed, but based on a real implementation, example is shown at ﬁgure 1.

Fig. 1. Sample of records used by the algorithm

From this ﬁgure it can be seen that the lowest level value is 10, so that the algorithm begins to select a group of records having this level value. In our case it is only one record which replace all found Roman numbers with the following schema ’’ for example a number ’II’ will be replaced to and

88

G. Ilczuk and A. Wakulicz-Deja

’IV’ results in . After this processing a next higher level value is selected (50) together with a group of records having the same value of their level attribute. In the shown example there are only one such record. But it is possible and common that, there are a lot of records which have the same level value and thus can be processed in parallel. It is important to note that a deﬁnition of a mask used for searching (ﬁeld ’FIND TEXT’) contains not only a correct version of a phrase but also several misspelled combinations stored using the alternation operation of regular expressions for example (pectoralis|pectoris|...). If a phase is found then it will be replaced with a speciﬁed replace string independently if it was written correctly or incorrectly. This allows to correct a simple type errors and focus on sentence analysis. As an example records with their level value 100 and 110 can be used. These two records search for a combination of symbolic replacements previously replaced by records with the level value 50, so that these two records can correctly assign an ’<END I20 0> code not only to a properly written ’angina pectoris’ diagnose but also to a whole bunch of misspelled combinations of these two words as for example ’angin pectoralis’. The described algorithm has following advantages: – it allows ﬁltering of redundant and noisy information at entry processing stage – it correctly recognizes misspelled diagnoses – the process of interpreting whole sentences is simpliﬁed because only connections between symbolic phrases must be analyzed and not all possible combinations which can be found in an input text – it is possible to stop the algorithm at stage and analyze or eventually correct the replacing process – in our implementation we only use already found in input text combinations of words what decreases a possibility of false positive recognitions In the next section we will compare the recognition accuracy and scalability of the described algorithm with our previous dictionary based algorithm.

5

Dataset Preparation and Experimental Environment

Data used in our research was obtained from the Cardiology Department of Silesian Medical Academy in Katowice - the leading Electrocardiology Department in Poland specializing in hospitalization of severe heart diseases. For our experiments we took a data set of 4000 patients hospitalized in this Department between 2003 and 2005. This data were imported into a PostgreSQL database and then divided in eight groups (G-C1, ..., G-C8), where G-C1 contained ﬁrst 500 records from the database and each next group had 500 more records then the previous group, so that the last group G-C8 contained all 4000 records. Each record in a group contained a single free-text report which was then analyzed of the presence of one of the following diseases: – Essential (primary) hypertension - if found mapped to I10 code – Old myocardial infarction - if found mapped to I25-2 code

Data Preparation for Data Mining in Medical Data Sets

– – – –

89

Atrioventricular block, ﬁrst degree - if found mapped to I44-0 code Atrioventricular block, second degree - if found mapped to I44-1 code Atrioventricular block, complete - if found mapped to I44-2 code Sick sinus syndrome - if found mapped to I49-5 code

We implemented the presented algorithm in Java version 1.5. and used the Java implementation of regular expressions from the ’java.util.regex.Pattern’ class.

6

Results

Results presented the table 1 show an absolute number of cases recognized by the described in this paper method within each of the tested group. These results are additionally compared with the dictionary method and this comparison is shown as a number in brackets, where a positive number means a number of cases additionally recognized by the method based on regular expressions. Visualization of these numbers is shown at ﬁgure 2, where it can be seen that the proposed method recognized more cases then the dictionary method but with a diﬀerent, depending on a selected disease, characteristic. For hypertension and old myocardial infarction a number of additionally recognized cases is rather low what can be attributable to the fact, that the most diagnosis variants are already covered by the dictionary method. Recognition of atrioventricular block poses a bigger challenge, so that a diﬀerence in a number of recognized cases for all three types of this disease oscillates between 20-40% additional cases identiﬁed by the proposed method. The most spectacular results were achieved for recognizing Sick sinus syndrome what can be assignable with a huge number of possible combinations used to specify this diagnosis. These combinations were better covered by regular expressions and a diﬀerence to the dictionary method was almost 42%. It can be also seen, that a number of identiﬁed cases, shown at ﬁgure 3, increased for all tested diseases almost linearly. This satisfactory result shows a good ability of the presented method to recognize new records with a relatively small number of deﬁnitions (500 regular expressions compared to more then 4800 dictionary phrases). We had also randomly selected a set of 100 records and with a help from domain experts from the Cardiology Department manually identiﬁed them for three diseases. These numbers were then compared with a results achieved by both the regular expression and the dictionary method. This comparison is shown in the table 2. From this table it can be seen that for a relatively small group of records the method based on regular expression recognized all hypertension and atrioventricular block (ﬁrst degree) cases. Of course it will be only a matter of additional time eﬀort needed to extend the recognition accuracy of the dictionary method but this is exactly the advantage of the proposed algorithm, which with a signiﬁcant smaller number of records presents better scalability and in case of new data also a better update ability.

90

G. Ilczuk and A. Wakulicz-Deja

Number of records additionaly recognized by the new model

800 I10 I25-2 I44-0 I44-1 I44-2 I49-5

700

600

500

400

300

200

100

0

-100 500

1000

1500

2000

2500 3000 Number of records

3500

4000

4500

Fig. 2. Additionally recognized cases by the method based on regular expressions

3000

Number of recognized cases

2500

I10 I25-2 I44-0 I44-1 I44-2 I49-5

2000

1500

1000

500

0 500

1000

1500

2000

2500 3000 Number of records

3500

4000

Fig. 3. Number of recognized diseases by the regular expression method

4500

Data Preparation for Data Mining in Medical Data Sets

91

Table 1. Number of recognized cases by the proposed method Group Number of I10 records

I25-2

I44-0

I44-1

I44-2

I49-5

G-C1 G-C2 G-C3 G-C4 G-C5 G-C6 G-C7 G-C8

91 (-15) 190 (-13) 276 (3) 368 (25) 442 (30) 524 (44) 600 (40) 704 (43)

45 (0) 95 (12) 134 (36) 183 (61) 236 (96) 305 (120) 373 (125) 427 (131)

82 (3) 161 (31) 248 (99) 329 (158) 410 (220) 502 (258) 590 (267) 678 (276)

135 (-4) 245 (17) 354 (68) 483 (131) 604 (192) 728 (238) 863 (237) 1007 (233)

220 (72) 414 (150) 618 (241) 824 (338) 1029 (441) 1265 (550) 1493 (640) 1693 (714)

500 1000 1500 2000 2500 3000 3500 4000

325 (1) 636 (3) 978 (12) 1314 (21) 1645 (30) 1959 (38) 2275 (41) 2616 (41)

Table 2. Recognition accuracy comparison between methods

7

Disease Number of Regular exp. cases found cases

Regular exp. accuracy [%]

Dictionary Dictionary found cases

accuracy [%]

I10 I44-0 I44-1

100.0 100.0 93.3

61 9 10

100.0 81.9 66.7

61 11 15

61 11 14

Conclusions

In this paper we presented an algorithm for recognition of free-text Latin medical reports which is based on hierarchically organized records. These records use regular expressions to ﬁnd a speciﬁed phrase in an input text and replace it with a user deﬁned text. The hierarchically organized records convert an input text step by step replacing ﬁrstly simple words into symbolic phrases then these symbolic phrases into more complicated expressions and at the end the whole sentences are mapped to user deﬁned codes. Such codes can be then easily used to construct a decision table used by next data mining algorithms. Our experiments shown that the presented method achieves better recognition accuracy then the method based on ﬁxed dictionary phrases and this result can be achieved with a signiﬁcant smaller number of records used for deﬁnition. This small number of easily modiﬁable and very ﬂexible records is truly an advantage of the described method. Our idea to reduce the complexity of recognizing Latin diagnosis through deﬁning a short parts of the whole sentence using regular expressions and then to join hierarchically such pieces of information together allowed us to cover with a ﬁnite, small number of records a huge number of possible combinations. This advantage and the fact that the presented method fulﬁll all the speciﬁed

92

G. Ilczuk and A. Wakulicz-Deja

requirements it is used in our data exploration system during a preprocessing stage for processing not only Latin free-text reports but also laboratory, electrocardiogram (ECG) and cardiovascular ultrasound descriptions.

Acknowledgements We would like to thank Rafal Mlynarski from the Cardiology Department of Silesian Medical Academy in Katowice, Poland for providing us the data and giving us feedbacks.

References 1. Pyle, D.: Data preparation for data mining. Morgan Kaufmann, San Francisco (1999) 2. Kozen, D.: On Kleene Algebras and Closed Semirings. In: Mathematical Foundations of Computer Science, Bansk´ a Bystrica (1990) 26–47 3. McCulloch, W. and Pitts, W.: A logical calculus of the ideas immanent in nervous activity. In: Bulletin of Mathematical Biophysics, (1943) 115–133 4. Pawlak, Z.: Rough sets. International Journal of Computer and Information Science 11 (1982) 341–356 5. Sipser, M.: Introduction to the Theory of Computation. Course Technology, (2006) 6. Pawlak, Z.: Knowledge and Uncertainty: A Rough Set Approach. SOFTEKS Workshop on Incompleteness and Uncertainty in Information Systems (1993) 34–42 7. Pawlak, Z. and Grzymala-Busse, J. W. and Slowinski, R. and Ziarko, W.: Rough Sets. Commun. ACM 38 (1995) 88–95 8. Ilczuk, G. and Wakulicz-Deja, A.: Rough Sets Approach to Medical Diagnosis System. In: AWIC 2005, Lodz (2005) 204–210 9. Ilczuk, G. and Wakulicz-Deja, A.: Attribute Selection and Rule Generation Techniques for Medical Diagnosis Systems. In: RSFDGrC 2005, Regina (2005) 352–361 10. Wakulicz-Deja, A. and Paszek, P.: Applying Rough Set Theory to Multi Stage Medical Diagnosing. Fundam. Inform. 54 (2003) 387–408 11. Grzymala-Busse, J. W.: MLEM2 - Discretization During Rule Induction. In: IIS 2003, Zakopane (2003) 499–508 12. Ilczuk, G. and Mlynarski, R. and Wakulicz-Deja, A. and Drzewiecka, A. and Kargul, W.: Rough Sets Techniques for Medical Diagnosis Systems. In: Computers in Cardiology 2005, Lyon (2005) 837–840 13. Mlynarski, R. and Ilczuk, G. and Wakulicz-Deja, A. and Kargul, W.: Automated Decision Support and Guideline Veriﬁcation in Clinical Practice. In: Computers in Cardiology 2005, Lyon (2005) 375–378 14. Chan, C. C. and Grzymala-Busse, J. W.: On the two local inductive algorithms: PRISM and LEM2. Foundations of Computing and Decision Sciences 19 (1994) 185–203 15. Chan, C. C. and Grzymala-Busse, J. W.: On the attribute redundancy and the learning programs ID3, PRISM, and LEM2.Department of Computer Science, University of Kansas,TR-91-14, (1991) 16. Grzymala-Busse, J. W.: A new version of the rule induction system LERS. Fundam. Inform. 31 (1997) 27–39

Data Preparation for Data Mining in Medical Data Sets

93

17. Komorowski, H. J. and Pawlak, Z. and Polkowski, L. T. and Skowron, A.: Rough Sets: A Tutorial. Springer-Verlag, Singapore (1999) 18. Farion, K. and Michalowski, W. and Slowinski, R. and Wilk, S. and Rubin, S.: Rough Set Methodology in Clinical Practice: Controlled Hospital Trial of the MET System. Rough Sets and Current Trends in Computing. 3066 (2004) 805–814 19. Grzymala-Busse, J. W. and Goodwin, L. K.: Predicting pre-term birth risk using machine learning from data with missing values. Bulletin of the International Rough Set Society (IRSS). 1 (1997) 17–21 20. Paszek, P. and Wakulicz-Deja, A.: The Application of Support Diagnose in Mitochondrial Encephalomyopathies. Rough Sets and Current Trends in Computing. 2475 (2002) 586–593 21. Pawlak, Z. and Slowinski, K. and Slowinski, R.: Rough Classiﬁcation of Patients After Highly Selective Vagotomy for Duodenal Ulcer. International Journal of ManMachine Studies. 24 (1986) 413–433 22. Tsumoto, S. and Wakulicz-Deja, A. and Boryczka, M. and Paszek, P.: Discretization of continuous attributes on decision system in mitochondrial encephalomyopathies. Proceedings of the First International Conference on Rough Sets and Current Trends in Computing. 1424 (1998) 483–490 23. Slowinski, K. and Slowinski, R. and Stefanowski, J.: Rough sets approach to analysis of data from peritoneal lavage in acute pancreatitis. Medical Informatics. 13 (1988) 143–159 24. Tsumoto, S. and Tanaka, H.: Induction of Disease Description based on Rough Sets. 1st Online Workshop on Soft Computing. (1996) 19–30 25. Komorowski, H.J. and Øhrn, A.: Modelling prognostic power of cardiac tests using rough sets. Artiﬁcial Intelligence in Medicine. 15 (1999) 167–191

A Wistech Paradigm for Intelligent Systems Andrzej Jankowski1,2 and Andrzej Skowron3 1

Institute of Decision Processes Support 2 AdgaM Solutions Sp. z o.o. W¸awozowa 9 lok. 64, 02-796 Warsaw, Poland [email protected] 3 Institute of Mathematics, Warsaw University Banacha 2, 02-097 Warsaw, Poland [email protected] If controversies were to arise, there would be no more need of disputation between two philosophers than between two accountants. For it would suﬃce to take their pencils in their hands, and say to each other: ‘Let us calculate’. – Gottfried Wilhelm Leibniz, Dissertio de Arte Combinatoria (Leipzig, 1666). ... Languages are the best mirror of the human mind, and that a precise analysis of the signiﬁcation of words would tell us more than anything else about the operations of the understanding. – Gottfried Wilhelm Leibniz, New Essays on Human Understanding (1705) Translated and edited by Peter Remnant and Jonathan Bennett Cambridge: Cambridge UP, 1982 Abstract. The problem considered in this article is how does one go about discovering and designing intelligent systems. The solution to this problem is considered in the context of what is known as wisdom technology (wistech), an important computing and reasoning paradigm for intelligent systems. A rough-granular approach to wistech is proposed for developing one of its possible foundations. The proposed approach is, in a sense, the result of the evolution of computation models developed in the Rasiowa–Pawlak school. We also present a long-term program for implementation of what is known as a wisdom engine. The program is deﬁned in the framework of cooperation of many Research & Development (R & D) institutions and is based on a wistech network (WN) organization. Keywords: wisdom technology, adaptive rough-granular computing, rough sets, wisdom engine, open innovation, wisdom network. J.F. Peters et al. (Eds.): Transactions on Rough Sets VI, LNCS 4374, pp. 94–132, 2007. c Springer-Verlag Berlin Heidelberg 2007

A Wistech Paradigm for Intelligent Systems

1

95

Introduction

Huge technological changes occurred during the second half of the 20th century aﬀecting every one of us. These changes aﬀect practically all objects manufactured by man such as spoons, clothing, books, and space rockets. There are many indications that we are currently witnessing the onset of an era of radical changes. These radical changes depend on the further advancement of technology to acquire, represent, store, process, discover, communicate and learn wisdom. In this paper, we call this technology wisdom technology (or wistech, for short). The term wisdom commonly means “judging rightly” [50]. This common notion can be reﬁned. By wisdom, we understand an adaptive ability to make judgements correctly to a satisfactory degree (in particular, correct decisions) having in mind real-life constraints. One of the basic objectives of the paper is to indicate the potential directions for the design and implementation of wistech computation models. An important aspect of wistech is that the complexity and uncertainty of real-life constraints mean that in practise we must reconcile ourselves to the fact that our judgements are based on non-crisp concepts and also do not take into account all the knowledge accumulated and available to us. This is why consequences of our judgements are usually imperfect. But as a consolation, we also learn to improve the quality of our judgements via observation and analysis of our experience during interaction with the environment. Satisfactory decision-making levels can be achieved as a result of improved judgements. The intuitive nature of wisdom understood in this way can be expressed metaphorically as shown in (1). wisdom = KSN + AJ + IP,

(1)

where KSN, AJ, IP denote knowledge sources network, adaptive judgement, and interactive processes, respectively. The combination of the technologies represented in (1) oﬀers an intuitive starting point for a variety of approaches to designing and implementing computational models for wistech. In this paper, (1) is called the wisdom equation. There are many ways to build wistech computational models. In this paper, the focus is on an adaptive rough-granular approach. The issues discussed in this article are relevant for the current research directions (see, e.g., [16,15,31,38,51,90,108] and the literature cited in these articles). This paper is organized as follows.

2

Wisdom Technology

This section brieﬂy introduces the wistech paradigm. 2.1

What Do We Mean by Wistech?

On the one hand, the idea expressed by (1) (the wisdom equation paradigm) is a step in the direction of a new philosophy for the use of computing machines

96

A. Jankowski and A. Skowron

in our daily life, referred to as ubiquitous computing (see [66]). This paradigm is strongly connected with various applications of autonomic computing [64]. On the other hand, it should be emphasized that the idea of integrating many basic AI concepts (e.g., interaction, knowledge, network, adaptation, assessment, pattern recognition, learning, network, simulation of behavior in an uncertain environment, planning and problem solving) is as old as the history of AI itself. Many examples of such an approach adopted by researchers in the middle of the 20th century can be found in [27]. This research was intensively continued in the second half of the 20th century. For example, the abstracts of thousands of interesting reports from the years 1954 -1985 can be found in [91,92]. This paper contains the conclusions of the authors’ experiences during numerous practical projects implementing wistech technologies in speciﬁc applications, e.g., fraud detection (MERIX – a prototype system for Bank of America), dialogue based search engine (EXCAVIO – intelligent search engine), UAV control (WITAS project), Intelligent marketing (data mining and optimization system for Ford Motor Company, General Motors), robotics, EVOLUTIONARY CHECKERS (adaptive checker R&D program at the University of North Carolina at Charlotte) and many other applications. These experiences are summarized by the authors in the metaphoric wisdom equation (1). This equation can also be illustrated using the following diagram presented in Figure 1. In Figure 1 the term ‘data’ is understood as a stream of symbols without any interpretation of their meaning. From the perspective of the metaphor expressed in the wisdom equation (1), wistech can be perceived as the integration of three technologies (corresponding to three components in the wisdom equation (1)). At the current stage two of them seem to be conceptually relatively clear, namely 1. knowledge sources network – by knowledge we traditionally understand every organized set of information along with the inference rules; in this context one can easily imagine the following examples illustrating the concept of knowledge sources network: – representation of states of reality perceived by our senses (or observed by the “receptors” of another observer) are integrated as a whole in our minds in a network of sources of knowledge and then stored in some part of our additional memory, – a network of knowledge levels represented by agents in some multi-agent system and the level of knowledge about the environment registered by means of receptors; 2. interactive processes – interaction understood as a sequence of stimuli and reactions over time; examples are: – the dialogue of two people, – a sequence of actions and reactions between an unmanned aircraft and the environment in which the ﬂight takes place, or – a sequence of movements during some multi-player game. Far more diﬃcult conceptually seems to be the concept of adaptive judgement distinguishing wisdom from the general concept of problem solving. Intuitions behind this concept can be expressed as follows:

A Wistech Paradigm for Intelligent Systems

Wistech

Database Technology

HIERARCHY

wisdom = knowledge sources network + adaptive judgment + interactive processes knowledge = information + information relationships + inference rules

LEVELS

Information Technology

TECHNOLOGY

Knowledge Management Technology

97

information = data + interpretation

data

COMPLEXITY LEVELS OF THE SOLUTION PROBLEM SUPPORT

Perception

Prediction

Questions: about data values and additionally questions about data context, like: Who? What? When? Where? How Much?

Questions: information type questions and additionally questions about explanation and prediction, like: How? Why? What if?

Objects: Data and data explanation by a description, picture or other presentations.

Objects: Information and rules for information transformation (reasoning, ...), constrains, relationships between concepts, ideas and thought patterns.

Time context: Usually information is a posteriori, known after the fact. Measures: Logical values, uncertainty, completeness, amount.

Time context: Usually knowledge is a priori, known before the fact and provides its meaning. Measures: Efficiency of problem solutions by applying theory to information, quality of problem solutions.

Questions: knowledge type questions and additionally questions about correct judgments and decisions, action/ interaction planning / executing and justification, like: What to do? Why to do it? When to do it? How to do it? Objects: Knowledge and correct judgments, decisions based on a hierarchy of being values or believes, action plans, incorporation of vision, design, plans and implementation standards based on being preferences. Time context: Usually deals with management interactions with environment to achieve future objectives. Measures: Priorities, culture values, profits, quality of action results and plan implementation.

Understanding Questions:

Objects:

Time context:

Measures:

Fig. 1. Wisdom equation context

1. adaptive judgement – understood here as arriving at decisions resulting from the evaluation of patterns observed in sample objects. This form of judgement is made possible by mechanisms in a metalanguage (meta-reasoning) which on the basis of selection of available sources of knowledge and on the basis of understanding of history of interactive processes and their current status enable us to perform the following activities under real-life constraints: – identiﬁcation and judgement of importance (for future judgement) of sample phenomena, available for observation, in the surrounding environment;

98

A. Jankowski and A. Skowron

– planning current priorities for actions to be taken (in particular, on the basis of understanding of history of interactive processes and their current status) toward making optimal judgements; – selection of fragments of ordered knowledge (hierarchies of information and judgement strategies) satisfactory for making a decision at the planned time (a decision here is understood as a commencing interaction with the environment or as selecting the future course to make judgements); – prediction of important consequences of the planned interaction of processes; – adaptive learning and, in particular, reaching conclusions deduced from patterns observed in sample objects leading to adaptive improvement in the adaptive judgement process. One of the main barriers hindering an acceleration in the development of wistech applications lies in developing satisfactory computation models implementing the functioning of “adaptive judgement”. This diﬃculty primarily consists in overcoming the complexity of the process of integrating the local assimilation and processing of changing non-crisp and incomplete concepts necessary to make correct judgements. In other words, we are only able to model tested phenomena using local (subjective) models and interactions between them. In practical applications, usually, we are not able to give global models of analyzed phenomena (see, e.g., [110,62,64,45,25,21]). However, we can only approximate global models by integrating the various incomplete perspectives of problem perception. One of the potential computation models for “adaptive judgement” might be the rough-granular approach. 2.2

Main Diﬀerences Between Wisdom and Inference Engine

In natural language, the concept of wisdom is used in various semantic contexts. In particular, it is frequently semantically associated with such concepts as inference, reasoning, deduction, problem solving, judging rightly as a result of pattern recognition, common sense reasoning, reasoning by analogy, and others. As a consequence this semantic proximity may lead to misunderstandings. For example, one could begin to wonder what the diﬀerence is between the widely known and applied concept in AI of “inference engine” and the concept of “wisdom engine” deﬁned in this paper? In order to avoid this type of misunderstanding it is worth explaining the basic diﬀerence between the understanding of wisdom and such concepts as inference, reasoning, deduction and others. Above all, let as start with explaining how we understand the diﬀerence between problem solving and wisdom. The widespread concept of problem solving is described as some slight modiﬁcation of this notion deﬁned in the context of solving mathematical problems by George P´ olya in [84]. The concept of problem solving is understood in [84] as the following set of activities: 1. First, you have to understand the problem. 2. After understanding, then make a plan.

A Wistech Paradigm for Intelligent Systems

99

3. Carry out the plan. 4. Look back on your work. How could it be better? An attempt at explaining the concept of wisdom can be taken using the concept of problem solving in the following manner: wisdom is the ability to identify important problems, search for suﬃciently correct solutions to them, having in mind real life, available knowledge sources, personal experience, constraints, etc. Having in mind this understanding of wisdom we get at once the ﬁrst important diﬀerence. Namely, in the problem solving process we do not have the following important wisdom factor in the above sequence (1-4) of activities: 0. Learning to recognize patterns that identify important problems and problem solution constraints. Certainly, this is not the only diﬀerence. Therefore, one can illustrate the general diﬀerence between the concept of problem solving and wisdom as the diﬀerence between the concept of ﬂying in an artiﬁcially controlled environment (e.g., using a ﬂying simulator and problem solving procedures) and the concept of ﬂying Boeing 767 aeroplane in real-life dangerous environment (wisdom in a particular domain). One can therefore think that wisdom is very similar to the ability of problem solving in a particular domain of application, which in the context of the world of computing machines is frequently understood as an inference engine. The commonly accepted deﬁnition of the concept of inference engine can be found for example in Wikipedia (http://en.wikipedia.org/wiki/Inference engine). It refers to understanding of “problem solving” in the spirit of the book [84]. It reads as follows: An inference engine is a computer program that tries to derive answers from a knowledge base. It is the “brain” that expert systems use to reason about the information in the knowledge base, for the ultimate purpose of formulating new conclusions. An inference engine has three main elements. They are: 1. An interpreter. The interpreter executes the chosen agenda items by applying the corresponding base rules. 2. A scheduler. The scheduler maintains control over the agenda by estimating the eﬀects of applying inference rules in light of item priorities or other criteria on the agenda. 3. A consistency enforcer. The consistency enforcer attempts to maintain a consistent representation of the emerging solution. In other words, the concept of inference engine relates to generating strategies for the inference planning from potentially varied sources of knowledge which are in interaction together. So this concept is conceptually related to the following two elements of the wisdom equation: 1. knowledge sources network, 2. interactive processes.

100

A. Jankowski and A. Skowron

However, it should be remembered that wisdom in our understanding is not only some general concept of inference. The basic characteristic of wisdom, distinguishing this concept from the general understanding of inference, is adaptive ability to make correct judgements having in mind real-life constraints. The signiﬁcant characteristic diﬀerentiating wisdom from the general understanding of such concepts as problem solving or inference engine is adaptive judgement. In analogy to what we did in the case of problem solving, we can now attempt to explain the concept of wisdom based on the notion of an inference engine in the following manner: Wisdom is an inference engine interacting with a reallife environment, which is able to identify important problems and to ﬁnd for them suﬃciently correct solutions having in mind real-life constraints, available knowledge sources and personal experience. In this case, one can also illustrate the diﬀerence between the concept of inference engine and the concept of wisdom using the metaphor of ﬂying a plane. One could ask the question of which is the more general concept: wisdom or problem solving? Wisdom is a concept carrying a certain additional structure of adaptive judgement which in a continuously improving manner assists us in identifying the most important problem to resolve in a given set of constraints and what an acceptable compromise between the quality of the solution and the possibility of achieving a better solution is. Therefore, the question of what the more general concept is closely resembles the question from mathematics: What is the more general concept in mathematics: the concept of a ﬁeld (problem solving), or the concept of the vector space over a ﬁeld (wisdom understood as problem solving + adaptive judgement)? The vector space is a richer mathematical structure due to the action on vectors. Analogously to wisdom it is a richer process (it includes adaptive judgement - a kind of meta-judgement that encompasses recognition of patterns common to a set of sample objects that leads to judgements relating to problem solving). On the other hand, research into single-dimensional space can be treated as the research of ﬁelds. In this sense, the concept of vector space over a ﬁeld is more general than the concept of a ﬁeld. 2.3

Why Does Wistech Seem to Be One of the Most Important Future Technologies?

Nobody today doubts that technologies based on computing machines are among the most important technology groups of the 20th century, and, to a considerable degree, have been instrumental in the progress of other technologies. Analyzing the stages in the development of computing machines, one can quite clearly distinguish the following three stages in their development in the 20th century: 1. Database Technology (gathering and processing of transaction data). 2. Information Technology (understood as adding to the database technology the ability to automate analysis, processing and visualization of information). 3. Knowledge Management Technology (understood as systems supporting organization of large data sets and the automatic support for knowledge processing and discovery (see, e.g., [59,18])).

A Wistech Paradigm for Intelligent Systems

101

The three stages of development in computing machine technology show us the trends for the further development in applications of these technologies. These trends can be easily imagined using the further advancement of complexity of information processing (Shannon Dimension) and advancement of complexity of dialogue intelligence (Turing Dimension), viz., – Shannon Dimension level of information processing complexity (representation, search, use); – Turing Dimension the complexity of queries that a machine is capable of understanding and answering correctly. One of the objectives of AI is for computing machines to reach the point in Turing Dimension that is wellknown Turing Test (see [114]). In this framework, the development trends in the application of computing machines technology can be illustrated in Figure 2.

Technology

Additional

Shannon

Turing

attributes

Dimensions

Dimensions

Database

data is the

How to

Technology

most basic

represent

SQL

level

information?

Information

information =

Where to find

Who? What?

Technology

data +

information?

When? Where?

interpretation

How much?

Knowledge

knowledge =

How to use

How? Why?

Management

information +

information?

What if?

Technology

information relationships + inference rules

Fig. 2. Computing machines technology

Immediately from the beginning of the new millennium one can see more and more clearly the following new application of computing machine technology, viz., wisdom technology (wistech) which put simply can be presented in table (see Figure 3, being an extension of the table presented in Figure 2). In other words, the trends in the development of technology of computing machines can be presented using the so-called DIKW hierarchy (i.e., Data, Information, Knowledge, Wisdom). Intuitively speaking, each level of the DIKW hierarchy adds certain attributes over and above the previous one. The hierarchy is presented graphically in Figure 4.

102

A. Jankowski and A. Skowron Technology

Additional

Shannon

attributes

Dimensions

Wisdom

Wisdom equation, Learn when to

Technology

i.e. wisdom

(Wistech)

knowledge

=

use

Turing Dimensions How to make

information correct judgements (in.

sources network

Learn how

+ adaptive

to get important correct

judgment +

information

particular

decisions)

interactive

heaving in mind

processes

real-life constraints?

Fig. 3. Computing machine technology (continued)

DIKW hierarchy can be traced back to the well-known poem by T. S. Eliot, “The Rock”, written in 1932. He wrote: Where is the life we have lost in living? Where is the wisdom we have lost in knowledge? Where is the knowledge we have lost in information? It is a truism to state that the eﬀects of any activity depend to a decisive degree on the wisdom of the decisions taken, both in the start, during the implementation, improvement and completion of the activity. The main objective of wistech is to automate support for the process leading to wise actions. These activities cover all areas of man’s activities, from the economy, through medicine, education, research, development, etc. In this context, one can clearly see how important role may have the wistech development in the future. The following comment from G.W. Leibniz on the idea to automate the processing of concepts representing thoughts should not surprise us either: No one else, I believe, has noticed this, because if they had ... they would have dropped everything in order to deal with it; because there is nothing greater that man could do. 2.4

A General Approach to Wistech Computation Model

In order to create general Wistech computational models let us start an analysis of the concept of adaptive judgement. For better familiarization of adaptive judgement we shall use the visualization of processes based on the IDEFO standard. Put simply, this means of visualisation is described in the diagram presented in Figure 5. An intrinsic part of the concept of judgement is relating it to the entity implementing the judgement. Intuitively this can be a person, animal, machine, abstract agent, society of agents, etc. In general, we shall call the entity making

A Wistech Paradigm for Intelligent Systems

103

Fig. 4. DIKW hierarchy

Controls constrain and direct activities

Inputs are transformed into Output

Movements or activity result items

Controls

Activity Inputs

Outputs

Mechanisms

Physical aspects of the activity e.g people, machines, tools

Activity, process or subprocess

Fig. 5. Activity

a judgement the judge. We shall also assume that knowledge sources network is divided into external sources, i.e., sources of knowledge that are also available to other judges, internal sources, which are only available to the speciﬁc judge in question.

104

A. Jankowski and A. Skowron Evolution rules

Real Life Constrain

Reception of External interactive process Explanation and modification proposals for External Communication Language

Input wisdom communication messages from external knowledge sources network

Adaptive Judgment

External environment conditions („weather”)

Output wisdom communication messages to external knowledge sources network - in particular knowledge messages and proposals of modifications of External Communication Language (for example new communication rules, new concepts, names, ...) External actions in order to implement desired interactive processes

Implementation issues

Fig. 6. The ﬁrst level of the model

The ﬁrst level of the model is presented in Figure 6. Of course, successive levels of the model are more complex. Its details may depend on the assumed paradigms for the implementation of adaptive judgement. However, these details should include such elements as: 1. Learning of the External Communication Language understood as a language based on concepts used to communicate and process knowledge with a network of external sources of knowledge; 2. Learning of the Internal Communication Language understood as a hierarchy of meta-languages based on concepts used to process and improve External Communication Language and a language based on concepts used to communicate and process knowledge with a network of internal sources of knowledge; 3. Receiving in memory signals from signal receptors and interactive processes and expressing their signiﬁcance in the External Communication Language and the Internal Communication Language; 4. Planning the current priorities for internal actions (mainly related to the processing of wisdom) on the basis of an assessment in relation to the hierarchy of values controlling the adaptive judgement process; 5. Selection of fragments of ordered knowledge (hierarchies of information and judgement strategies) suﬃcient to take a decision at the planned time (a decision here is understood as commencing interaction with the environment or selecting the future course to resolve the problem);

A Wistech Paradigm for Intelligent Systems

105

6. Output wisdom communication messages to external knowledge sources network, in particular, knowledge messages and proposals of modiﬁcations of the External Communication Language (e.g., new communication rules, new concepts, names); 7. External actions in order to implement the desired interactive processes. All elements occurring in the above list are very complex and important but the following two problems are particularly important for adaptive judgement computational models: 1. Concept learning and integration - this is the problem of computational models for implementation of learning concepts important for the representation, processing and communicating of wisdom and, in particular, this relates to learning of concepts improving the quality of approximation of the integration of incomplete local perceptions of a problem (arising during local assimilation and processing of vague and incomplete concepts (see, e.g., [78,79])). 2. Judge hierarchy of habit habits controls - this is the problem of computational models for implementation of process of the functioning of a hierarchy of habit controls by a judge controlling the judgement process in an adaptive way. Now, we sketch the idea of a framework for solution of the problem of implementation of judge hierarchy of habit controls. In this paper, we treat a concept of habit as an elementary and repeatable part of behavioral pattern. In this context, the meaning of elementary should be considered by comparison to the required reasoning (knowledge usage) complexity necessary for the behavioral pattern implementation. In other words, by a habit we mean any regularly repeated behavioral pattern that requires little or no reasoning eﬀort (knowledge usage). In general, any behavioral pattern could be treated as a sequence of habits and other activities which use knowledge intensively. Among such activities those leading to new habits are especially important. We assume that such habit processing is controlled by so-called habit controls which support the following aspects of adaptive judgement process for a considered situation by a judge: 1. Continuous habit prioritization to be used in a particular situation after identiﬁcation of habits. This is a prioritization from the point of view of the following three criteria: – The predicted consequences of the phenomena observed in a considered situation; – Knowledge available to a judge; – The actual plans of a judge’s action. 2. Knowledge prioritization is used if we do not identify any habit to be used in a considered situation, then we have to make prioritization of pieces of available knowledge which could be used to choose the best habit or for a construction of a new habit for the considered situation. 3. Habit control assessment for continuous improvement of adaptive judgement process and for construction of new habits and habit controls.

106

A. Jankowski and A. Skowron

SIGNAL FOR ACTION

JUDGE judging in order to take action

ACTION TAKEN

Fig. 7. Judge diagram

As it can be seen from the above considerations, one of the key components of wistech, judge hierarchy of habit control, is essential for optimal decision making and is closely correlated with the knowledge held and interactions with the environment. Judge hierarchy also means the desire of the judge to satisfy his/her needs in interactions with his/her environment. Put very simply, the judge receives and sends out signals according to the diagram presented in Figure 7. The interior of the box is the place for the judge to process signals and to take an action. By the judge environment adaptation we understand the interaction of the following two adaptive processes: 1. adaptation of the environment, in which the judge lives to the needs and objectives of the judge so as to best ﬁt the needs and objectives of the environment, 2. adaptation of the internal processes taking place in a judge in such a way as to best realize his/her needs and objectives based on the resources available in the environment. The judge environment adaptation is the basis for computational models of judge learning. The key part of this is the evolution of judge hierarchy of habit controls. The judge hierarchy of habits controls constitutes a catalyst for evolutionary processes in the environment, and also constitutes an approach to expressing various paradigms of computation models to be used in the machine implementation of this concept. For example, these paradigms can be based on the metaphorically understood principle of Newtonian dynamics (e.g., action = reaction), thermodynamics (e.g., increase in entropy of information), quantum mechanics (the principle of it being impossible to determine location and speed simultaneously) and quantum computational models [44], psychology (e.g., based on metaphorical understanding of Maslow’s hierarchy of needs; see also [53,80,40]). Particularly worthy of attention in relation to wistech is the metaphoric approach to Maslow’s hierarchy of needs in reference to the abstractly understood community of agents. Put simply, this hierarchy looks as in Figure 8. It could be used for direct constructions of computational models of judge hierarchy of habit controls.

A Wistech Paradigm for Intelligent Systems

PHYSICAL

EMOTIONAL

MENTAL

SPIRITUAL

Transcendental

Self-fulfillment

Aesthetic Needs Cognitive Needs

Respect and an appreciation of one's own value Allegiance to a group and the desire for love

Security and intactness

Physiological needs

107

. Transcendental (the highest level, confirming the existence of higher being) - spiritual needs. Self-realisation (the need to have objectives, the need to satisfy one's own potential) - the needs of self-realization - are expressed in the striving of man to develop his possibilities; they constitute a means to satisfying physiological needs or reactions compensating for the failure to satisfy other needs. Aesthetic needs (the need for harmony and beauty) Cognitive needs (the need for knowledge, understanding, novelty)

Respect (the need to believe in oneself, a feeling of one's own value, competencies, authority) - the need of recognition (respect) and prestige in one's own eyes and the eyes of other people; desiring power, achievement and freedom; a need for respect and recognition from other people, good social standing, fame, dominance, attention. Allegiance (the need for ties, affiliations, love and to be loved) occur in attempts to overcome loneliness, elimination and alienation, a tendency to make close intimate relations, participation in the life of a group, in efforts to create gangs.

Security needs (defined by dependence, care and support, protection, need of comfort, peace, freedom from fear) stimulates for action, secures intactness; they reveal when the existing habits are becoming less useful. Physiological needs (lowest level needs such as food, water, oxygen, sexual needs, lack of tension) if not satisfied they dominate over all other needs, taking them on the back seat; decide on human behavior processes

Hierarchy of Human Needs A. Maslow (about 1934)

Fig. 8. The Maslow Hierarchy of human needs (about 1934) as an example of judge hierarchy of habit controls

2.5

A Rough-Granular Computing Approach to Wistech Computation Models

In this section, we outline basic ideas for the rough-granular approach to wisdom. 2.5.1 Evolution of Reasoning Computation Models in the Rasiowa–Pawlak School By the Rasiowa–Pawlak school we mean a continuation of approaches to computational models of approximate reasoning developed by Rasiowa [86], Pawlak [74,87], and their students. In some sense, it is a continuation of ideas initiated by Leibniz, Boole and currently continued in a variety of forms over the world. Of course, the Rasiowa–Pawlak school is also some kind of continuation of the Polish School of Mathematics and Logics. The achievements of this school led to the development of the modern understanding of the basic computational aspects of logic, epistemology, ontology, foundations of mathematics and natural deduction (S. Banach, S. Eilenberg, R. Ingarden, S. Ja´skowski, K. Kuratowski,

108

A. Jankowski and A. Skowron

S. Le´sniewski, A. Lindenbaum, J. L ukasiewicz, S. Mazur, A. Mostowski, H. Rasiowa, R. Sikorski, W. Sierpi´ nski, A. Tarski, S. Ulam, and many others). Two fundamental tools of the Rasiowa–Pawlak school are the following: – Computation models of a logical concept (especially of such concepts as deduction or algebraic many-valued models for classical, modal, and constructive mathematics). The Rasiowa–Pawlak approach is based on the method of treating the sets of logically equivalent statements (or formulas) as abstract algebras known as the Lindenbaum–Tarski algebras. – Computation models of vague concept. L ukasiewicz originally has proposed to treat uncertainty (or vague concepts) in logic as concepts of many-valued logic. However, software developed for today’s computers is based on two-valued Boolean algebra. Therefore it is more practical to treat uncertainty and vagueness using the classical logic concept based on two-valued Boolean algebra. The concept of a rough set introduced by Pawlak [74] and developed in the Rasiowa–Pawlak school is based on the classical two-valued logic and, hence, the rough set approach is important and suitable for the applications mentioned above. The rough set approach intended to deal with uncertainty and vagueness has been developed to deal with uncertainty and vagueness. The rough set approach makes it possible to reason precisely about approximations of vague concepts. These approximations are tentative, subjective, and varying accordingly to changes in the environment [75,76,77,8]. Both the above mentioned fundamental tools can be applied in many contexts. It is interesting to illustrate evolution of the both above fundamental tools from the Rasiowa–Pawlak school perspective (see Figure 9 and Figure 10). 2.5.2 Rough-Granular Computing (RGC) Solving complex problems by multi-agent systems in distributed environments requires approximate reasoning methods based on new computing paradigms. One such emerging recently computing paradigm is RGC. Computations in RGC are performed on information granules representing often vague, partially speciﬁed, and compound concepts delivered by agents engaged in tasks such as knowledge representation, communication with other agents, and reasoning. We discuss the rough-granular approach for modeling computations in complex adaptive systems and multiagent systems. Information granules are any objects constructed when modeling of computations, and in approximating compound concepts, and approximate reasoning about these concepts. Information granules are constructed in an optimization process based on the minimal length principle. This process is aiming at constructing approximations of concepts satisfying some (vague and/or uncertain) constraints. Examples of information granules are information systems and decision systems, elementary information granules deﬁned by indiscernibility neighborhoods, families of elementary granules (e.g., partitions and coverings),

A Wistech Paradigm for Intelligent Systems Domain & Operators

Natural Numbers Calculus

Algebra of subsets

Boolean Algebra

Logical concepts Semantical models in Lindenbaum – of constructive Tarski algebra mathematics

X is smaller than Y

X is a subset of Y

X is smaller than Y in Boolean algebra

Y can be deduced from Y

0

Zero

Empty set

The smallest element

Falsity

1

One

Full set

The biggest element

Truth

+

Addition

Join of two sets

Maximum

Disjunction

*

Multiplication

Intersection of Minimum two sets

Conjunction

Exponentiation of X to power Y

Join of (–Y) and X

Join of (–Y) and X

Implication (Y implies X)

Mod (X)

Modulo X calculus

Logical values

True False

Quotient algebra of the filter generated by set X True False

Quotient Boolean algebra of the filter generated by set X True False

Lindenbaum – Tarski algebra for a theory generated by a set of axioms X Algebra of logical values

X t+ (u |=+ deg u )} for u ∈ NInf . Granule u votes “for” granules from N + (u) (relative to t+ ) (see [46]). Let us assume N − (u) = {u ∈ D : ∃deg < t− (u |=− deg u )}

for u ∈ NInf . Then granule u votes “against” granules from N − (u) (relative to t− ). We assume that B is a distinguished set of information granules called behavioral patterns of J (e.g., decisions, actions, plans [34,115]) and Lab : D −→ B is the (partial) labeling function assigning the behavioral patterns to (some) information granules from D. • S is one of the judgemental strategies of J making it possible to select a particular behavioral pattern as a reaction to the perceived information about the environment. In particular, S uses granules from Lab(N + (u)) and Lab(N − (u)), where u = Inf (x) and x is the current state of the environment, and the labeling of these sets of granules by behavioral patterns. Observe that the strategy S should resolve conﬂicts arising due to the fact that information granules should satisfy some constraints. For example, some information granules cannot be matched by one information granule to a degree higher than a given threshold t+ . • Quality strategy of J. Q is the quality strategy of J for estimation of the closeness (similarity) between granules. The closeness estimation is

A Wistech Paradigm for Intelligent Systems

115

based on arguments “for” and “against” the satisﬁability of the compound concept of “closeness” represented by Q. In this judgement J uses relevant granules from available granules representing knowledge accessible for J, often distributed among other judges, as well as the relationships between granules represented by matching degrees. • Adaptation strategy of J. Adap is the adaptation strategy transforming a tuple − (N, G, Inf, B, Lab, |=+ deg , |=deg , S, Q) into a new such tuple. Observe that judgements performed by J during adaptation can, in particular, lead to construction of new granules (e.g., through cooperation with other judges [2]), changing some strategies such as the matching strategy, the labeling strategy, the selection strategy for relevant behavioral patterns, and the strategy for estimation of closeness of granules. Adap can also be changed, e.g., by tuning some of its parameters. 2.5.5 Basic Cycle of Judge Each judge realizes some goals using behavioral patterns. The basic cycle of each judge J is the following: 1. Step 1: Initialization. − (N, G, Inf, B, Lab, |=+ deg , |=deg , S, Q) := − (N0 , G0 , Inf0 , B0 , Lab0 , |=+ deg,0 , |=deg,0 , S0 , Q0 ). 2. Step 2: Perception granule construction by J representing the current state. u := Inf (x); where u is the granule representing perception by J of the current environment state x. 3. Step 3: J selects the relevant granules from N + (u), N − (u) and performs judgements to select (construct) the relevant behavior b toward achieving the current goal (target). During selection of b the judge J is also predicting the information Infpred (b, x) returned from EN VJ as a reaction to the behavior b applied to the current state x of EN VJ . This is realized by another special judgemental strategy of J. By applying S to Lab(N + (u)) and Lab(N − (u)) J searches for a relevant behavior b. 4. Step 4: Estimation of the closeness. The judge J uses the quality measure Q for estimation of the closeness (similarity) of Infpred (b, x) and Infreal (b, x) by Q(Infpred (b, x), Infreal (b, x)), where Infreal (b, x) is information about the real reaction of the environment in state x to the behavior b.

116

A. Jankowski and A. Skowron

5. Step 5: J uses a special judgemental strategy in testing whether the closeness is satisfactory. If the closeness is satisfactory, then J continues from Step2; otherwise J goes to the next step. 6. Step 6: Adaptation step. − (N, G, Inf, B, Lab, |=+ deg , |=deg , S, Q) := − Adapt(N, G, Inf, B, Lab, |=+ deg , |=deg , S, Q). 7. Step 7: Go to Step 2. All constructive objects involved in computations realized by means of the above judgement schemes are information granules. 2.5.6 Remark on Task Solving by Systems of Judges The above examples illustrate the complexity and richness of the information granule spaces we deal with when modeling adaptive processes and reasoning about such processes. Systems of judges solve tasks by searching in the information granule spaces for information granules satisfying the task speciﬁcation to a satisfactory degree (not necessarily exactly), i.e., matching information granules representing the task speciﬁcation to a satisfactory degree. The requirement of “matching to a degree” used instead of “matching exactly” often makes searching for solutions feasible in information granule spaces [122]. In a number of papers (see, e.g., [99,105,106]), we have developed methods for construction of information granules (satisfying a given speciﬁcation to a satisfactory degree) by means of operations on information systems called constrained sums. In particular, this approach proved to be general enough for modeling compound spatio-temporal information granules (e.g., information granules representing processes or behavioral patterns speciﬁed by vague concepts) and interactions between them.

3

Wistech Network (WN)

In this section, we discuss shortly the organization of cooperation for the projects based on wistech. 3.1

What We Mean by Wistech Network

The huge complexity of the problem of designing eﬀective wistech computation models means that wistech progress signiﬁcantly depends on forming eﬀective and systematic cooperation between the numerous interdisciplinary teams verifying the Wistech calculation models developed in practical experiments. Moreover, in order to make a really essential progress in wistech it is important to involve the best possible specialists for making it possible to combine in wistech based projects knowledge of such areas as: psychology, sociology, ethics and domain dependent knowledge, e.g., neuroscience, medicine, economics,

A Wistech Paradigm for Intelligent Systems

117

security, law, robotics, telecommunications, banking. This research, like all other research, requires a signiﬁcant eﬀort in other fundamental sciences, such as logic, epistemology, ontology, mathematics, computer science, philosophy and others. Of course such activity is very expensive. Moreover, in general, research of this type does not translate directly into economic results. No private company can aﬀord to implement such extensive research by itself. It is also unlikely that there would be any signiﬁcant commitment by government agencies in the coordination and development of research on such a wide scale. Unfortunately, current attempts at extending the international coordination of such type of research are not eﬀective. A dilemma therefore arises whether to develop wistech within the framework of expensive and highly risky closed research programs, or to support open programs in which the costs and risk are spread among many entities? It is our opinion that both directions are equally important and the key to the success is an environment for creating and developing harmony mechanisms between open and closed research (see [19]). In [19], among others, the contrasting principles of closed and open innovation are clariﬁed (see Figure 11). At the current stage of building an environment for creating and developing harmony mechanisms between open and closed research it is very important to develop a powerful framework for eﬀective Open Innovation Wistech R&D network. The current stage of development in wistech above all requires the development of coordinated interdisciplinary basic research with a well-coordinated and easily accessible environment for experiments. Such activities are not possible in hermetically sealed companies, which are paralyzed by security procedures and guided by the criterion of rapid economic return. This is also why it is proposed to start up mechanisms for the systematized and relatively coordinated cooperation of centers interested in developing Wistech under a Wistech Network (WN) cooperating with one another in accordance with jointly perfected open principles based on Open Innovation Principles. It is worth stressing that organizations preferring Closed Innovation Principles may also draw great beneﬁts from active participation in WN. This participation gives the possibility of testing solutions that have little chance of giving rapid market results, and also in the case of the appearance of such opportunities they can be translated into economic results in accordance with the principles accepted. At the same time, in the case of basic research, which in general does not translate directly into market eﬀects, the understanding of progress in basic research gives greater opportunities for developing new market applications of one’s own. A further great beneﬁt of active participation in WN should be the possibility of comparing the various paradigms for building calculation models for Wistech. The times have long since gone when people believed that there is only one perfect paradigm in AI. Hybrid solutions adapted to the speciﬁc nature of the sphere of application dominate in applications. Hybrid applications themselves also use a variety of construction paradigms in a platform for integrating various approaches. Similarly we are also assuming that the WN environment would be represented in

118

A. Jankowski and A. Skowron Contrasting Principles of Closed and Open Innovation

Closed Innovation Principles

Open Innovation Principles

The smart people in our field work

Not all smart people work for us.

for us.

We need to work with smart people inside and outside of our company.

To profit from R&D, we must

External R&D can create significant

discover it, develop it, and ship

value; internal R&D in needed to

it ourselves.

claim some portion of that value.

If we discover it ourselves, we

We don’t have to originate the

will get it to the market first.

research to profit from it.

The company that gets an innovation Building a better business model is to the market first will win.

better than getting to the market first.

If we create the most and the best

If we make the best use of internal

ideas in the industry, we will win. and external ideas, we will win. We should control our intellectual

We should profit from others’ use

properties (IP), so that our

of our IP, and we should buy

competitors don’t profit from our

others’ IP whenever it advances our

ideas.

own business model.

Fig. 11. Contrasting principles of closed and open innovation [19]

the form of a sub-network with various paradigms for the construction of an integration platform. WN would provide the data and criteria to assess the results of experiments used for the assessment of various paradigms. In the remainder of this work we present, among others, a proposal to start up a sub-network based on a paradigm for the integration of various technologies based on an adaptive rough-granular computing approach (RGC). 3.2

A Potential Example Scenario of WN Establishment

3.2.1 WN Long-Term Vision and Role The basic objectives of WN are supporting open innovation and the development of wistech and its applications through: 1. creating new paradigms and trends in Wistech and its applications, 2. creating a platform (e.g. intranet, symposia, training programs, e-learning, etc.) for communication and the exchange of knowledge and experience on the practical applications and achievements of basic research, 3. preparing educational and research programs, 4. starting up projects for speciﬁc practical applications, as well as for basic research,

A Wistech Paradigm for Intelligent Systems

119

5. establishing the conditions and criteria used to compare the quality of various approaches to Wistech (especially having in mind applications in medicine, economy, agriculture, energy and forex market), 6. popularization of Wistech. 3.2.2 WN Organization and Financial Support We assume that participation in WN is absolutely voluntary in nature, and WN itself also does not assume any additional ﬁnancial fees or obligatory participation in conferences. The organization is open in nature and any person or organization can take part in it. The form of organization is based on communities cooperating together, which jointly use and develop open software (see, e.g., http://www.opensource.org/). At the same time we assume that at some stage WN may take part in commercial projects. The project participants will mutually agree upon the principles for cooperation in every such case. It is expected that in the long-term some products or components created by WN will function according to the principles of open software (e.g. similar to the principles of http://www.opensource.org/). We continue to assume the organization of working groups in the network which would deal with jointly agreed packets of problems and projects. It is expected in our exemplary scenario that WN will develop in accordance with the stages for development of a mature organization modeled on the ideas of Carnegie Mellon Capability Maturity Model (http://www.sei.cmu.edu/cmm/). This model consists of the six stages presented in Figure 12 and Figure 13. The basic assumption to WN is the realization of projects ﬁnanced by WN participants who cover the costs and risk of their own activities in the network. It is also assumed that in WN there will be several specialist centers which will coordinate the activities in individual areas (competency centers), e.g. the multi-agent approach, the rough mereology approach. The coordination work of these centers would be ﬁnanced from voluntary ﬁnancial contributions from participants of the group in question. It follows from this that the intensity and quality of work in a given group will to a large degree depend on the level of ﬁnancial support from participants in the group.

4

Wisdom Engine

We discuss some exemplary projects proposed as pilot projects in development of wistech. 4.1

Wisdom Engine Concept

By wisdom engine we understand a machine system which implements the concept of wisdom. In other words, the basic functions of the wisdom engine would be acquiring, processing, discovering, learning and communicating wisdom. One of the main ﬁrst objectives of WN can be to create an open international R&D environment for the design and implementation of the concept of universal domainindependent wisdom engine. A universal wisdom engine implementation should

120

A. Jankowski and A. Skowron

Name of stage ESTABLISHMENT

INITIAL

Organization Starting up the first projects in the network and defining the principles for the cooperation of the first group of participants who confirm their participation in WN. Starting up the first forms of communication.

Content Developing the initial catalogue of paradigms for approaches to development of wistech (e.g., multiagents, evolution, symbolic processing, neural nets, statistics, adaptive rough granular approach, formal concepts, ontology engineering, information semiotics, cognitive and epistemological approach, etc., and their combinations). The preliminary allocation Developing a common of categorized paradigms language to describe the for approaches to wistech concepts relating to to their respective starting up, implementing and closing projects in WN. competency centers. Allocating a paradigm to a competency center, e.g. multi-agent approach, adaptive rough granular approach, etc. This does not mean that at a given competency center only and exclusively this method will be developed. On the contrary, it is assumed that every competency center will develop hybrid solutions combining various approaches. At the same time, a competency center will particularly strongly develop aspects relating to the paradigms allocated to this center.

Fig. 12. Six stages of the Carnegie Mellon Capability Maturity Model

be independent of any speciﬁc application domain. At the same time, functionality of the universal wisdom engine should enable the conﬁguration and tuning of modules for it in the form of a series of products dependent on speciﬁc application domains such as, e.g., medicine, economics, stock market, forex market, security, law, tourism, telecommunications, banking, job market. In particular universal wisdom engine should be able to learn domain knowledge by reading, discussing with experts and gathering wisdom from experience. Of course, the design and implementation of a universal wisdom engine is an extremely diﬃcult task and probably unrealistic today in a short term. First of all, we have to do some experiments with several application domains and several diﬀerent paradigms for wistech implementation. Based on an analysis of the results of such experiments we can create a more general wistech ontology which should provide a better formal framework for the implementation of a universal wisdom engine.

A Wistech Paradigm for Intelligent Systems REPEATABLE

121

Establishing the principles for

Establishing the mutually tied

selecting good practices

objectives to achieve at the

specific for the implementation

individual competency centers

of a project in wistech,

in order to verify the

designed to repeat the

effectiveness and

successes of projects realized

possibilities of developing

in similar conditions and to

various approaches.

avoid failures. Establishing the list of first conditions and criteria used to compare the quality of various approaches to wistech.

DEFINED

Putting in writing and the

Starting up the first projects

effective implementation of a

realized in the common

list of joint standards for

standards by a variety of

organization and management of

centers within the WN

projects specific to wistech, that will be binding for the WTN community. MEASURABLE

Enhancing the standards arising

Starting up mechanisms for

at the previous stage to

competitiveness between

include sets of measurable

communities working on various

indices used to verify and

approaches to wistech in the

optimize the benefits to costs

network.

of wistech projects.

CONTINUOUS IMPROVEMENT

Enhancing the standards and

Developing the optimum methods

indices defined at the

for harmonious co-operation

MEASURABLE stage to set out in

between WN and commercial

writing and effectively

companies.

implement procedures for continuously improving the functioning of WN.

Fig. 13. Six stages of the Carnegie Mellon Capability Maturity Model (continued)

Thus, it is assumed that in parallel with the work on a universal concept of a wisdom engine, work would also be conducted on utilizing the wisdom engine in selected areas of application, e.g., medicine, economics, stock market, forex market, security, law, tourism, telecommunications, banking, or job market. The long-term vision is as follows: “wisdom engineers” will receive the task to create the conﬁguration for the wisdom engine for applications in a speciﬁc ﬁeld of life, and then, after having carried out the necessary analytical and design work, to

122

A. Jankowski and A. Skowron

conﬁgure the wisdom engine and to enter the necessary initial data. The wisdom engine should have properties for self-growth and adaptation to changing conditions of its environment, as well as advances in wisdom in the ﬁelds of application. This is why one should strongly emphasize the planned property of automatic adaptation of the system – a feature not taken into account in the construction of the numerous systems in the past that were intended to perform similar tasks. A classic example here is the long-standing MYCIN project implemented by Stanford University. The implementation of the idea expressed by the wisdom equation is very diﬃcult and it would be unreasonable to expect its full implementation in a short period of time. We assume that the creativity cycle for the ﬁrst product prototypes implementing this concept would take several years of intensive work with cooperation of product managers, scientists, engineers, programmers and domain experts. On the other hand, it is not desirable to implement such long projects without any clear interim eﬀects. This is why we assume that the wisdom engine implementation project would go through several phases. For example, initially we assume they will go through ﬁve phases in the implementation of the wisdom engine. We propose a route, to achieving the target wisdom engine products through continuously improving intermediary products that meet successive expansions in functionality. The ﬁve phases are called as follows: 1. 2. 3. 4. 5.

Summary, Spider, Conceptual Clustering and Integration, Wisdom Extraction, Wisdom Assistant.

The eﬀect of each of these phases will be a prototype product that after acceptance would be interesting for the WN community. Stated in simple terms the functional eﬀects of the individual phases would be as presented in Figure 14. 4.2

Examples of Wisdom Engine Domain-Dependent Product Lines

The above ﬁve phases (i.e., Summary, Spider, Conceptual Clustering and Integration, Wisdom Extraction, and Wisdom Assistant) should be applied to several directions for potential product lines which would be developed in the WN. Of course, it can theoretically be any product relating to applications in robotics, unmanned aircraft, space rockets, etc. However, if we wish to have as many people as possible cooperating in the WN, then the product lines must be chosen so that experimenting with them does not prove expensive. On the other hand, these product lines must be suﬃciently attractive so as to interest as many people as possible. We propose that these product lines relate to applications in such areas as medicine, economics, the stock market, forex market, security, law, tourism, telecommunications, banking, job market and others. The list of products that could be expanded in accordance with the above scheme is potentially unlimited. The proposals for the descriptions of speciﬁc

A Wistech Paradigm for Intelligent Systems Phase

Key functions

Summary

Spider

Conceptual Clustering and Integration key document document concept searching clustering extraction related to based on key key concept concept and and indexing Integration documents

123

Adaptive Wisdom Extraction

Adaptive Wisdom Assistant

extracting data structures, information structures, knowledge structures and adaptive wisdom structures from documents, in particular generating thesauruses, conceptual hierarchies and constraints to acceptable solutions

user query / answering processing in order to support users in solving their problems as effectively as possible

Fig. 14. Functional eﬀects of the individual phases

Summary

Spider

Document Manager

Phase/Product

Document Summary

Document Spider

Conceptual Clustering Document Conceptual Clustering

Job Market

Job Market Summary

Job Market Spider

Job Market Conceptual Clustering

Brand Monitoring

Brand Monitoring Summary

Brand Monitoring Spider

Brand Monitoring Conceptual Clustering

World Communication

World Communication Summary

World World Communication Communication Spider Conceptual Clustering

World Forex

World Forex Summary

World Forex Spider

World Stock Market

World Stock Market Summary

World Stock World Stock Market Spider Market Conceptual Clustering

World Tourist

World Tourist Summary

World Tourist World Tourist Spider Conceptual Clustering

Physician

Physician Summary

Physician Spider

Lawyer

Lawyer Summary

Lawyer Spider Lawyer Conceptual Clustering

Economy Monitoring

Economy Monitoring Summary

Economy Monitoring Spider

World Forex Conceptual Clustering

Physician Conceptual Clustering

Economy Monitoring Conceptual Clustering

Wisdom Extraction Document adaptive wisdom Extraction Job Market adaptive wisdom Extraction Brand Monitoring adaptive wisdom Extraction World Communication adaptive wisdom Extraction World Forex adaptive wisdom Extraction World Stock Market adaptive wisdom Extraction World Tourist adaptive wisdom Extraction Physician adaptive wisdom Extraction Lawyer adaptive wisdom Extraction Economy Monitoring adaptive wisdom Extraction

Wisdom Assistant Document adaptive wisdom Assistant Job Market adaptive wisdom Assistant Brand Monitoring adaptive wisdom Assistant World Communication adaptive wisdom Assistant World Forex adaptive wisdom Assistant World Stock Market adaptive wisdom Assistant World Tourist adaptive wisdom Assistant Physician adaptive wisdom Assistant Lawyer adaptive wisdom Assistant Economy Monitoring adaptive wisdom Assistant

Fig. 15. Proposed products

products, included in the later part of this report, should be treated as ﬂexible and primarily constitute material for discussion, and not a ﬁnal decision. On the other hand, the list of products described is not entirely accidental in nature.

A. Jankowski and A. Skowron

Job Market

Document Manager

Product / Phase

124

Summary

Spider

Conceptual Clustering

Adaptive Wisdom Adaptive Wisdom Extraction Assistant

automatic summarizing of a document and groups of documents, the contents of which are not connected with any specific field

automatic searching and downloading of any documents

conceptual clustering of documents on any subject

automatic summarizing of a document and groups of documents relating to job market, carried out from the perspective of the following groups of users: potential employers and potential employees

automatic searching and downloading of documents relating to job market, carried out with particular emphasis on the needs of the following groups of users: potential employers and potential employees

conceptual clustering of documents relating to job market with particular emphasis on the specific nature of queries submitted by the following types of users: potential employers and potential employees

extracting data structures, information structures, knowledge structures and adaptive wisdom structures from documents, in particular generating thesauruses, conceptual hierarchies and constraints to acceptable solutions in any domain extracting data structures, information structures, knowledge structures and adaptive wisdom structures from documents, in particular generating thesauruses, conceptual hierarchies and constraints to acceptable solutions in job market domain

general user query / answering processing in order to support users in solving their problems as effectively as possible

job market related to user query / answering processing in order to support users in solving their problems as effectively as possible

Fig. 16. Functionality of individual products

This is because they form a certain logical continuity, connected both with the degree of diﬃculty in successive products and current preferences resulting from the previous experiences of the human resources that would be engaged to carry out the work on individual products. The initial selection of product lines is as follows: – – – – – – – – – –

Document Manager, Job Market, Brand Monitoring, World Communication, World Forex, World Stock Market, World Tourist, Physician, Lawyer, Economy Monitoring.

World Communication

Brand Monitoring

Product / Phase

A Wistech Paradigm for Intelligent Systems

125

Summary

Spider

Conceptual Clustering

Adaptive Wisdom Adaptive Wisdom Extraction Assistant

automatic summarizing of a document and groups of documents relating to brand, carried out from the perspective of the following groups of users: brand owners, detectives looking for frauds, buyers

automatic searching and downloading of documents relating to brand, carried out with particular emphasis on the needs of the following groups of users: brand owners, detectives looking for frauds, buyers

conceptual clustering of documents relating to brand with particular emphasis on the specific nature of queries submitted by the following types of users: brand owners, detectives looking for frauds, buyers

extracting data structures, information structures, knowledge structures and adaptive wisdom structures from documents, in particular generating thesauruses, conceptual hierarchies and constraints to acceptable solutions in brand monitoring domain

brand monitoring related to user query / answering processing in order to support users in solving their problems as effectively as possible

automatic summarizing of a document and groups of documents relating to communication , carried out from the perspective of people looking for optimal connections

automatic searching and downloading of documents relating to communication carried out with particular emphasis on the needs of the following groups of users: people looking for optimal connections

conceptual clustering of documents relating to communication with particular emphasis on the specific nature of queries submitted by the following types of users: people looking for optimal connections

extracting data structures, information structures, knowledge structures and adaptive wisdom structures from documents, in particular generating thesauruses, conceptual hierarchies and constraints to acceptable solutions in communication domain

communication related to user query / answering processing in order to support users in solving their problems as effectively as possible

Fig. 17. Functionality of individual products (continued)

This initial selection for the product list generates several dozen of products that would be the eﬀect of work on the individual phases of implementing each of the products, i.e., Summary, Spider, Conceptual Clustering, Wisdom Extraction, Wisdom Assistant. We present the proposed products in Figure 15. The scope of the program described in this paper should be considered as dynamic and more as a basis for further discussion than a ﬁnal version of the speciﬁc deﬁnitions of the projects. This is why the innovative ideas presented and the vision for their implementation do not contain any detailed cost beneﬁts analysis. It will only be possible to specify revenues, costs and cash ﬂow forecasts with any accuracy after the planned scope of work and the role of the WN has stabilized. As there are as yet no ﬁnal decisions on the scope of operations or role of the WN, this means that at the current stage it is impossible to

126

A. Jankowski and A. Skowron

precisely estimate the planned requirements for human resources. This is why in this document we present only the general human resources requirements and a description of the general mechanisms for acquiring these resources to implement WN. Each of these products would have their own individual functionality which would result from adapting the wisdom engine to the speciﬁc characteristics of their specialist ﬁelds. Figure 16 and Figure 17 show the functionality of the individual products.

5

Conclusions

We have discussed the main features of wistech and its importance for further progress in the development of intelligent systems. The proposed approach is based on Rough Granular Computing (RGC). One of the central problems of science today is to develop methods for approximation of compound vague concepts and approximate reasoning about them [32,81]. Today, we do not have yet satisfactory tools for discovery of relevant patterns for approximation of compound concepts directly from sample objects. However, we have developed methods for compound concept approximation using sample objects and domain knowledge acquired from experts (this is the approach pioneered by Zdzislaw Pawlak in [73]). The performed experiments based on approximation of concept ontology (see, e.g., [3,5,6,7,8,22,68,69,70], [78,79,93,94,95,98,99], [100,101,102,105,106]) showed that domain knowledge enables to discover relevant patterns in sample objects for compound concept approximation. Our approach to compound concept approximation and approximate reasoning about compound concepts is based on the rough-granular approach. One of the RGC challenges is to develop approximate reasoning techniques for reasoning about dynamics of distributed systems of judges. These techniques should be based on systems of evolving local perception logics rather than on a global logic [94,95]. Approximate reasoning about global behavior of judges’ system is infeasible without methods for approximation of compound vague concepts and approximate reasoning about them. One can observe here an analogy to phenomena related to the emergent patters in complex adaptive systems [21]. Let us observe that judges can be organized into a hierarchical structure, i.e., one judge can represent a coalition of judges in interaction with other agents existing in the environment [2,56,62]. Such judges representing coalitions play an important role in hierarchical reasoning about behavior of judges’ populations. Strategies for coalition formation and cooperation [2,62,64] are of critical importance in designing systems of judges with dynamics satisfying to a satisfactory degree a given speciﬁcation. Developing strategies for discovery of information granules representing relevant (for the given speciﬁcation) coalitions and cooperation protocols is another challenge for RGC.

A Wistech Paradigm for Intelligent Systems

127

RGC will become more and more important for analysis and synthesis of the discussed compound adaptive processes.The impact of RGC on real-life applications will be determined by techniques based on the rough-granular approach to modeling of relevant computations on compound information granules and methods for approximate reasoning about complex adaptive processes over such information granules. RGC techniques for modeling of complex processes will also have impact on the development of new non-conventional computation models.

Acknowledgments The research of Andrzej Jankowski was supported by Institute of Decision Process Support. The research of Andrzej Skowron has been supported by the grant 3 T11C 002 26 from Ministry of Scientiﬁc Research and Information Technology of the Republic of Poland. Many thanks to Professors James Peters and Anna Gomoli´ nska for their incisive comments and for suggesting many helpful ways to improve this article.

References 1. G. Antoniou, F. van Harmelen, A Semantic Web Primer (Cooperative Information Systems) The MIT Press, 2004. 2. R. Axelrod, The Complexity of Cooperation, Princeton, NJ: Princeton University Press, 1997. 3. A. Bargiela, W. Pedrycz, Granular Computing: An Introduction, Dordrecht: Kluwer Academic Publishers, 2003. 4. J. Barwise, J. Seligman, Information Flow: The Logic of Distributed Systems, Cambridge University Press, 1997. 5. J. G. Bazan, J. F. Peters, and A. Skowron, Behavioral pattern identiﬁcation through rough set modelling, in [71], pp. 688–697. 6. J. Bazan, A. Skowron, On-line elimination of non-relevant parts of complex objects in behavioral pattern identiﬁcation, in [111], pp. 720–725. 7. J. Bazan, A. Skowron, Classiﬁers based on approximate reasoning schemes, in Monitoring, Security, and Rescue Tasks in Multiagent Systems (MSRAS 2004), B. Dunin-Keplicz, A. Jankowski, A. Skowron, and M. Szczuka, Eds., Advances in Soft Computing, pp. 191-202, Heidelberg: Springer, 2005. 8. J. Bazan, A. Skowron, R. Swiniarski, Rough sets and vague concept approximation: From sample approximation to adaptive learning, Transactions on Rough Sets V: Journal Subline, Lecture Notes in Computer Science, vol. 3100, pp. 39-63, Heidelberg: Springer, 2006. 9. R. Baeza-Yates, B. Ribeiro-Neto, Modern Information Retrieval, Addison Wesley, 1999. 10. M. W. Berry, Survey of Text Mining : Clustering, Classiﬁcation, and Retrieval, Springer, 2003. 11. R. Brachman, H. Levesque, Knowledge Representation and Reasoning, Morgan Kaufmann, 2004. 12. L. Breiman, Statistical modeling: The two Cultures, Statistical Science 16(3) (2001) 199–231.

128

A. Jankowski and A. Skowron

13. S. Brin, L. Page, The Anatomy of a Large-Scale Hypertextual Web Search Engine, Stanford University, 1998. 14. C. Carpineto, G. Romano, Concept Data Analysis: Theory and Applications, John Wiley & Sons, 2004. 15. N. L. Cassimatis, A cognitive substrate for achieving human-level intelligence, AI Magazine 27(2) (2006) 45-56. 16. N. L. Cassimatis, E. T. Mueller, P. H. Winston, Achieving human-level intelligence through integrated systems and research, AI Magazine 27(2) (2006) 12-14. 17. S. Chakrabarti, Mining the Web: Analysis of Hypertext and Semi Structured Data, The Morgan Kaufmann Series in Data Management Systems, Morgan Kaufmann, 2002. 18. Mu-Y. Chen, An-P. Chen, Knowledge management performance evaluation: a decade review from 1995 to 2004, Journal of Information Science 32 (1) 2006 17-38. 19. H. W. Chesbrough, Open Innovation: The New Imperative for Creating and Profiting from Technology, Cambridge MA: Harvard Business School Publishing, 2003. 20. J. Coleman, Introducing Speech and Language Processing (Cambridge Introductions to Language and Linguistics), Cambridge University Press, 2005. 21. A. Desai, Adaptive complex enterprises, Communications ACM 48(5) (2005) 3235. 22. P. Doherty, W. L ukaszewicz, A. Skowron, and A. Szalas, Knowledge Engineering: A Rough Set Approach, Studies in Fuzziness and Soft Computing, vol. 202, Heidelberg: Springer, 2006. 23. R. Dornfest, T. Calishain, Wistech Network Hacks O’Reilly Media, Inc.; 2004. 24. R. Duda, P. Hart, and R. Stork, Pattern Classiﬁcation, New York, NY: John Wiley & Sons, 2002. 25. B. Dunin-K¸eplicz, A. Jankowski, A. Skowron, M. Szczuka, Monitoring, Security, and Rescue Tasks in Multiagent Systems (MSRAS’2004), Series in Soft Computing, Heidelberg: Springer, 2005. 26. A. E. Eiben, J. E. Smith, Introduction to Evolutionary Computing, Natural Computing Series, Springer, 2003. 27. E. Feigenbaum, J. Feldman (Eds.), Computers and Thought, New York: McGraw Hill, 1963. 28. S. Feldman, Why Search is Not Enough (white paper), IDC, 2003. 29. S. Feldman, Enterprise Search Technology: Information Disasters and the High Cost of Not Finding Information (Special IDC Report), IDC, 2004. 30. D. Fensel, Ontologies: A Silver Bullet for Knowledge Management and Electronic Commerce, Springer, 2003. 31. K. D. Forbus, T. R. Hinrisch, Companion cognitive systems: A step toward human-level AI, AI Magazine 27(2) (2006) 83-95. 32. M. Gell-Mann, The Quark and the Jaguar, NY: Freeman and Co., 1994. 33. J. H. Friedman, T. Hastie, R. Tibshirani, The Elements of Statistical Learning: Data Mining, Inference, and Prediction, Heidelberg: Springer, 2001. 34. M. Ghallab, D. Nau, and P. Traverso, Automated Planning: Theory and Practice, CA: Morgan Kaufmann, 2004. 35. S. Ghemawat, H. Gobioﬀ, Shun-Tak Leung, The Wistech Network File System, Wistech Network, 2005. 36. D. E. Goldberg, Genetic Algorithms in Search, Optimization, and Machine Learning, Addison-Wesley Professional, 1989.

A Wistech Paradigm for Intelligent Systems

129

37. A. Gomez-Perez, O. Corcho, M. Fernandez-Lopez, Ontological Engineering with examples from the areas of Knowledge Management, e-Commerce and the Semantic Web (Advanced Information and Knowledge Processing), Springer, 2004. 38. R. Granger, Engines of the brain: The computational instruction set of human cognition, AI Magazine 27(2) (2006) 15-31. 39. S. Grimes, The Developing Text Mining Market, A white paper prepared for Text Mining Summit 2005, Boston, June 7-8 2005, Alta Plana Corporation, 2005. 40. A. OHagan, C. E. Buck, A. Daneshkhah, J. R. Eiser, P. H. Garthwaite, D. J. Jenkinson, J. E. Oakley, T. Rakow, Uncertain Judgements: Eliciting Expert Probabilities, Wiley, New York, 2006. 41. J. Heaton, Programming Spiders, Bots, and Aggregators in Java, Sybex, 2002. 42. K. Hemenway, T. Calishain, Spidering Hacks O’Reilly Media, Inc.; 2003. 43. M. Henzinger, S. Lawrence, Extracting knowledge from the World Wide Web, Wistech Network, 2004. 44. M. Hirvensalo, Quantum Computing, Springer-Verlag, Heidelberg 2001. 45. M. N. Huhns, M. P. Singh, Readings in Agents, Morgan Kaufmann, 1998. 46. P. Jackson, I. Moulinier, Natural Language Processing for Online Applications: Text Retrieval, Extraction, and Categorization (Natural Language Processing, 5), John Benjamins Publishing Co, 2002. 47. Z. Janiszewski, On needs of mathematics in Poland (O potrzebach matematyki w Polsce) (in Polish), In: Nauka Polska. Jej Potrzeby, Organizacja i Rozw´ oj, Warszawa, 1918; see also reprint in Wiadomo´sci Matematyczne VII (1963) 3-8. 48. A. Jankowski, An alternative characterization of elementary logic, Bull. Acad. Pol. Sci., Ser. Math. Astr. Phys.XXX (1-2) (1982) 9-13. 49. A. Jankowski, Galois structures, Studia Logica 44(2) (1985) 109-124. 50. S. Johnson, Dictionary of the English Language in Which the Words are Deduced from Their Originals, and Illustrated in their Diﬀerent Signiﬁcations by Examples from the Best Writers, 2 Volumes. London: F.C. and J. Rivington, 1816. 51. R. M. Jones, R. E. Wray, Comparative analysis of frameworks for knowledgeintensive intelligent agents, AI Magazine 27(2) (2006) 57-70. 52. D. Jurafsky, J. H. Martin, Speech and Language Processing: An Introduction to Natural Language Processing, Computational Linguistics and Speech Recognition, Prentice Hall, 2000. 53. D. Kahneman, P. Slovic, A. Tversky, A. (Eds.). Judgement under Uncertainty: Heuristics and Biases, Cambridge University Press, New York, 1982. 54. R. Keefe, Theories of Vagueness, Cambridge, UK: Cambridge Studies in Philosophy, 2000. ˙ 55. W. Kloesgen, J. Zytkow, Handbook of Knowledge Discovery and Data Mining, New York: Oxford University Press, 2002. 56. S. Kraus, Strategic Negotiations in Multiagent Environments, Massachusetts: The MIT Press, 2001. 57. J. Lambek, P. J. Scott, Introduction to Higher-Order Categorical Logic (Cambridge Studies in Advanced Mathematics 7), Cambridge University Press, 1986. 58. M. Lamb, Build Your Own Army of Web Bots Within 24 Hours (Army of Web Bots Series) Authorhouse, 2003. 59. P. Langley, H. A. Simon, G. L. Bradshaw, J. M. Zytkow, Scientiﬁc Discovery: Computational Explorations of the Creative Processes, MIT Press, 1987. 60. P. Langley, Cognitive architectures and general intelligent systems, AI Magazine 27(2) (2006) 33-44. 61. G. W. Leibniz, New Essays on Human Understanding, Cambridge UP, 1982.

130

A. Jankowski and A. Skowron

62. J. Liu, Autonomous Agents and Multi-Agent Systems: Explorations in Learning, Self-Organization and Adaptive Computation, Singapore: World Scientiﬁc Publishing, 2001. 63. J. Liu, L. K. Daneshmend, Spatial Reasoning and Planning: Geometry, Mechanism, and Motion, Springer, 2003, Hardcover. 64. J. Liu, X. Jin, K. Ch. Tsui, Autonomy Oriented Computing: From Problem Solving to Complex Systems Modeling, Heidelberg: Kluwer Academic Publisher/Springer, 2005. 65. S. MacLane, I. Moerdijk, Sheaves in Geometry and Logic: A First Introduction to Topos Theory (Universitext), Springer, 1994. 66. A. Madhavapeddy, N. Ludlam, Ubiquitious Computing needs to catch up with Ubiquitous Media, University of Cambridge Computer Laboratory, Interceptor Communications Ltd., 2005. 67. C. D. Manning, H. Sch¨ utze, H., Foundations of Statistical Natural Language Processing, The MIT Press, 1999. 68. S. H. Nguyen, J. Bazan, A. Skowron, and H. S. Nguyen, Layered learning for concept synthesis, Transactions on Rough Sets I: Journal Subline, Lecture Notes in Computer Science, vol. 3100, pp. 187-208, Heidelberg: Springer, 2004. 69. T. T. Nguyen, Eliciting domain knowledge in handwritten digit recognition, in [71] 762-767. 70. S. K. Pal, L. Polkowski, and A. Skowron (Eds.), Rough-Neural Computing: Techniques for Computing with Words, Cognitive Technologies, Heidelberg: SpringerVerlag, 2004. 71. S. K. Pal, S. Bandoyopadhay, and S. Biswas (Eds.), Proceedings of the First International Conference on Pattern Recognition and Machine Intelligence (PReMI’05), December 18-22, 2005, Indian Statistical Institute, Kolkata, Lecture Notes in Computer Science vol. 3776, Heidelberg: Springer, 2005. 72. T. B. Passin, Explorer’s Guide to the Semantic Web Mining Publications, 2004. 73. Pawlak, Z.: Classiﬁcation of objects by means of attributes, Research Report PAS 429, Institute of Computer Science, Polish Academy of Sciences, ISSN 138-0648, January (1981). 74. Z. Pawlak, Rough Sets: Theoretical Aspects of Reasoning about Data, System Theory, Knowledge Engineering and Problem Solving 9, Dordrecht: Kluwer Academic Publishers, 1991. 75. Z. Pawlak, A. Skowron, Rudiments of rough sets. Information Sciences. An International Journal. 177(1) (2007) 3-27. 76. Z. Pawlak, A. Skowron, Rough sets: Some extensions. Information Sciences. An International Journal. 177(1) (2007) 28-40. 77. Z. Pawlak, A. Skowron, Rough sets and Boolean reasoning. Information Sciences. An International Journal. 177(1) (2007) 41-73. 78. J. F.Peters, Rough ethology: Toward a biologically-inspired study of collective behavior in intelligent systems with approximation spaces. Transactions on Rough Sets III: Journal Subline, Lecture Notes in Computer Science, vol. 3400, pp. 153174, Heidelberg: Springer, 2005. 79. J. F. Peters, C. Henry, C., Reinforcement learning with approximation spaces. Fundamenta Informaticae 71(2-3) (2006) 323-349. 80. S. Plous, The Psychology of Judgement and Decision Making, McGraw-Hill, New York, 1993. 81. T. Poggio, S. Smale, The mathematics of learning: Dealing with data, Notices of the AMS 50(5) (2003) 537-544.

A Wistech Paradigm for Intelligent Systems

131

82. L. Polkowski, A. Skowron, Rough mereology: A new paradigm for approximate reasoning, International Journal of Approximate Reasoning 15 (1996) 333-365. 83. L. Polkowski, S. Tsumoto, T. Y. Lin (Eds.), Rough Set Methods and Applications: New Developments in Knowledge Discovery in Information Systems, Studies in Fuzziness and Soft Computing vol. 56, Physica-Verlag Heidelberg, 2000. 84. G. P´ olya, How to Solve It, 2nd ed., Princeton University Press, 1957; see also http://en.wikipedia.org/wiki/How to Solve It. 85. H. Rasiowa, W. Marek, On reaching consensus by groups of intelligent agents, In: Z. W. Ras (Ed.), Methodologies for Intelligent Systems, North-Holland, Amsterdam, 1989, 234-243. 86. H. Rasiowa, Algebraic Models of Logics, Warsaw University, 2001. 87. H. Rasiowa, R. Sikorski, The Mathematics of Metamathematics, Monograﬁe Matematyczne vol. 41, PWN Warsaw, 1963. 88. S. Read, Thinking about Logic. An Introduction to the Philosophy of Logic, Oxford, New York: Oxford University Press, 1995. 89. P. Saint-Dizier, E. Viegas, B. Boguraev, S. Bird, D. Hindle, M. Kay, D. McDonald, H. Uszkoreit, Y. Wilks, Computational Lexical Semantics (Studies in Natural Language Processing), Cambridge University Press, 2005. 90. C. Schlenoﬀ, J. Albus, E. Messina, A. J. Barbera, R. Madhavan, S. Balakirsky, Using 4D/RCS to address AI knowledge integration, AI Magazine 27(2) (2006) 71-81. 91. Scientiﬁc Datalink, The Scientiﬁc DataLink index to artiﬁcial intelligence research, 1954-1984, Scientiﬁc DataLink, 1985. 92. Scientiﬁc Datalink, The Scientiﬁc DataLink index to artiﬁcial intelligence research, 1985 Supplement, Scientiﬁc DataLink, 1985. 93. A. Skowron, Approximate reasoning in distributed environments, in N. Zhong, J. Liu (Eds.), Intelligent Technologies for Information Analysis, Heidelberg: Springer, pp. 433-474. 94. A. Skowron, Perception logic in intelligent systems (keynote talk), In: S. Blair et al (Eds.), Proceedings of the 8th Joint Conference on Information Sciences (JCIS 2005), July 21-26, 2005, Salt Lake City, Utah, USA, X-CD Technologies: A Conference & Management Company, ISBN 0-9707890-3-3, 15 Coldwater Road, Toronto, Ontario, M3B 1Y8, 2005, pp. 1-5. 95. A. Skowron, Rough sets in perception-based computing (keynote talk), in [71], pp. 21-29. 96. A. Skowron, R. Agrawal, M. Luck, T. Yamaguchi, O. Morizet-Mahoudeaux, J. Liu, N. Zhong (Eds.), Proceedings of the 2005 IEEE/WIC/ACM International Conference on WEB Intelligence, Compiegne, France, September 19-22, 2005, IEEE Computer Society Press, Los Alamitos, CA, 2005, pp. 1-819. 97. A. Skowron, J.-P. Barthes, L. Jain, R. Sun, P. Morizet-Mahoudeaux, J. Liu, N. Zhong (Eds.), Proceedings of the 2005 IEEE/WIC/ACM International Conference on Intelligent Agent Technology, Compiegne, France, September 19-22, 2005, IEEE Computer Society Press, Los Alamitos, CA, 2005, pp. 1-766. 98. A. Skowron, J. Stepaniuk, Tolerance approximation spaces, Fundamenta Informaticae 27 (1996) 245-253. 99. A. Skowron, J. Stepaniuk, Information granules: Towards foundations of granular computing, International Journal of Intelligent Systems 16(1) (2001) 57-86. 100. A. Skowron, J. Stepaniuk, Information granules and rough-neural computing, in [70], pp. 43-84. 101. A. Skowron, P. Synak, Complex patterns, Fundamenta Informaticae 60(1-4) (2004) 351-366.

132

A. Jankowski and A. Skowron

102. A. Skowron, R. Swiniarski, and P. Synak, Approximation spaces and information granulation, Transactions on Rough Sets III: Journal Subline, Lecture Notes in Computer Science, vol. 3400, pp. 175-189, Heidelberg: Springer, 2005. 103. J. F. Sowa, Knowledge Representation: Logical, Philosophical, and Computational Foundations, Course Technology, 1999. 104. S. Staab, R. Studer, Handbook on Ontologies, in International Handbooks on Information Systems, Heidelberg: Springer 2004. 105. J. Stepaniuk, J. Bazan, and A. Skowron, Modelling complex patterns by information systems, Fundamenta Informaticae 67 (1-3) (2005) 203-217. 106. J. Stepaniuk, A. Skowron, J. Peters, and R. Swiniarski, Calculi of approximation spaces, Fundamenta Informaticae 72 (1-3) (2006) 363-378. 107. P. Stone, Layered Learning in Multi-Agent Systems: A Winning Approach to Robotic Soccer, Cambridge, MA: The MIT Press, 2000. 108. W. Swartout, J. Gratch, R. W. Hill, E. Hovy, S. Marsella, J. Rickel, D. Traum, Towards virtual humans, AI Magazine 27(2) (2006) 96-108. 109. R. Sun (Ed.), Cognition and Multi-Agent Interaction. From Cognitive Modeling to Social Simulation. New York, NY: Cambridge University Press, 2006. 110. K. Sycara, Multiagent systems, in AI Magazine, Summer 1998, 79-92. ´ ezak, J. T. Yao, J. F. Peters, W. Ziarko, and X. Hu (Eds.), Proceedings of 111. D. Sl¸ the 10th International Conference on Rough Sets, Fuzzy Sets, Data Mining, and Granular Computing (RSFDGrC’2005), Regina, Canada, August 31-September 3, 2005, Part II, Lecture Notes in Artiﬁcial Intelligence, vol. 3642, Heidelberg: Springer, 2005. 112. A. S. Troelstra, H. Schwichtenberg, Basic Proof Theory, Cambridge University Press, 2000. 113. A. S. Troelstra, D. Van Dalen, Constructivism in Mathematics: An Introduction, Studies in Logic and the Foundations of Mathematics Vol. 1 & 2, Elsevier Science Publishing Company, 1988. 114. A. Turing, Computing machinery and intelligence, Mind LIX(236) (October 1950) 433-460. 115. W. Van Wezel, R. Jorna, and A. Meystel, Planning in Intelligent Systems: Aspects, Motivations, and Methods. Hoboken, New Jersey: John Wiley & Sons, 2006. 116. V. Vapnik, Statistical Learning Theory, New York: John Wiley & Sons, 1998. 117. S. Weiss, N. Indurkhya, T. Zhang, F. Damerau, Text Mining: Predictive Methods for Analyzing Unstructured Information, Springer, 2004. 118. I. H. Witten, E. Frank, Data Mining: Practical Machine Learning Tools and Techniques, Second Edition, Morgan Kaufmann Series in Data Management Systems, Morgan Kaufmann, 2005. 119. I. H. Witten, A. Moﬀat, T. C. Bell, Managing Gigabytes: Compressing and Indexing Documents and Images, The Morgan Kaufmann Series in Multimedia and Information Systems, Morgan Kaufmann, 1999. 120. L. A. Zadeh, Fuzzy sets, Information and Control 8 (1965) 333-353. 121. L. A. Zadeh, From computing with numbers to computing with words - from manipulation of measurements to manipulation of perceptions, IEEE Transactions on Circuits and Systems - I: Fundamental Theory and Applications 45(1) (1999) 105-119. 122. L. A. Zadeh, A new direction in AI: Toward a computational theory of perceptions, AI Magazine 22 (1) (2001) 73-84. 123. W. Ziarko, Variable precision rough set model, Journal of Computer and System Sciences 46 (1993) 39-59.

The Domain of Acoustics Seen from the Rough Sets Perspective Bozena Kostek Multimedia Systems Department, Gdansk University of Technology and Excellence Center Communication Process: Hearing and Speech, PROKSIM, Warsaw, Poland [email protected]

Abstract. This research study presents rough set-based decision systems applications to the acoustical domain. Two areas are reviewed for this purpose, namely music information classiﬁcation and retrieval and noise control. The main aim of this paper is to show results of both measurements of the acoustic climate and a survey on noise threat, conducted in schools and students’ music clubs. The measurements of the acoustic climate employ multimedia noise monitoring system engineered at the Multimedia Systems Department of the Gdansk University of Technology. Physiological eﬀects of noise exposure are measured using pure tone audiometry and otoacoustic emission tests. All data are gathered in decision tables in order to explore the signiﬁcance of attributes related to hearing loss occurence and subjective factors that attribute to the noise annoyance. Future direction of experiments are shortly outlined in Summary.

1

Opening Thoughts

Before introducing the particular topic of research presented in the paper, I would like to share a few thoughts. This Section is devoted to some personal aspects of the research carried out by the author for many years. It concerns the fascination of the rough set methodology and the philosophy that lies behind it, and also (or rather in the ﬁrst place) the fascination of the rough set method creator, Professor Zdzislaw Pawlak [34,37]. His personality stands out very clearly amongst other researchers. It happened that his plenary talk I’ve listened to on the occasion of the 2nd International Conference on Rough Sets in Banﬀ guided me toward new interests, namely the applications of decisions rule-based systems which are formidably ﬁtted for uncertainty so often found in acoustics and its analysis. From this time on, we have met many times on various occasions, and I was always inspired by his presentations that led me into new directions and horizons. Professor Pawlak was a mentor to me and I beneﬁted greatly because he was very kind to write Foreword for my two books showing his interest in the research carried out by me and my colleagues. These books perhaps would not happen without his wise patronage. J.F. Peters et al. (Eds.): Transactions on Rough Sets VI, LNCS 4374, pp. 133–151, 2007. c Springer-Verlag Berlin Heidelberg 2007

134

B. Kostek

This is very valuable to me and I will be always grateful to him. Altogether within the rough set society, a clear interest appeared to pursue the rough set-acoustic applications [7,21,22], especially the domain of music evoked many research studies [2,3,8,9,12,13,16,17,18,21,22,23,24,25,29,30,31,40,41,42,43]. For some years now, many researches have published in the joint area of rough sets and acoustic/music, thus some of these names being recalled in References. Lately, also Chinese and Korean contributors to this domain appeared. As a result of the interest in this area a new domain of applications emerged, which focuses on interests such as musical instrument recognition based on timbre descriptors, musical phrase classiﬁcation based on its parameters or contour, melody classiﬁcation (e.g. query-by-humming systems), rhythm retrieval (diﬀerent approaches), high-level-based music retrieval such as looking for emotions in music or diﬀerences in expressiveness, music search based on listeners’ preferences, and others. One may also ﬁnd research studies which try to correlate low-level descriptor analysis to high-level human perception. The semantic description is becoming a basis of the next web generation, i.e., the Semantic Web. Several important concepts have been introduced recently by the researchers associated with the rough set community with regard to semantic data processing including techniques for computing with words [20,33]. Moreover, Zdzislaw Pawlak in his papers [35,36] promoted his new mathematical model of ﬂow networks which can be applied to mining knowledge in databases. Such topics are reﬂected also in papers that followed Prof. Pawlak’s original idea on ﬂow graphs [9,26]. Studies performed on the verge of two domains: soft computing (and particularly rough sets) and acoustics enabled the author to apply for many research grants and many of these projects have been successfully awarded. Once again, the current research is also a good example of the need for employing decision systems to the area which at ﬁrst glance seems far away from the soft computing interests.

2

Introduction

This paper deals with a particular topic which is noise threat-related. As indicated in numerous reports, noise threats occur very frequently nowadays. Occupational exposure limits (OELs) for noise are typically given as the maximum duration of exposure permitted for various noise levels. Environmental noise regulations usually specify a maximum outdoor level of 60 to 65 dB(A), while occupational safety organizations recommend that the maximum exposure to noise is 40 hours per week at 85 to 90 dB(A). For every additional 3 dB(A), the maximum exposure time is reduced by a factor of 2, e.g. 20 hours per week at 88 dB(A). Sometimes, a factor of 2 per additional 5 dB(A) is used. However, these occupational regulations are recognized by the health literature as inadequate to protect against hearing loss and other health eﬀects, especially for sensitive individuals, adverse subjective eﬀects might be expected to appear earlier than for others [4,5,38].

The Domain of Acoustics Seen from the Rough Sets Perspective

135

The background of this study is the fact that younger and younger people experience a noticeable loss in hearing. In previous decades a procedure was established within the audiology ﬁeld, that a group of young males of the age between 18-21 could constitute a reference group for hearing measurements. However, during the last decade numerous studies have shown that this statement is no longer valid. Also, hearing characteristics of students measured during psychoacoustic laboratory sessions at the Multimedia Systems Department have shown that students of this age typically have a threshold shift at 6 kHz. On average, this accounts to 20 dB HL (hearing loss), which is for this age rather unexpected taking into account that students that have any history of ear illnesses were excluded from the experiments. That is why the starting point is to look for the causes of loss in hearing in younger groups of population. The study aimed at showing results of a survey on noise threat which was conducted in schools and students’ music clubs. Noise has an enormous impact on health and life quality of human beings. Noise pollution in Poland is greater than in others UE countries, moreover recently it has been reported to be on the increase [14]. Taking into account the European 2002/49/WE directive related to the control and assessment of environmental noise, monitoring of these threats becomes a necessity [38]. That is why a thorough study on many aspects of noise was envisioned and is carried out for some time at the Multimedia Systems Department [6,10,27,28]. First of all, measurements of the acoustic climate that employed telemetry stations for continuous noise monitoring engineered at the Multimedia Systems Department were conducted. Also, physiological eﬀects of noise were measured among pupils and students. Hearing tests were performed twice, before and after the exposure to noise. For this purpose a so-called distortion product otoacoustic emission method (DPOAE) was utilized. As derived from numerous studies, otoacoustic emission is treated as an early indicator of the occurrence of hearing loss for which reason this method was chosen. The obtained results of noise measurements revealed that an unfavorable noise climate was found in examined schools and music clubs. This was also conﬁrmed by the results of a subjective examination. For the latter purpose students and pupils ﬁlled in a questionnaire expressing their feelings as to noise presence and its annoyance. The noise dose analysis based on average time spent by pupils in schools was also calculated. It revealed that noise in schools did not constitute a risk to the pupils’ hearing system, however, it may be considered as an essential source of annoyance. On the other hand, noise in music clubs surpassed all permitted noise limits, thus could be treated as dangerous to hearing. Hearing tests revealed changes in the cochlea activity of examined students, also the Tinnitus (ringing in the ear) eﬀect was experienced temporarily. In addition, noise annoyance and noise threat criteria and analysis were proposed and veriﬁed based on the acquired and analyzed data. All factors recognized in the study constitute the basis of two types of decision tables that were created. The ﬁrst one consists of the attributes derived from the measurements and calculation of the noise dose, also the presence or absence of the Tinnitus (ringing in the ear) eﬀect is included in this table. The second

136

B. Kostek

decision table gathers data from the survey on noise annoyance. The conditional attributes are subject-driven in this case. Examples of the questions included in this survey are shown in the following sections. The paper aims at showing that a complex and thorough study may lead to better understanding of noise threats and the correlation between the measurement data and the survey responses concerning noise annoyance. Another aim is to show that current regulations are not adequate to predict PTS (Permanent Threshold Shift) early. Finally the data dependency is analyzed, in the reduced database, to ﬁnd the minimal subset of attributes called reduct. The analysis of collected data is done by employing the rough set decision system.

3

Multimedia Noise Monitoring System

The MNMS (Multimedia Noise Monitoring System), developed at the Multimedia Systems Department of the Gdansk University of Technology enables to proceed with the environmental noise measurements in cities on an unparalleled scale now. In general, the MNMS consists of a central database which serves as a repository of measurement results, and numerous equipment tools which execute noise meter functions. One of the proposed devices is a mobile noise monitoring station. The station realizes all measuring functions typical for a sound level meter. It also includes special solutions for long-term measurements and introduces a new type of noise indicators. The application of wireless data transmission technology enables to send data to the server and to remotely control the performance of the station. Since this subject was already published [6], thus its main features are only outlined above.

4

Noise and Hearing Measurements

The noise measurement results, obtained by means of the MNMS, are presented below. The measurements were done in selected schools, musical clubs, and during a musical band rehearsals. Participation in music bands concerts and staying in students’ clubs are a common way of entertainment amongst students. This is why the investigation was carried out also in these locations. The acquired data were utilized to perform the noise dose analysis. This is done to determine the noise exposure of a person staying in the considered places. In selected cases (i.e. schools and musical clubs), the noise dose analysis was expanded by the assessment of hearing. To achieve this, a so-called distortion product otoacoustic emission (DPOAE) measurement and pure tone audiometry were applied. Hearing was examined twice. First, directly before the exposure to noise of a given type, and then immediately after. The performed analysis combined the obtained noise and hearing measurement results. Hearing examinations employed the DPOAE method using GSI 60 DPOAE system. The following parameters of the stimuli were used during tests: L1 equals 65 dB, L2 equals 55 dB, f2/f1 = 1.2, DP frequency (geometric mean): 1062, 1312, 1562, 1812, 2187, 2625, 3062, 3687, 4375, 5187, 6187, 7375 Hz. A DP signal level

The Domain of Acoustics Seen from the Rough Sets Perspective

137

and a noise ﬂoor for every stimuli were registered. The test result was accepted if the diﬀerence between the evoked otoacoustic emission signals and the noise ﬂoor was not less than 10 dB. For pure tone audiometry only selected frequencies were examined: 1000, 1500, 2000, 3000, 4000, 6000, 8000 Hz. The stimuli for each frequency were presented starting from the minimal loudness. The reason of such a selection of parameters was because the noise impact on the hearing system is the strongest for middle and high frequencies. The test was carried out in rooms specially adapted for this purpose. Some measurements performed in schools were interfered with sounds coming from adjoining rooms. Typical response to stimuli is shown in Figure 1. The DP otoacoustic response was found at 3187 Hz, and because the diﬀerence between the noise level and DP signal is larger than 10 dB, thus this measurement is accepted.

Fig. 1. Example of the DPOAE analysis

The following noise parameters LAF min , LAeq , LAmax (see the Equation and deﬁnitions below) were measured independently over broadband and in one-third octave bands. A histogram of time history of LAF instantaneous levels was also calculated. A measuring microphone was located 1.9 m above the ﬂoor level for every measurement. For all measuring series, a place where people gather most often was selected. This was to determine correctly a real noise dose to which they were exposed. N 1 0.1·LAdBi 10 (1) Leq = 10 log N i=1 where: Leq – A-weighted equivalent continuous noise level, N –number of LAdBi values, LAdBi – A-weighted instantaneous sound levels,

138

B. Kostek

On the other hand, LAF min , LAF max denote the lowest and highest A-weighted sound levels for fast time weighting that occurred during the measurement. 4.1

Noise Dose Analysis

The evaluation of both the occupational noise exposure and the risk of developing a permanent hearing loss that may result from the noise exposure are shown in Fig. 2.

Fig. 2. Evaluation of the occupational noise exposure (LEX,8h (D) – daily noise exposure level, LAmax – maximum sound level in working conditions, LCpeak – peak sound level)

The presented evaluation of the occupational noise exposure and the risk of developing a permanent hearing loss is based on noise regulations [4]. The regulations recommend a limit for occupational noise exposure of 85 decibels, A-weighted, as an 8-hour time-weighted average [85 dBA as an 8-hr TWA]). This enables to evaluate whether occupational noise may cause hearing loss, and (or) whether personal a hearing protector (i.e. earmuﬀs, formable earplugs, earplugs, etc.) is required. The A ﬁlter is a weighting curve which approximates equal loudness perception characteristics of the human hearing for pure tones with reference to 40 dB SPL at 1 KHz. It is worth noticing that this curve was established for pure tones, and a potential noise is typically broadband, which means that A-weighting may not provide the best estimate of potential noiseinduced hearing loss. The formal deﬁnition of the noise dose deﬁnes the dose as the amount of actual exposure relative to the amount of allowable exposure, and for which 100% and above represents exposures that are hazardous. The noise dose is calculated according to the following formula:

The Domain of Acoustics Seen from the Rough Sets Perspective

D = [C1 /T1 + C2 /T2 + ... + Cn /Tn ]100%

139

(2)

where D is a dose in the allowable percent, Cn refers to the total time of exposure at a speciﬁed noise level, and Tn denotes the exposure time at which noise for this level becomes hazardous. This deﬁnition allows to calculate a T W A (Timeweighted average) from a Noise Dose: T W A = 90 + 16.61log(D/100)

(3)

where T W A is the 8-hour time-weighted average noise exposure, and D denotes the dose. Unfortunately, the deﬁnition given above is arguable since other formulas of the T W A are also known. The above regulations are recalled before the noise dose calculated for this study is presented, because it is to show that the norms may not provide the best estimate of the potential occurrence of the permanent hearing loss. One may ﬁnd also such conclusions in many research studies [1,15,11,32,39]. Time of the noise exposure for the presented activities is much longer in real conditions than in a time-controlled experiment. A simple survey that included questions about how long pupils/students stay in clubs, play or listen to loud music, stay in school, etc. was also carried out. On the basis of the answers, an average time of the exposure for diﬀerent type of activities was speciﬁed. The total time of the noise exposure in schools, clubs and rehearsing musicians’ is respectively equal to 3600 s, 13500 s, and 5400 s. Based on the assumption that in the indicated places the noise climate is the same, it is possible to obtain the noise dose for people staying in these places. The noise dose for school amounts to not more than 26%, for rehearsing in a musical band to 673%, for club No.1 - 506% and for club No. 2 - 1191% of the daily dose. 4.2

Noise Investigation Results

The obtained noise measurement results are presented in Table 1. Noise investigation was performed in three diﬀerent schools. They diﬀered from each other in the age of the pupils. The youngest pupils attended a primary school (school No. 1). The second school was for children between the age of 13 and 15. The third school was a high school attended by the youth aged from 16 to 19. The biggest noise occurred at the primary school. This is because small children are the main source of noise in schools. They behave extremely vigorously at this age. This entailed a very high noise level. In this school, additional source of noise was loud music played from loudspeakers. In high school No. 3 the noise was produced by loud conversations. It should also be mentioned that in all investigated schools there was no suﬃcient absorption materials covering walls and ceilings, which fact further increased sound level. The fourth measuring series was done during a rehearsal of a small students’ music band. The band consisted of a drummer, a bass player and a keyboard player. This measurement revealed high dynamics of noise level. This is because the musicians often paused and

140

B. Kostek

Table 1. Noise measurement results. Time of noise exposure is expressed in seconds, noise levels in dB and noise dose in per cent of allowed daily noise dose. Measurem. No. Exposure LAF min 1 School No. 1 67.4 2 School No. 2 67.2 3 School No. 3 72.0 4 Music band 52.5 5 Club No. 1 76.2 6 Club No. 2 68.9

LAeq 89.0 85.5 83.6 100.5 95.3 99.0

LAF max 105.5 106.8 97.4 114.4 108.2 114.2

Exposure Time Noise Dose 600 5.2 900 3.5 600 1.5 4058 506.1 4529 169.9 5330 470.0

Fig. 3. LAF histogram measured during lessons

Fig. 4. LAF histogram measured during breaks

consulted each. The 5th and 6th measurement series were carried out in two students’ clubs. They diﬀered in the type of music. In the ﬁrst club a pop music was dominating, while in the second only rock was played. The results of the

The Domain of Acoustics Seen from the Rough Sets Perspective

141

noise dose analysis for the measured time exposures are presented in Table 1. Figures 3 and 4 show histograms measured during lessons and breaks in schools. The breaks were 15 minutes long. The results of noise measurements during two breaks were similar (see Fig. 4). Leq for the breaks was approx. 85dB. Sounds louder than 90 dB appeared at 3-4% of measured time span, however a noise level of 122 dB was also registered. On the other hand, noise during lessons turned out to be less annoying, Leq equaled to 62 dB. In all examined places, a very high noise level was observed. Staying for a long time in places such as music clubs, discotheques, etc. where noise level reaches the values shown in Table 1, can be hazardous to hearing (i.e. may produce a permanent threshold shift). 4.3

Hearing Measurement Results

Several dozens of persons took part in the presented hearing tests. The total number of people examined for diﬀerent types of exposure is presented in Table 2. Table 2 also includes the results of average changes of a hearing threshold (pure tone audiometry), and the results of the DPOAE tests. The average changes of a hearing threshold after the exposure to noise for individual frequencies obtained for pure tone audiometry are presented in Table 3. Two diﬀerent aspects were taken into consideration while analyzing the DPOAE results. First, the number Table 2. Results of hearing testing (in [%])

Measurem. No. 1 2 3 4 5 6

Pure tone audiometry No. of Decrease of Increase of persons threshold threshold 9 21.4 37.3 10 14.3 62.9 12 12.5 75.0

No change 41.3 22.8 12.5

DPOAE test results No. of +Pass -Pass persons 10 11.0 13.6 5 10.0 19.2 5 3.3 12.5 9 5.1 11.6 11 3.4 10.6 12 4.5 20.5

No change 75.4 70.8 84.2 83.3 86.0 75.0

Table 3. The average changes of hearing threshold for pure tone audiometry (in [dB]) Type of noise exposure Music band

Club No. 1

Club No. 2

L R AVG L R AVG L R AVG

1000 2.8 0.0 1.4 2.5 -1.5 0.5 2.9 3.8 3.3

1500 0.0 -1.1 -0.6 6.5 3 4.75 6.7 6.7 6.7

2000 2.2 1.1 1.7 7.5 7 7.25 8.8 10.0 9.4

3000 2.2 0.6 1.4 10.5 10.5 10.5 10.4 12.1 11.3

4000 3.3 2.2 2.8 11.5 10.5 11.0 15.9 12.9 14.4

6000 1.7 3.3 2.5 5 5.5 5.25 12.1 10.8 11.5

8000 1.1 0.5 0.8 1 3.5 2.25 -1.7 0.8 -0.4

142

B. Kostek

Table 4. The average changes of DP signal level for particular types of exposures (in [%]) Measurement No. 1 2 3 Increase 30.3 27.5 36.7 Decrease 28.1 30.0 34.2 No change 41.6 42.5 29.1

4 25.0 44.4 30.6

5 27.3 45.4 27.3

6 16.3 53.5 30.2

of “passed” and “failed” tests for the DPOAE examination were determined. The result of the ﬁrst examination served as reference. The symbol “+Pass” indicates that a pupil failed the ﬁrst examination and passed the second one. The symbol “–Pass” signiﬁes a reverse situation (a test passed in the ﬁrst examination and failed after the exposure to noise). The results are presented in Table 2, in the “DPOAE test results” column. The second kind of analysis determined how the DP signal level changed under the inﬂuence of the exposure to noise. The results of this analysis are presented in Table 4. As seen from the tables, in most cases the hearing threshold was increased. It means that almost every person had a TTS after noise exposure. Also some people reported the perception of the Tinnitus eﬀect as well. The most signiﬁcant TTS occurred for 4000 Hz. The data obtained from DPOAE tests conﬁrm negative after-eﬀects of the noise exposure analysis acquired by means of pure tone audiometry.

5

Psychophysiological Noise Dosimeter

Methods of the estimation of noise-induced hearing loss presented in this paper are based mainly on the equal energy hypothesis [5]. This approach focuses on an assessment of the quantity of energy which aﬀects the hearing system. The time characteristics of noise is neglected, and the main emphasis is placed on the

Fig. 5. General scheme of the psychophysiological noise dosimeter

The Domain of Acoustics Seen from the Rough Sets Perspective

143

assessment of the equivalent sound level value. However, in many cases this approach could be insuﬃcient. The analysis of numerous literature data, including testing of the exposure to noise from diﬀerent sources, provides a knowledge that time characteristics and noise spectrum have an essential signiﬁcance in generating loss in hearing [1,11,15]. Taking these data into consideration, a method of estimating the risk of hearing impairment has been proposed by J. Kotus, the author’s Ph.D. student. The proposed model is based on a modiﬁed Johnston’s psychoacoustical model [19], which provides a global distribution of a basilar membrane deﬂection within critical bands. In Figure 5, a scheme of the psychophysiological noise dosimeter is presented. The model is based on the analysis of a basilar membrane’s answer to noise in critical bands. First, the power spectrum of noise is determined. Afterwards, it is corrected by taking into account the transition from the outer to the inner ear. Subsequently, particular spectrum coeﬃcients are grouped into critical bands, according to the bark scale. Then the noise level in particular critical bands is calculated. The result let us assess the extent to which the basilar membrane is stimulated. Its answer is determined by the multiplication of the instate stimulation level value by the characteristics of hearing ﬁlters for the particular critical band. A basilar membrane displacement value obtained in this way is exponentially averaged. This action reﬂects the inertia of the processes occurring in the inner ear. The obtained averaged values are used for the assessment of the asymptotic hearing threshold shift. Finally, these values are subjected to exponential averaging, that reﬂects a process of the hearing threshold shift. Therefore this model enables to assess TTS in critical bands and the recovery time of a hearing threshold to its initial value. The model enables to determine the hearing threshold shift for a given noise during the exposure. The initial simulation for two selected noise exposures was done (series 4 and 6). A theoretical time series of noise was created based on the histogram of the LAF levels. The obtained TTS levels for a particular measurement series amounted to: TTS = 13.5 dB (measurement No. 4), and TTS = 18.5 dB (measurement No. 6). A comparison of the noise exposure for series 4 and 6 has shown that the time characteristic of noise inﬂuences the occurrence and level of the TTS. The results obtained through the use of the presented model, conﬁrmed a greater TTS for the 6th exposition (club No. 2). Overall, this conﬁrms harmful eﬀects of noise.

6

Survey on Noise Annoyance

An objective noise measurement was extended by a subjective measurement by means of a dedicated survey. The survey consisted of three parts. The ﬁrst part involved getting information such as age, sex, class, school. The second part included questions about noise in places of residence and exposure to noise related to musical preferences. The last part concentrated on noise climates in schools in typical circumstances (lessons, breaks, etc.). In the following Figures (Figs. 6-10) some sample results are shown.

144

B. Kostek

Fig. 6. Main type of noise sources in the neighborhood during day and night

Fig. 7. Subjective evaluation of noise annoyance during day (CL – comfortable)

Fig. 8. Subjective evaluation of noise annoyance during night

Main type of noise sources in the neighborhood As seen in Figure 6 the main sources of noise during day and night are communication and the neighborhood. They constitute 60% of all noise sources. On the other hand, industrial noise does not present a threat to the place of living.

The Domain of Acoustics Seen from the Rough Sets Perspective

145

Subjective evaluation of noise annoyance Overall, 8 % of asked pupils evaluated their place of living as quiet, 40% said that the neighborhood is either too loud or loud. On the other hand, during the night most places were evaluated as quiet, apart from special events occurring during the night. It should be mentioned that 70% of our responders live in towns or in close proximity to towns. Subjective evaluation of noise annoyance in schools Almost 70% of pupils evaluate noise as too loud during breaks, and the remaining 30% as loud. As to noise evaluation during lessons pupils diﬀer in their opinions.

Fig. 9. Subjective evaluation of noise annoyance during breaks

Fig. 10. Subjective evaluation of noise annoyance during lessons

Evaluation of noise sources related to personal way of living In addition, pupils were asked how often they listen to music and how loud this music is, and also how often they use personal music players and listen to music via headphones. It occurs that on average younger groups of pupils do not often listen to loud music, contrarily to older pupils and students. The latter group

146

B. Kostek

also attends musical events very frequently. As a result, one may notice problems with speech understanding when noise is present, which in students’ subjective opinions occurred in 50% of situations.

7

Data Analysis Employing Rough Sest

Regulations and gathered data enabled to deﬁne attributes that may provide ways to assess the risk of developing PTS (Permanent Threshold Shift) and noise annoyance. Conditional attributes from the decision table (Table 5) are derived from objective measurements and noise dose calculation. One of the attributes driven more subjectively, namely the absence and presence of the Tinnitus eﬀect after the exposure to noise, is also taken into consideration. As mentioned before, otoacoustic emission is treated as an early indicator of hearing loss, that is why the DPOAE analysis should be included in the decision table. In a way the decision table follows results included in Tables 1 and 2. Another decision table (Table 6) includes answers gathered at schools based on a questionnaire form published in the Internet (MNMS service [7]). The decision attribute are in this case noise annoyance/nuisance or noise threat. A legal deﬁnition of a nuisance says that this is the noise that oﬀends or upsets the receiver because it is occurring at the wrong time, in the wrong place, or is of a character that annoys due to excessive tonal components or impulses. When looking at the table, one may expect that sensitivity to noise is more frequent among children and increases with age. This could be explained by the fact that small children get easily tired in noisy conditions, and on the other hand, people active professionally prefer some rest after work, the same as older people. This means that the age of the survey respondents’ should be taken into consideration. Also, it is quite obvious that females are less noisy than males while entertaining. When analyzing other attributes from Table 5 the relation between them and the decision attribute is not always that clear, and since this is a subjectively-driven questionnaire many contradictions may occur in the acquired data. A good example of such a confusing attribute is ‘hearing’. Hearing impairments of the inner ear are characterized by the so-called recruitment phenomenon, thus people with the inner-ear hearing loss may suﬀer more in noisy conditions. This means that they do not hear soft sounds, and perceive loud sounds as too loud. This phenomenon is due to an abnormally rapid rise of the loudness sensation with only a small increase in sound stimulus intensity. In this early study, the survey did not include questions related to for example social class, since it was directed at speciﬁc groups (pupils attending schools and students). The observations and earlier researches on noise annoyance indicate that these factors should also be taken into account. All these factors constitute huge data, especially as the survey will be progressively ﬁlled in by the Internet respondents. This enables to consolidate data on noise annoyance on a very large scale. The only way to analyze subjectivelydriven data is to use a soft computing approach, and particularly the rough set

The Domain of Acoustics Seen from the Rough Sets Perspective

147

method which seems to be most suitable for this problem. The rough set-based analysis of the decision tables enables to derive rules in which attributes and their values support a claim that noise dose analysis based on average daily exposure is not only insuﬃcient to assess noise annoyance, but also to predict the risk of permanent threshold shift early enough. There are a lot of data in which we see noise events recalled by the survey respondents even if they live in a quiet neighborhood with no evident source of noise during the day and night. This means that they evaluate the noise annoyance or nuisance as high because of noise events that happen sporadically and were of short duration. In such a case, factors derived from the noise dose analysis will not exceed any norm or regulations, but still these noise events may be highly stressful and harmful. Table 5. Decision table based on measurements and calculation according to noise regulations (denotation same as before, in addition: Ts denotes absence or presence of the Tinnitus eﬀect) No. 1 2 ... n

LAF min 67.4 52.5 ... ...

Leq 89 100.5 ... ...

LAF max 105.5 114.4 ... ...

Exposure Time . . . 600 4058 ... ... ... ...

D 5.2 506.1 ... ...

DPOAE ... ... ... ...

TTS YES YES ... ...

Ts YES YES ... ...

PTS NO YES ... ...

Table 6. Decision table based on survey (denotations: StoN – sensitivity to noise, H – hearing, SC – social class, WP/S – type of work performed/school, SR/D – Survey results/neighborhood day, SR/N – Survey results/neighborhood night, SR/NS – Survey results/noise source, NE/AdNS – Noise events/additional noise sources, NA/N – noise annoyance/nuissance) No. Age Sex StoN H SC WP/S SR/D SSR/N SR/NS 1 41-50 M high good high educat. quiet quiet no 2 10-15 F high good NO high noisy quiet road school traﬃc ... ... ... ... ... ... ... ... ... ... n ... ... ... ... ... ... ... ... ...

. . . NE/AdNS NA/N . . . dogs HIGH . . . ambul. HIGH siren ... ... ... ... ... ...

A rule prototype derived from the rough set-based analysis is presented below: (attribute 1) = (val 1) ∧ ... ∧ (attribute k) = (val k) => (P T SY ES/N O )

(4)

It should be mentioned that some of the numerical attributes require quantization, however in the case of the noise-related attributes the quantization should be based on norms and standards. This means a process of replacing the original values of the input data with for example the number of an interval to which a selected parameter value belongs should be performed taking into account speciﬁc constraints. For example, it is well-known that values of LAF max , Leq , Exposure

148

B. Kostek

Time exceeding the limits for which these values start to be harmful constitute the cut-points. It is obvious that after rule derivation not all attributes are to be retained and some of the rules are not longer valid. Most important attributes are to be found in reducts. Preliminary experiments show that amongst the most signiﬁcant attributes resulting from measurements one may ﬁnd: LAF max , Leq , Exposure Time and TTS, though some of these parameters are interrelated, and are eliminated in the rough set-based analysis, reducing the number of the correlated attributes. On the other hand, looking into the Decision Table based on survey results, one may ﬁnd that such attributes as: age of a person, speciﬁc type of noise events, duration of these events, neighborhood noise background are present in the evaluation of noise annoyance/nuissance. Having the rough set analysis performed and rules derived based on both approaches (objective and subjective), this may constitute a scientiﬁc basis of an advertising campaign against noise pollution and its adverse consequences, which in addition may contribute to better regulations on noise. This is why noise monitoring stations should be installed in all agglomerations larger than 250,000 inhabitants in the coming year, thus both measurement quantities shown in a form of noise maps and subjective opinions would be available for the same place, in future. However, it is worth noticing that without changing our habits, regulations may still be ineﬀective as seen from the survey.

8

Summary

To sum up, on the basis of the investigations, it was conﬁrmed that noise climates in schools is adverse to pupils’ and teachers’ health. The main reasons of the high noise level in schools are: the behavior of pupils, loudspeaker systems and low sound absorption of the classrooms and corridors. The data analysis of the hearing measurements at schools does not conﬁrm negative inﬂuence of noise on the hearing system. Especially because the time of exposure to noise is too short to produce measurable changes in the activity of the inner ear. Noise measured during of the students’ music band rehearsals and in clubs reaches very high levels and exceeds all related norms. Measurements of the hearing characteristics of people working or entertaining in these places conﬁrmed harmful eﬀects of noise. A signiﬁcant TTS and the reduction of the DP level were observed. A comparison of the noise exposure for series 4 and 6 has shown that the time characteristic of noise inﬂuences the occurrence of the TTS. The results obtained through the use of the presented model, conﬁrmed a greater TTS for the 6th exposition (club No. 2). In addition, the Tinnitus eﬀect was perceived by some students. The data presented are very complex, interrelated, and in some cases contradictory, thus for an adequate analysis they require the use of a decision system. For this particular problem, it is thought that the rough set method is the most suitable solution for the analysis, since it allows to derive rules/reduct capable of identifying the most signiﬁcant attributes. It is then possible to determine the weights that should be taken into consideration when constructing a feature vector.

The Domain of Acoustics Seen from the Rough Sets Perspective

149

On the basis of the survey, it may be said that even if norms are not exceeded, noise can be still perceived as annoying, stressful and in consequence harmful. The analysis of the acquired data has shown that all measurement quantities should ﬁrst be included as conditional attributes in the constructed decision tables along with subjective factors. At the moment, two types of decision tables have been proposed. The ﬁrst one provides indications as to the signiﬁcance of some attributes in relation to the PTS occurrence (decision attribute). The second gathers results of the survey, and relates subjective factors to the noise annoyance. The results of this analysis may provide the basis for an adequate campaign against noise pollution and lead to better regulations. Acknowledgment. This work was supported by the Polish Ministry of Science and Education within the projects No. 3T11E02829 and No. R0201001. The author wishes to acknowledge her Ph.D. student J. Kotus for his valuable input to the presented research.

References 1. Borg, E., Engstrom, B.: Noise level, inner hair cell damage, audiometric features and equal-energy hypothesis, J Acoust. Soc. Am. 86 (5) (1989) 1776–1782 2. Budzynska, L., Jelonek, J., Lukasik, E., Slowinski, R.,: Supporting Experts in Ranking Generic Audio and Visual Objects, Proc. IEEE Workshop ”Signal Processing’2004”, Poznan (2004) 81–86. 3. Budzynska, L., Jelonek, J., Lukasik, E., Susmaga, R., Slowinski R.: Multistimulus ranking versus pairwise comparison in assessing quality of musical instruments sounds, 118 AES Convention Paper, 6482, Barcelona (2005) 4. Criteria for a recommended standard, Occupational Noise Exposure, U.S. Department of Health and Human Services (1998) 5. http://www.cdc.gov/niosh/98-126.html (CRITERIA FOR A RECOMMENDED STANDARD) 6. Czyzewski, A., Kostek, B., Skarzynski, H.: Intelligent System for Environmental Noise Monitoring, in Monitoring, Security, and Rescue Techniques in Multiagent Systems, Series: Advances in Soft Computing, Dunin-Keplicz, B.; Jankowski, A.; Skowron, A.; Szczuka, M.(eds.), chapter, 397–410 , XII, Springer Verlag, Heidelberg, New Yorkc(2005) 7. Czyzewski, A., Kostek, B., Skarzynski, H.: IT applications for the remote testing of communication senses” chapter in INFORMATION TECHNOLOGY SOLUTIONS FOR HEALTH CARE, Spinger-Verlag (2006) 8. Czyzewski, A., Szczerba M., Kostek B.: Musical Phrase Representation and Recognition by Means of Neural Networks and Rough Sets, Rough Set Theory and Applications (RSTA), vol. 1, 259-284, Advances in Rough Sets, Subseries of SpringerVerlag Lecture Notes in Computer Sciences, LNCS 3100, Transactions on Rough Sets, Grzymala-Busse, J.W., Kostek, B., Swiniarski, R.W., Szczuka M. (eds.) (2004) 9. Czyzewski A., Kostek B.: Musical Metadata Retrieval with Flow Graphs, in Rough Sets and Current Trends in Computing, RSCTC, Uppsala, Sweden, Lecture Notes in Atiﬁcial Intelligence, LNAI 3066, Springer Verlag, Berlin, Heidelberg, New York (2004) 691–698

150

B. Kostek

10. Czyzewski, A., Kotus, J., Kostek, B.,: Comparing Noise Levels and Audiometric Testing Results Employing IT Based Diagnostic Systems, The 33rd International Congress and Exposition on Noise Control Engineering INTERNOISE’2004, August 22-24, Prague (2004) 11. Dunn, D.E., Davis, R.R., Merry, C.J., Franks, J.R.: Hearing loss in the chinchilla from impact and continuous noise exposure, J Acoust. Soc. Am. 90 (4) (1991) 1979–1985 12. Dziubinski, M., Dalka, P., Kostek, B.: Estimation of Musical Sound Separation Algorithm Eﬀectiveness Employing Neural Networks, J. Intelligent Information Systems, Special Issue on Intelligent Multimedia Applications, 24, 2(2005) 133– 157 13. Dziubinski, M., Kostek, B.: Octave Error Immune and Instantaneous Pitch Detection Algorithm, J. of New Music Research, vol. 34, 292-273, Sept. 2005. 14. Engel, Z.W., Sadowski J., et al.: Noise protection in Poland in European Legislation, The Committee on Acoustics of the Polish Academy of Science & CIOP-PIB, Warsaw, (2005) (in Polish) 15. Henderson, D., Hamernik, R.P.: Impulse noise: Critical review, J Acoust. Soc. Am. 80(2) (1986) 569–584 16. Hippe, M.P.: Towards the Classiﬁcation of Musical Works: A Rough Set Approach Third International Conference, RSCTC 2002, Malvern, PA, USA, October 14-16, 2002. Proceedings Editors: J.J. Alpigini, J.F. Peters, A. Skowron, N. Zhong (eds.) (2002) 546-553 17. Jelonek, J., Lukasik, E., Naganowski, A., Slowinski, R.: Inferring Decision Rules from Jurys’ Ranking of Competing Violins, Proc. Stockholm Music Acoustic Conference, KTH, Stockholm (2003) 75–78 18. Jelonek, J., Lukasik, E., Naganowski, A., Slowinski, R.: Inducing jury’s preferences in terms of acoustic features of violin sounds, Lecture Notes in Computer Science, LNCS 3070, Springer (2004) 492–497 19. Johnston, J.D.: Transform Coding of Audio Signals Using Perceptual Noise Criteria. IEEE Journal on Selected Areas in Communications, vol. 6(2) (1988) 314–323 20. Komorowski, J, Pawlak, Z, Polkowski, L, Skowron, A. Rough Sets: A Tuto-rial. In: Pal SK, Skowron A (eds) Rough Fuzzy Hybridization: A New Trend in DecisionMaking. Springer-Verlag (1998), 3–98 21. Kostek, B.: Soft Computing in Acoustics, Applications of Neural Networks, Fuzzy Logic and Rough Sets to Musical Acoustics, Physica Verlag, Heidelberg, New York (1999) 22. Kostek, B.: Perception-Based Data Processing in Acoustics. Applications to Music Information Retrieval and Psychophysiology of Hearing, Springer Verlag, Series on Cognitive Technologies, Berlin, Heidelberg, New York (2005) 23. Kostek, B.: Musical Instrument Classiﬁcation and Duet Analysis Employing Music Information Retrieval Techniques, Proc. of the IEEE, 92, 4 (2004) 712–729 24. Kostek, B.: Intelligent Multimedia Applications - Scanning the Issue, J. Intelligent Information Systems, Special Issue on Intelligent Multimedia Applications, 24, 2 (2005) 95–97 (Guest Editor) 25. Kostek, B., Wojcik, J.: Machine Learning System for Estimating the Rhythmic Salience of Sounds, International J. of Knowledge-based and Intelligent Engineering Systems, 9 (2005), 1–10

The Domain of Acoustics Seen from the Rough Sets Perspective

151

26. Kostek, B., Czyzewski, A.: Processing of Musical Metadata Employing Pawlak’s Flow Graphs, Rough Set Theory and Applications (RSTA), vol. 1, 285–305, Advances in Rough Sets, Subseries of Springer-Verlag Lecture Notes in Computer Sciences, LNCS 3100, Transactions on Rough Sets, Grzymala-Busse, J.W., Kostek, B., Swiniarski, R.W., Szczuka, M., (eds.) (2004) 27. Kotus, J., Kostek, B.: Investigation of Noise Threats and Their Impact on Hearing in Selected Schools, OSA’ 2006, Archives of Acoustics (2006) (in print). 28. Kotus, J.: Evaluation of Noise Threats and Their Impact on Hearing by Employing Teleinformatic Systems, (Kostek, B., supervisor) (2007) (in preparation). 29. Lukasik, E.: AMATI-Multimedia Database of Violin Sounds. In: Proc Stockholm Music Acoustics Conference, KTH Stockholm (2003a) 79–82 30. Lukasik, E.: Timbre Dissimilarity of Violins: Speciﬁc Case of Musical Instruments Identiﬁcation. Digital Media Processing for Multimedia Interactive Services, World Scientiﬁc, Singapore (2003b) 324–327 31. Lukasik, E., Susmaga, R.: Unsupervised Machine Learning Methods in Timbral Violin Characteristics Visualization. In: Proc Stockholm Music Acoustics Conference, KTH Stockholm (2003) 83–86 32. Melnick, W.: Human temporary threshold shift (TTS) and damage risk, J Acoust. Soc. Am. 90(1) (1991) 147–154 33. Pal, S.K., Polkowski, L., Skowron, A. Rough-Neural Computing. Techniques for Computing with Words. Springer Verlag, Berlin Heidelberg New York (2004) 34. Pawlak, Z.: Rough Sets. International J Computer and Information Sciences (1982) 35. Pawlak, Z.: Probability, Truth and Flow Graph. Electronic Notes in Theoretical Computer Science 82, International Workshop on Rough Sets in Knowledge Discovery and Soft Computing, Satellite event of ETAPS 2003, Elsevier, Warsaw (2003) 36. Pawlak, Z.: Elementary Rough Set Granules: Towards a Rough Set Processor. In: Pal SK, Polkowski L, Skowron A (eds) Rough-Neural Computing. Techniques for Computing with Words. Springer Verlag, Berlin Heidelberg New York, 5–13(2004) 37. Pawlak, Z.: A Treatise on Rough Sets. Transactions on Rough Sets IV, Peters, J.F., Skowron, A. (Eds) 1–17 (2005) 38. Polish Standard PN-N-01307 (1994), Permissible sound level values in work-places and general requirements concerning taking measurements (in Polish). 39. Seixas, N., et al.: Alternative Metrics for Noise Exposure Among Construction Workers, Ann Occup Hyg. 49 (2005) 493–502 40. A. Wieczorkowska, P. Synak, R. Lewis, Z. W. Ras, Creating Reliable Database for Experiments on Extracting Emotions from Music. In: M. A. Klopotek, S. Wierzchon, K. Trojanowski (eds.), Intelligent Information Processing and Web Mining, Proceedings of the International IIS: IIPWM’05 Conference, Gdansk, Poland Advances in Soft Computing, Springer (2005), 395-402 41. Wieczorkowska, A., Synak, P., Lewis, R., Ras, Z.W.: Extracting Emotions from Music Data, in: M.-S. Hacid, Murray, N.V., Ras Z.W., Tsumoto, S. (eds.), Foundations of Intelligent Systems, 15th International Symposium, ISMIS 2005, Saratoga Springs, NY, USA, 2005, Proceedings; LNAI 3488, Springer, 456-465 42. Wieczorkowska, A., Ras, Z.W.: Do We Need Automatic Indexing of Musical Instruments?, in: Warsaw IMTCI, International Workshop on Intelligent Media Technology for Communicative Intelligence, Warsaw, Poland, September 13–14, Proceedings, PJIIT - Publishing House (2004), 43–38 43. Wieczorkowska, A.: Towards Extracting Emotions from Music. In: Warsaw IMTCI, International Workshop on Intelligent Media Technology for Communicative Intelligence, Warsaw, Poland, September 13–14, Proceedings, PJIIT - Publishing House (2004) 181–183

Rule Evaluations, Attributes, and Rough Sets: Extension and a Case Study Jiye Li1 , Puntip Pattaraintakorn2, and Nick Cercone3 1

2

David R. Cheriton School of Computer Science, University of Waterloo 200 University Avenue West, Waterloo, Ontario, Canada N2L 3G1 [email protected] Department of Mathematics and Computer Science, Faculty of Science King Mongkut’s Institute of Technology Ladkrabang, Thailand [email protected] 3 Faculty of Science and Engineering, York University 4700 Keele Street, North York, Ontario, Canada M3J 1P3 [email protected]

Abstract. Manually evaluating important and interesting rules generated from data is generally infeasible due to the large number of rules extracted. Diﬀerent approaches such as rule interestingness measures and rule quality measures have been proposed and explored previously to extract interesting and high quality association rules and classiﬁcation rules. Rough sets theory was originally presented as an approach to approximate concepts under uncertainty. In this paper, we explore rough sets based rule evaluation approaches in knowledge discovery. We demonstrate rule evaluation approaches through a real-world geriatric care data set from Dalhousie Medical School. Rough set based rule evaluation approaches can be used in a straightforward way to rank the importance of the rules. One interesting system developed along these lies in HYRIS (HYbrid Rough sets Intelligent System). We introduce HYRIS through a case study on survival analysis using the geriatric care data set.

1

Introduction

The general models of knowledge discovery in databases (KDD) contains processes including data preprocessing, knowledge discovery algorithms, rule generations and evaluations. Rule evaluation is a signiﬁcant process in KDD. How to automatically extract important, representative rules to the human beings instead of selecting those useful rules manually are the main problems. Speciﬁc diﬃculties make the research of rule evaluation very challenging. One of the diﬃculties is that real-world large data sets normally contain missing attribute values. They may come from the collecting process, or redundant scientiﬁc tests, change of the experimental design, privacy concerns, ethnic issues, unknown data and so on. Discarding all the data containing the missing attribute values cannot fully preserve the characteristics of the original data, and J.F. Peters et al. (Eds.): Transactions on Rough Sets VI, LNCS 4374, pp. 152–171, 2007. c Springer-Verlag Berlin Heidelberg 2007

Rule Evaluations, Attributes, and Rough Sets: Extension and a Case Study

153

wastes part of the data collecting eﬀort. Knowledge generated from missing data may not fully represent the original data set, thus the discovery may not be as suﬃcient. Understanding and utilizing of original context and background knowledge to assign the missing values seem to be an optimal approach for handling missing attribute values. In reality, it is diﬃcult to know the original meaning for missing data from certain application domains. Another diﬃculty is that huge amount of rules are generated during the knowledge discovery process, and it is infeasible for humans to manually select useful and interesting knowledge from such rule sets. Rough sets theory, originally proposed in the 1980’s by Pawlak [1], was presented as an approach to approximate concepts under uncertainty. The theory has been widely used for attribute selection, data reduction, rule discovery and many knowledge discovery applications in the areas such as data mining, machine learning and medical diagnoses. We are interested in tackling diﬃcult problems in knowledge discovery from a rough sets perspective. In this paper, we introduce how rough sets based rule evaluations are utilized in knowledge discovery systems. Three representative approaches based on rough sets theory are introduced. The ﬁrst approach is to provide a rank of how important is each rule by rule importance measure (RIM) [2]. The second approach is to extract representative rules by considering rules as condition attributes in a decision table [3]. The third approach is applied to data containing missing values. This approach provides a prediction for all the missing values using frequent itemsets as a knowledge base. Rules generated from the complete data sets contain more useful information. The third approach can be used at the data preprocessing process, combining with the ﬁrst or second approach at the rule evaluation process to enhance extracting more important rules. It can also be used alone as preprocessing of missing attribute values. An interesting system based on this rule-enhanced knowledge discovery system, HYRIS (HYbrid Rough sets Intelligent System) [4], is developed. Case studies on using HYRIS on survival analysis are further demonstrated. We address particular problems from real-world data sets, using recent missing attribute value techniques and rule evaluations based on rough sets theory to facilitate the tasks of knowledge discovery. The rule discovery algorithm focuses on association rule algorithms, although it can be classiﬁcation algorithm, decision tree algorithm and other rule discovery algorithms from data mining and machine learning. We demonstrate the rule evaluation approaches using a real-world geriatric care medical data set. We discuss related work on rough sets theory, current knowledge discovery system based on rough sets, and rule evaluations in Section 2. Section 3 presents three rough sets based rule evaluations methods. We show experiments on the geriatric care data set in Section 4. Section 5 contains a case study of HYRIS system developed based on the proposed approaches, and experiments on survival analysis are demonstrated. Section 6 gives the concluding remarks.

154

2

J. Li, P. Pattaraintakorn, and N. Cercone

Related Work

We introduce related work to this paper including rough sets theory, knowledge discovery systems based on rough sets theory and existing rule evaluations approaches. 2.1

Rough Sets Theory

Rough sets theory, proposed in the 1980’s by Pawlak [1], has been used for attribute selection, rule discovery and many knowledge discovery applications in the areas such as data mining, machine learning and medical diagnoses. We brieﬂy introduce rough sets theory [1] as follows. U is the set of objects we are interested in, where U = φ. Deﬁnition 1. Equivalence Relation. Let R be an equivalence relation over U , then the family of all equivalence classes of R is represented by U/R. [x]R means a category in R containing an element x ∈ U . Suppose P ⊆ R, and P = φ, IN D(P ) is an equivalence relation over U . For any x ∈ U , the equivalence class of x of the relation IN D(P ) is denoted as [x]P . Deﬁnition 2. Lower Approximation and Upper Approximation. X is a subset of U , R is an equivalence relation, the lower approximation of X and the upper approximation of X is deﬁned as: RX = ∪{x ∈ U |[x]R ⊆ X}

(1)

RX = ∪{x ∈ U |[x]R ∩ X = φ}

(2)

respectively. From the original deﬁnitions [1], reduct and core are deﬁned as follows. R is an equivalence relation and let S ∈ R. We say, S is dispensable in R, if IN D(R) = IN D(R − {S}); S is indispensable in R if IN D(R) = IN D(R − {S}). We say R is independent if each S ∈ R is indispensable in R. Deﬁnition 3. Reduct. Q is a reduct of P if Q is independent, Q ⊆ P , and IN D(Q) = IN D(P ). An equivalence relation over a knowledge base can have many reducts. Deﬁnition 4. Core. The intersection of all the reducts of an equivalence relation P is deﬁned to be the Core, where Core(P ) = ∩All Reducts of P. Reduct and core are among the most important concepts in this theory. A reduct contains a subset of condition attributes that are suﬃcient enough to represent the whole data set. The reducts can be used in attribute selection process. There may exist more than one reduct for each decision table. Finding all the reduct sets for a data set is NP-hard [5]. Approximation algorithms are used to obtain reduct sets [6]. The intersection of all the possible reducts is called the core.

Rule Evaluations, Attributes, and Rough Sets: Extension and a Case Study

155

The core is contained in all the reduct sets, and it is the essential of the whole data. Any reduct generated from the original data set cannot exclude the core attributes. Reduct Generations. There are several reduct generation approaches, such as ROSETTA [7], RSES [8], ROSE2 [9], QuickReduct algorithm [10] and Hu et al. [11]’s reduct generation combining the relational algebra with the traditional rough sets theory. ROSETTA rough set system GUI version 1.4.41 [7] provides Genetic reducer, Johnson reducer, Holte1R reducer, Manual reducer, Dynamic reducer, RSES Exhaustive reducer and so on. Genetic reducer is an approximation algorithm based on genetic algorithm for multiple reducts generation. Johnson reducer generates only a single reduct with minimum length. In this research, we use both genetic and Johnson’s reduct generations to develop rule evaluations approaches. Core Generation. Hu et al. [11] introduced a core generation algorithm based on rough sets theory and eﬃcient database operations, without generating reducts. The algorithm is shown in Algorithm 1, where C is the set of condition attributes, and D is the set of decision attributes. Card denotes the count operation in databases, and Π denotes the projection operation in databases. Algorithm 1. Hu’s Core Generating Algorithm [11] input : Decision table T (C, D), C is the condition attributes set; D is the decision attribute set. output: Core, Core attributes set. Core ← φ; for each condition attribute A ∈ C do if Card(Π(C − A + D)) = Card(Π(C − A)) then Core = Core ∪ A; end end return Core;

This algorithm is developed to consider the eﬀect of each condition attribute on the decision attribute. The intuition is that, if the core attribute is removed from the decision table, the rest of the attributes will bring diﬀerent information to the decision making. Theoretical proof of this algorithm is provided in [11]. The algorithm takes advantage of eﬃcient database operations such as count and projection. This algorithm requires no inconsistency in the data set. 2.2

Rough Sets Based KDD Systems

We brieﬂy survey current rough sets based knowledge discovery systems. We discuss the individual functions of each system based on general characteristics, such as the input data sets, the preprocessing tasks, the related rough sets tasks, the rule generations and so on.

156

J. Li, P. Pattaraintakorn, and N. Cercone

1. ROSETTA. ROSETTA [7] software is a general purpose rough set toolkit for analyzing the tabular data, and is freely distributed. The downloadable versions for both the Windows and Linux operating systems are available. The software supports the complete data mining process, from data preprocessing, including processing incomplete data, data discretization, generating reduct sets which contain essential attributes for the given data set, to classiﬁcation, rule generation, and cross validation evaluation. Some discretization and reducts generation packages are from RSES library [8]. 2. RSES2.2. RSES [8] stands for Rough Set Exploration System. There are downloadable versions for both the Windows and Linux operating systems. It is still maintained and being developed. The system supports data preprocessing, handling incomplete data, discretization, data decomposition into parts that share the same properties, reducts generation, classiﬁcation, and cross validations and so on. 3. ROSE2. ROSE [9] stands for Rough Sets Data Explorer. This software is designed to process data with large boundary regions. The software supports data preprocessing, data discretization, handling missing values, core and reducts generation, classiﬁcations and rule generation, as well as evaluations. This software provides not only the classical rough set model, but also the variable precision model, which is not provided by [7] and [8]. 4. LERS. LERS [12] stands for Learning from Examples based on Rough Sets. It is not publicly available. The system was designed especially to process missing values of attributes and inconsistency in the data set. Certain rules and possible rules are both extracted based on the lower and upper approximations. In addition to the rough sets based systems mentioned above, there are other available knowledge discovery systems based on the methodologies of rough sets such as GROBIAN [13] and DBROUGH [14]. 2.3

Current Research on Rule Evaluations

Rule generation often brings a large amount of rules to analyze. However, only part of these rules are distinct, useful and interesting. How to select only useful, interesting rules among all the available rules to help people understand the knowledge in the data eﬀectively has drawn the attention of many researchers. Research on designing eﬀective measures to evaluate rules comes from statistic, machine learning, data mining and other ﬁelds. These measures fall into two categories of evaluation measures. Rule Interestingness Measures. One category of evaluating rules is to rank the rules by rule interestingness measures. Rules with higher interestingness measures are considered more interesting. The rule interestingness measures, originated from a variety of sources, have been widely used to extract interesting rules. Diﬀerent applications may have diﬀerent interestingness measures emphasizing on diﬀerent aspect of the applications. Hilderman provided an extensive survey

Rule Evaluations, Attributes, and Rough Sets: Extension and a Case Study

157

on the current interestingness measures [15] for diﬀerent data mining tasks. For example, support and conﬁdence are the most common interestingness measures to evaluate the association rules. Not all the interestingness measures generate the same rank of interestingness for the same set of rules. Depending on diﬀerent application purpose, appropriate rule interestingness measures should be selected to extract proper rules. More than one measure can be applied together to evaluate and explain the rules. Tan et. al. [16] evaluate twenty one measures in their comparative experiments and suggest diﬀerent usage domains for these measures. They provide several properties of the interestingness measures so that one can choose a proper measure for certain applications. Their experiments also imply that not all the variables perform equally good at capturing the dependencies among the variables. Furthermore, there is no measure that can perform constantly better than the others in all application domains. Diﬀerent measure is designed towards diﬀerent domains. Rule Quality Measures. The concept of rule quality measures was ﬁrst proposed by Bruha [17]. The motivation for exploring this measure is that decision rules are diﬀerent with diﬀerent predicting abilities, diﬀerent degrees to which people trust the rules and so on. Measures evaluating these diﬀerent characteristics should be used to help people understand and use the rules more eﬀectively. These measures have been known as rule quality measures. The rule quality measures are often applied in the post-pruning step during the rule extraction procedure [18]. For example, some measures are used to evaluate whether the rules overﬁt the data. When removing an attribute-value pair, the quality measure does not decrease in value, this pair is considered to be redundant and will be pruned. As one of the applications, rule generation system uses rule quality measures to determine the stopping criteria for the rule generations and extract high quality rules. In [19] twelve diﬀerent rule quality measures were studied and compared through the ELEM2 [18] system on their classiﬁcation accuracies. The measures include empirical measures, statistical measures and measures from information theory.

3

Rule Evaluation on Knowledge Discovery

In this section, we ﬁrst examine a current rough set knowledge discovery system, and suggest the importance of rule evaluations. We propose rule evaluation approaches and their functions in knowledge discovery systems. 3.1

Analyzing RSES – Rough Set Exploration System

We take the RSES [8] system as an example system, and study in more detail of the role of rule evaluations. We show that current systems are limited with regard to rule evaluation, and we emphasize the importance of rule evaluation in current knowledge discovery systems.

158

J. Li, P. Pattaraintakorn, and N. Cercone

RSES (Rough Set Exploration System) is a well developed knowledge discovery system focusing on data analysis and classiﬁcation tasks, which is currently under development. Figure 1 shows a use of the system on a heart disease data set for classiﬁcation rule generation.

Fig. 1. Using Rough Set Exploration System on heart data

The data input to RSES is in the form of decision table T = (C, D), where C is the condition attribute set and D is the decision attribute set. Preprocessing is conducted once the data is imported to the system, during which stage the missing attribute values are handled and discretization is performed if necessary as well. Reducts are then generated, classiﬁcation rules based on the reducts are extracted. RSES provides four approaches on processing missing attribute values, such as removing data records with missing values, assigning the most common values of the missing attribute within the same decision class and without the same decision class, and considering missing attribute values as a special value of the attribute [8]. These approaches are used during the data preprocessing stage in the system. Although these approaches are fast and can be directly applied in the data, they lack the ability of preserving the semantic meanings of the original data set. Missing values may be assigned, however, the ﬁlled values may not be able to fully represent what is missing in the data. RSES provides rule postprocessing, which are “rule ﬁlter”, “rule shorten” and “rule generalize”. “Rule ﬁlter” removes from the rule set rules that do not satisfy certain support. “Rule shorten” shortens the length of the rules according to certain parameters [8]. “Rule generalization” generalizes rules according to a system provided parameter on the precision level. Although these rule postprocessing approaches provide an easier presentation of all the rule sets, these approaches do not provide ways to evaluate which rules are more interesting,

Rule Evaluations, Attributes, and Rough Sets: Extension and a Case Study

159

and which rules have higher quality. These functions cannot provide a rank of rules according to a rule’s signiﬁcance to the users. 3.2

Enhanced Knowledge Discovery System Based on Rough Sets

We present a rough set based knowledge discovery system, as shown in Figure 2.

Original Data

ItemRSFit on predicting missing values

Preprocessing

Processed Data

Attribute Selection (Rough Sets)

Rule Generation

Defining Rule Template

Rule Importance Measures Rule Evaluation Reduct Rules

Knowledge

Knowledge Representation visualization

Fig. 2. The Knowledge Discovery Based on Rough Sets Theory

In this general purpose knowledge discovery system, data from diﬀerent application domains are ﬁrst imported into the system. Preprocessings including missing attribute values processing, discretization, are conducted in this stage. After the data is preprocessed attribute selections are conducted. Depending on the output, diﬀerent attribute selection approaches can be applied here. Rule generation algorithms extract rules. After the rule sets are obtained, the important postprocessing - rule evaluations are performed in this stage. Rules are ﬁnally represented, possibly visualized in a certain format, as knowledge to the end users. We introduce three approaches integrated into this general purpose KDD system as shown in Figure 2. The ﬁrst approach ItemRSFit is used in the data preprocessing stage. The second approach, rule importance measure is used to

160

J. Li, P. Pattaraintakorn, and N. Cercone

rank rules during the rule evaluation process. The third approach of extracting reduct rules is also used during the rule evaluation stage. We will elaborate these approaches in the following. I. Predicting missing attribute values based on Frequent Itemset. ItemRSFit approach is a recently developed approach on predicting missing attribute values based on association rules algorithm and rough sets theory. It has been shown on both large scale real world data set and UCI machine learning data sets on the improved prediction accuracies. ItemRSFit approach is an integration of two other approaches from association rule algorithm and rough sets theory. Priori to the association rule generation, frequent itemsets are generated based on the item-item relations from the large data set according to a certain support. Thus the frequent itemsets of a data set represent strong correlations between diﬀerent items, and the itemsets represent probabilities for one or more items existing together in the current transaction. When considering a certain data set as a transaction data set, the implications from frequent itemsets can be used to ﬁnd to which attribute value the missing attribute is strongly connected. Thus the frequent itemset can be used for predicting the missing values. We call this approach “itemset-approach” for prediction. The larger the frequent itemsets used for the prediction, the more information from the data set itself will be available for prediction, hence the higher the accuracy will be obtained. However, generating frequent itemset for large data set is time-consuming. Although itemsets with higher support need less computation time, they restrict item-item relationships, therefore not all the missing values can be predicted. In order to balance the tradeoﬀ between computation time and the percentage of the applicable prediction, another approach must be taken into consideration. A reduct contains a subset of condition attributes that are suﬃcient enough to represent the whole data set. The intersection of all the possible reduct is the core. Therefore the attributes contained in the reduct or core are more important and representative than the rest of the attributes. Thus by examining only attributes within the same core or reduct to ﬁnd the similar attribute value pairs for the data instance containing the missing attribute values, we can assign the most relevant value for the missing attribute. Since this method only considers a subset of the data set, which is either the core or the reduct, the prediction is quite fast. This approach “RSFit” is recently proposed in [20], and it is an alternative approach designed for fast prediction. It can be used to predict missing attributes that cannot be predicted by the frequent itemset. We integrate the prediction based on frequent itemset and RSFit approach into a new approach ItemRSFit to predict missing attribute values. Frequent itemsets are used to predict missing values ﬁrst, and RSFit approach is used to predict the rest of the missing values that cannot be predicted by the frequent itemsets. This integrated approach can predict missing values from the data itself, therefore less noise is brought into the original data. The details on the ItemRSFit approach is presented in [21].

Rule Evaluations, Attributes, and Rough Sets: Extension and a Case Study

161

Properly processed data can improve the quality of the generated knowledge. Therefore the ItemRSFit approach is used in this system at the preprocessing stage. It helps to preserve the qualities of the original input data to this system, thus facilitate the rule evaluation process. II. Rule Importance Measures. Rule importance measure [2] is developed to provide a diverse rank of how important the association rules are, although this approach can also be applied to rules generated by other rule discovery algorithms. Association rules algorithm can be applied on this transaction data set to generate rules, which have condition attributes on the antecedent part and decision attributes on the consequent part of the rules. Rules generated from diﬀerent reduct sets can contain diﬀerent representative information. If only one reduct set is being considered to generate rules, other important information might be omitted. Using multiple reducts, some rules will be generated more frequently than other rules. We consider the rules that are generated more frequently more important. The Rule Importance is deﬁned to be important by the following deﬁnition. Deﬁnition 5. If a rule is generated more frequently across diﬀerent rule sets, we say this rule is more important than rules generated less frequently across those same rule sets. Rule importance measure is deﬁned as follows, Deﬁnition 6 Number of times a rule appears in all the generated rules from the reduct sets . Rule Importance Measure = Number of reduct sets The deﬁnition of the rule importance measure can be elaborated by Eq. 3. Let n be the number of reducts generated from the decision table T (C, D). Let RuleSets be the n rule sets generated based on the n reducts. rulesetj ∈ RuleSets (1 ≤ j ≤ n) denotes individual rule sets containing rules generated based on reducts. rulei (1 ≤ i ≤ m) denotes the individual rule from RuleSets. RIMi represents the rule importance measure for the individual rule. Thus the rule importance measures can be computed by the following |{rulesetj ∈ RuleSets|rulei ∈ rulesetj }| . (3) n The details of how to use rule importance measures can be found in [2]. Rule importance measure can be integrated into the current rough sets based knowledge discovery system to be used during the rule evaluation process. A list of ranked important rules can therefore be presented with their rule importance measures to facilitate the understanding of the extracted knowledge. RIMi =

162

J. Li, P. Pattaraintakorn, and N. Cercone

III. Extracting Reduct Rules. In [3] a method of discovering and ranking important rules by considering rules as attributes was introduced. The motivation comes from the concept of reduct. A reduct of a decision table contains attributes that can fully represent the original knowledge. If a reduct is given, rules extracted based on this reduct are representative of the original decision table. Can we take advantage of the concept of a reduct to discover important rules? We construct a new decision table Am×(n+1) , where each record from the original decision table u0 , u1 , ..., um−1 are the rows, and the columns of this new table consists of Rule0 , Rule1, ..., Rulen−1 and the decision attribute. We say a rule can be applied to a record in the decision table if both the antecedent and the consequent of the rule appear together in the record. For each Rulej (j ∈ [0, ..., n−1]), we assign 1 to cell A[i, j] (i ∈ [0, ..., m−1]) if the rule Rulej can be applied to the record ui . We set 0 to A[i, j] otherwise. The decision attribute A[i, n] (i ∈ [0, ..., m − 1]) remains the same as the original values of the decision attribute in the original decision table. Eq. 4 shows the conditions for the value assignments of the new decision table. ⎧ ⎨ 1, if j < n and Rulej can be applied to ui A[i, j] = 0, if j < n and Rulej cannot be applied to ui ⎩ di , if j = n and di is the corresponding decision attributes for ui (4) where i ∈ [0, ..., m − 1], j ∈ [0, ..., n − 1]. We further deﬁne Reduct Rule Set and Core Rule Set. Deﬁnition 7. Reduct Rule Set. We deﬁne a reduct generated from the new decision table A as Reduct Rule Set. A Reduct Rule Set contains Reduct Rules. The Reduct Rules are representative rules that can fully describe the decision attribute. Deﬁnition 8. Core Rule Set. We deﬁne the intersection of all the Reduct Rule Sets generated from this new decision table A as Core Rule Set. A Core Rule Set contains Core Rules. The Core Rules are contained in every Reduct Rule Set. By considering rules as attributes, reducts generated from the new decision table contain all the important attributes, which represent the important rules generated from the original data set; and it excludes the less important attributes. Core attributes from the new decision table A contain the most important attributes, which represent the most important rules. Other Enhancements. The three approaches discussed in our research have shown to eﬀectively evaluate rules. There are other techniques that can be used along with these approaches in Figure 2. For example, during the rule generation

Rule Evaluations, Attributes, and Rough Sets: Extension and a Case Study

163

process, properly deﬁned rule templates can not only reduce the computation of rule generations, but it also ensures high quality rules, or interesting rules generated according to the application purposes. Important attributes, such as probe attributes (discussed in Section 5) can be deﬁned in the data preprocessing stage for generating rules containing such attributes for generating expected rules. Our motivation is, proposing approaches to enhance the current knowledge discovery system, to facilitate the knowledge discovery process on discovering more interesting and higher quality rules.

4

Experiments

We demonstrate, through a series of experiments, that systems improved by the proposed rule evaluation approaches can help humans discover and understand more important rules. 4.1

Specifying Rule Templates

Apriori association rules algorithm is used to generate rules. Because our interest is to make decisions or recommendations based on the condition attributes, we are looking for rules with only decision attributes on the consequent part. Therefore, we specify the following 2 rule templates to extract rules we want as shown by Template 5, and to subsume rules as shown by Template 6. Attribute1 , Attribute2 , . . . , Attributen → DecisionAttribute

(5)

Template 5 speciﬁes only decision attributes can be on the consequent part of a rule, and Attribute1 , Attribute2 , . . . , Attributen lead to a decision of DecisionAttribute. We specify the rules to be removed or subsumed using Template 6. For example, given rule Attribute1 , Attribute2 → DecisionAttribute

(6)

the following rules Attribute1 , Attribute2 , Attribute3 → DecisionAttribute

(7)

Attribute1 , Attribute2 , Attribute6 → DecisionAttribute

(8)

can be removed because they are subsumed by Template 6. Take the geriatric care data in Table 1 as an example, in the rule set, a rule shown as Eq. 9 exists SeriousChestP roblem → Death

(9)

the following rule is removed because it is subsumed. SeriousChestP roblem, T akeM edicineP roblem → Death

(10)

164

4.2

J. Li, P. Pattaraintakorn, and N. Cercone

Geriatric Care Data Set

We perform experiments on a geriatric care data set as shown in Table 1. This data set is an actual data set from Dalhousie University Faculty of Medicine to determine the survival status of a patient giving all the symptoms he or she shows. The data set contains 8, 547 patient records with 44 symptoms and their survival status (dead or alive). We use survival status as the decision attribute, and the 44 symptoms of a patient as condition attributes, which includes education level, the eyesight, hearing, be able to walk, be able to manage his/her own meals, live alone, cough, high blood pressure, heart problem, cough, gender, the age of the patient at investigation and so on.1 There is no missing value in this data set. There are 12 inconsistent data entries in the medical data set. After removing these instances, the data contains 8, 535 records. 2 Table 1. Geriatric Care Data Set edulevel eyesight hearing health 0.6364 0.25 0.50 0.25 0.7273 0.50 0.25 0.25 0.9091 0.25 0.50 0.00 0.5455 0.25 0.25 0.50 0.4545 0.25 0.25 0.25 0.2727 0.00 0.00 0.25 0.0000 0.25 0.25 0.25 0.8182 0.00 0.50 0.00 ... ... ... ...

trouble livealone cough hbp heart 0.00 0.00 0.00 0.00 0.00 0.50 0.00 0.00 0.00 0.00 0.00 0.00 0.00 1.00 1.00 0.00 1.00 1.00 0.00 0.00 0.00 1.00 0.00 1.00 0.00 0.50 1.00 0.00 1.00 0.00 0.00 0.00 0.00 0.00 1.00 0.00 0.00 0.00 1.00 0.00 ... ... ... ... ...

stroke 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 ...

... ... ... ... ... ... ... ... ... ...

sex livedead 1 0 2 0 1 0 2 0 2 0 2 0 1 0 2 0 ... ...

The ItemRSFit approach is implemented by Perl and the experiments are conducted on Sun Fire V880, four 900Mhz UltraSPARC III processors. We use apriori frequent itemset generation [23] to generate frequent 5-itemset. The core generation in RSFit approach is implemented with Perl combining the SQL queries accessing MySQL (version 4.0.12). ROSETTA software [7] is used for reduct generation. 4.3

Experiments on Predicting Missing Attribute Values

In order to show the ItemRSFit approach obtains better prediction accuracy than the existing approach (i.e., RSFit), we perform the experiments on the geriatric care data set by randomly selecting 150 missing values from the original data. We then apply both RSFit approach and ItemRSFit approach on predicting missing values, and compare the accuracy of the prediction. Figure 3 demonstrates the comparison predicting abilities between RSFit and ItemRSFit approaches. We can see from the ﬁgure that the smaller the support is, the more accurate the 1 2

Refer to [22] for details about this data set. Notice from our previous experiments that core generation algorithm cannot return correct core attributes when the data set contains inconsistent data entries.

Rule Evaluations, Attributes, and Rough Sets: Extension and a Case Study

165

100 RSFit ItemRSFit 95

Accuracy (%)

90

85

80

75

70 10

20

30

40

50 Support (%)

60

70

80

90

Fig. 3. Accuracy Comparisons for Geriatric Care Data with 150 Missing Attribute Values

prediction of the missing attribute values for the ItemRSFit approach obtains; whereas for the RSFit approach, the accuracy remains the same as the value of support gets smaller; and the accuracy obtained by RSFit is always lower than the ItemRSFit approach. This result demonstrates that frequent itemsets as knowledge base can be eﬀectively applied for predicting missing attribute values. 4.4

Experiments on Rule Importance Measure

In our experiment, we use the genetic algorithm to generate multiple reduct sets with the option of full discernibility. The apriori algorithm [23] for large item sets generation. The core attributes for this data set are eartrouble, livealone, heart, highbloodpressure, eyetrouble, hearing, sex, health, educationlevel, chest, housework, diabetes, dental, studyage. Table 2. Reduct Sets for the Geriatric Care Data Set after Preprocessing No. Reduct Sets 1 {edulevel,eyesight,hearing,shopping,housewk,health,trouble,livealone, cough,sneeze,hbp,heart,arthriti,eyetroub,eartroub,dental, chest,kidney,diabetes,feet,nerves,skin,studyage,sex} 2 {edulevel,eyesight,hearing,phoneuse,meal,housewk,health,trouble,livealon, cough,sneeze,hbp,heart,arthriti,evetroub,eartroub,dental, chest,bladder,diabetes,feet,nerves,skin,studyage,sex} ... ... 86 {edulevel,eyesight,hearing,shopping,meal,housewk,takemed,health, trouble,livealone,cough,tired,sneeze,hbp,heart,stroke,arthriti, eyetroub,eartroub,dental,chest,stomach,kidney,bladder,diabetes, feet,fracture,studyage,sex}

166

J. Li, P. Pattaraintakorn, and N. Cercone Table 3. Rule Importance for the Geriatric Care Data No. 0 1 2 3 4 5 ... 10 ... 216 217

Selected Rules Rule Importance SeriousHeartProblem → Death 100% SeriousChestProblem → Death 100% SeriousHearingProblem, HavingDiabetes → Death 100% SeriousEarTrouble → Death 100% SeriousEyeTrouble → Death 100% Sex Female → Death 100% ... ... Livealone, HavingDiabetes, NerveProblem → Death 95.35% ... ... SeriousHearingProblem, ProblemUsePhone → Death 1.16% TakeMedicineProblem, NerveProblem → Death 1.16%

Table 2 shows selected reduct sets among the 86 reducts generated by ROSETTA. All of these reducts contain the core attributes. For each reduct set, association rules are generated with support = 30%, conf idence = 80%. 3 218 unique rules are generated over these 86 reducts. These rules as well as their rule importance are shown in Table 3. Among these 218 rules, 87 rules have rule importance of no less than 50% , 8 of which have rule importance of 100%. All the rules with rule importance of 100% contain only core attributes. 4.5

Experiments on Generating Reduct Rules

The new decision table A8535×219 is constructed by using the 218 rules 4 as condition attributes, and the original decision attribute as the decision attribute. Note that after reconstructing the decision table, we must check for inconsistency again before generating reduct rules for this table. After removing the inconsistent data records, there are 5709 records left in the new decision table. The core rule set is empty. We use Johnson’s reduct generation algorithm on this table A5709×219 and the reduct rule set is {Rule0, Rule1, Rule3 , Rule5, Rule19 , Rule173}. We show these rules in Table 4. From Table 4 we can see that the reduct rule sets contain 6 rules. There are 4 rules judged to be the most important. The rule importance for Rule0 , Rule1, Rule3 and Rule5 are all 100%. The Rule19 has the importance of 82.56%, which is more important among the 218 rules. 3

4

Note that the value of support and conﬁdence can be adjusted to generate as many or as few rules as required. There are 1615 rules generated by apriori algorithm from the original data set with support = 30%, conf idence = 80%, after applying the rule template. We can circumvent problems inherent in considering all 1615 generated rules using the 218 unique rules that are derived from the 86 reducts obtained by ROSETTA’s genetic algorithm.

Rule Evaluations, Attributes, and Rough Sets: Extension and a Case Study

167

Table 4. Reduct Rules for the Geriatric Care Data No. in Table 3 0 1 3 5 19 173

5

Reduct Rules

Rule Importance SeriousHeartProblem → Death 100% SeriousChestProblem → Death 100% SeriousEarTrouble → Death 100% Sex Female → Death 100% Livealon, OftenSneeze, DentalProblems, HavingDiabetes → Death 82.56% ProblemHandleYourOwnMoney → Death 27.91%

A Case Study: Survival Analysis in HYRIS

This section provides a case study to illustrate how a rough sets based KDD system provides a useful mechanism for analyzing and distilling essential attributes and rules from survival data, and evaluates the generated rules in postprocessing for survival analysis. Use of medical survival data challenges researchers because of the size of data sets and vagaries of their structures. Among prognostic modeling techniques that induce models from medical data, survival analysis warrants special treatment in the type of data required and its modeling. Data required for medical analysis includes demographic, symptoms, laboratory tests and treatment information. Special features for survival data are the events of interest, censoring, followup time and survival time speciﬁc for each type of disease. Such data demands powerful analytical models for survival analysis. The studies applying rough sets to survival analysis related to our work are [24][25]. They illustrated rough sets contribution to a medical expert system for throat cancer patients successfully. Rough sets and decision trees have been used to study kidney dialysis patients’ survival [26]. HYRIS (HYbrid Rough sets Intelligent System) [27][28] is designed speciﬁcally to consider survival analysis with several statistical approaches. HYRIS uses the CDispro algorithm from a previous study [4]. HYRIS successively derives dispensable, probe attribute, reduct and probe reduct together with life time table and Kaplan-Meier survival curves [27]. In addition to survival analysis, HYRIS provides a general data analysis and decision rule generation and evaluation [27] as well. HYRIS Case Study I. HYRIS is able to analyze censor variable and survival time attributes that are a speciality for survival analysis. Given the survival data set, the system can identify the covariant levels of particular attributes according to rough sets and several statistical approaches. The Kaplan-Meier method, hazard function, hypothesis testing, log-rank, Brewslow, Tarone-Ware tests, p-value and CDispro [4]. incorporate the rough sets framework to generate core, dispensable attributes, probe attribute, reducts, and probe reducts which are the informative attributes. Consequently, the rules are derived and validated with ELEM2 [18].

168

J. Li, P. Pattaraintakorn, and N. Cercone

We demonstrate the utility of HYRIS by investigating a particular problem using both actual and benchmark medical data sets: geriatric data, melanoma data [29], pneumonia data [30] and primary biliary cirrhosis data (PBC) [31]. For the geriatric care data set, time lived in months are used as the survival time target function. Inconsistent records are removed. Data is discretized according to percentile groups. The age is used as probe attribute, the reducts and probe reducts are generated successfully. The rules generated for geriatric data care are the decision rules for predicting survival time. Note that in Section 4 the experimental results showing important rules are used to predict the survival status, not the survival time. Two sample survival prediction rules out of 1,600 rules of geriatric care data set (when considering the probe attribute {ExperienceDiabetes}) generated from HYRIS are provided as follows: Rule 1: UnHealthy, SevereHearingDamage, NerveProblem, FootProblem, SeriousHeartProblem, DentalDisease, StomachDisease, HighBloodPressure, ExperienceDiabetes → SurvivalTime = 7-18 months. Rule 2: FemalePatient, LowEducationLevel, EyesightProblemLowToSeriousType, HealthProblemFromLowToSeriousType, HearQuiteWell, DoNotHaveDiabetesExperience, EasilyTired, FootProblem, → SurvivalTime = 56-73 months. When comparing the accuracy of rules that were generated from original attributes and those generated from reducts, the accuracy of all data sets range between 83.7851%–90.5686%. Rule performance outcomes are improved signiﬁcantly as reported in [4]. HYRIS Case Study II. HYRIS can accomplish preprocessing, learning and model construction and broaden further to use in rule evaluation and postprocessing. We continue a series of studies in [28]. In this case study, we propose an alternative approach for decision rule learning with rough sets theory in the postprocessing step called ROSERULE - Rough Sets Rule Reducts Learning Algorithm. ROSERULE learned and analyzed from the rule set to generate rule reducts which can be used to reduce the number of the rules. Results imply a reduced number of rules that successfully preserve the original classiﬁcation. The rule numbers of geriatric data set reduced from 1,600 to 1,150, melanoma data set reduced from 16 to 15, pneumonia data set reduced from 606 to 42 and PBC data set reduced from 83 to 72. At the same time, the prediction accuracy is preserved for all data sets.

6

Conclusion

We study the work of rough sets based rule evaluations on knowledge discovery system. We propose solutions to the challenging problems brought by large real world data sets, such as the existence of missing values and analyzing huge

Rule Evaluations, Attributes, and Rough Sets: Extension and a Case Study

169

amount of generated rules manually. Three rough set based approaches to enhance the current KDD systems on rule evaluations are introduced. The ItemRSFit approach is used to predict missing attribute values using frequent itemset as a knowledge base. Complete data can be obtained using this approach. The rule importance measure provides a ranking of how important is a rule. Finally, the reduct rules are extracted using the concept of reduct by considering rules as condition attributes in a decision table. Experimental results on a real world geriatric care data set demonstrate the utilities of applying rough sets based rule evaluations to enhance current KDD systems. A case study of a recent knowledge discovery system shows the applications of approaches which have been incorporated into HYRIS with an emphasis on survival analysis.

Acknowledgements This research is supported by the Natural Sciences and Engineering Research Council of Canada (NSERC). The research of Puntip Pattraintakorn has also been supported by a grant from the King Mongkut’s Institute of Technology Ladkrabang (KMITL) research fund, Thailand. We would also like to thank Arnold Mitnitski from Dalhousie University for providing the geriatric care data set, and the anonymous reviewers for their helpful comments.

References 1. Pawlak, Z.: Rough Sets. In Theoretical Aspects of Reasoning about Data. Kluwer Academic Publishers (1991) 2. Li, J. and Cercone, N.: Introducing A Rule Importance Measure. Transactions on Rough Sets, Springer LNCS, vol 5 (2006) 3. Li, J., Cercone, N.: Discovering and Ranking Important Rules. In Proceedings of IEEE International Conference on Granular Computing, vol 2, Beijing China 25-27 July (2005) 506–511 4. Pattaraintakorn, P., Cercone, N., Naruedomkul, K.: Hybrid Intelligent Systems: Selecting Attributes for Soft-Computing Analysis. In Proc. of the 29th Annual International Computer Software and Applications Conference (COMPSAC), vol 2 (2005) 319–325 5. Kryszkiewicz, M., Rybinski, H.: Finding Reducts in Composed Information Systems, Rough Sets, Fuzzy Sets Knowldege Discovery. In W.P. Ziarko (Ed.), Proceedings of the International Workshop on Rough Sets, Knowledge Discovery, Heidelberg/Berlin: Springer-Verlag (1994) 261–273 6. Jan Bazan, Hung Son Nguyen, Sinh Hoa Nguyen, Piotr Synak, and Jakub Wroblewski.: Rough set algorithms in classiﬁcation problems. Rough Set Methods and Applications: New Developments in Knowledge Discovery in Information Systems, volume 56 of Studies in Fuzziness and Soft Computing, pages 49-88. Physica-Verlag, Heidelberg, Germany (2000). 7. Øhrn, A.: Discernibility and Rough Sets in Medicine: Tools and Applications. PhD Thesis, Department of Computer and Information Science, Norwegian University of Science and Technology, Trondheim, Norway. (1999) 8. RSES 2.2 User’s Guide. Warsaw University. http://logic.mimuw.edu.pl/∼rses/

170

J. Li, P. Pattaraintakorn, and N. Cercone

9. Predki, B., Wilk, Sz.: Rough Set Based Data Exploration Using ROSE System. In: Foundations of Intelligent Systems. Ras, Z. W., Skowron, A., Eds, LNAI 1609, Springer-Verlag, Berlin (1999) 172-180 10. Chouchoulas, A. and Shen, Q.: Rough Set-Aided Keyword Reduction For Text Categorization. Applied Artiﬁcial Intelligence, vol 15 (2001) 843–873 11. Hu, X., Lin, T., Han, J.: A New Rough Sets Model Based on Database Systems. Fundamenta Informaticae 59 no.2-3 (2004) 135–152 12. Freeman, R. L., Grzymala-Busse, J. W., Riﬀel, L. A., Schroeder, S. R.: Analyzing the Relation Between Heart Rate, Problem Behavior, and Environmental Events Using Data Mining System LERS. In 14th IEEE Symposium on Computer-Based Medical Systems (CBMS’01) (2001) 13. Ivo, D., Gunther, G.: The Rough Set Engine GROBIAN. In Proc. of the 15th IMACS World Congress, vol 4, Berlin, August (1997) 14. Hu, T., Shan, N., Cercone, N. and Ziarko, W.: DBROUGH: A Rough Set Based Knowledge Discovery System, Proc. of the 8th International Symposium on Methodologies for Intelligent System, LNAI 869, Spring Verlag (1994) 386–395 15. Hilderman, R. and Hamilton, H.: Knowledge discovery and interestingness measures: A survey. Technical Report 99-04, Department of Computer Science, University of Regina, October (1999) 16. Pang-Ning Tan and Vipin Kumar and Jaideep Srivastava: Selecting the right interestingness measure for association patterns. Processings of SIGKDD. (2002) 32–41 17. Bruha, Ivan: Quality of Decision Rules: Deﬁnitions and Classiﬁcation Schemes for Multiple Rules. In Machine Learning and Statistics, The Interface, Edited by G. Nakh aeizadeh and C. C. Taylor. John Wiley & Sons, Inc. (1997) 107–131 18. An, A. and Cercone, N.: ELEM2: A Learning System for More Accurate Classiﬁcations. In: Proceedings of Canadian Conference on AI (1998) 426–441 19. An, A. and Cercone, N.: Rule Quality Measures for Rule Induction Systems: Description and Evaluation. Computational Intelligence. 17-3 (2001) 409–424. 20. Li, J. and Cercone, N.: Assigning Missing Attribute Values Based on Rough Sets Theory. In Proceedings of IEEE Granular Computing, Atlanta, USA. (2006) 21. Li, J. and Cercone, N.: Predicting Missing Attribute Values based on Frequent Itemset and RSFit. Technical Report, CS-2006-13, School of Computer Science, University of Waterloo (2006) 22. Li, J. and Cercone, N.: Empirical Analysis on the Geriatric Care Data Set Using Rough Sets Theory. Technical Report, CS-2005-05, School of Computer Science, University of Waterloo (2005) 23. Borgelt, C.: Eﬃcient Implementations of Apriori and Eclat. Proceedings of the FIMI’03 Workshop on Frequent Itemset Mining Implementations. In: CEUR Workshop Proceedings (2003) 1613-0073 http://CEUR-WS.org/Vol-90/borgelt.pdf 24. Bazan, J., Osmolski, A., Skowron, A., Slezak, D., Szczuka, M., Wroblewski, J.: Rough Set Approach to the Survival Analysis. In Alpigini, J. J., et al. (Eds.): The Third International Conference on Rough Sets and Current Trends in Computing (RSCTC), Proceedings, LNAI 2475, Springer-Verlag Berlin Heidelberg (2002) 522-529 25. Bazan, J., Skowron, A., Slezak, D., Wroblewski, J.: Searching for the Complex Decision Reducts: The Case Study of the Survival Analysis, LNAI 2871, SpringerVerlag, Berlin Heidelberg (2003) 160-168 26. A. Kusiak, B. Dixon, S. Shah: Predicting Survival Time for kidney Dialysis Patients: A Data Mining Approach, Computers in Biology and Medicine 35 (2005) 311–327

Rule Evaluations, Attributes, and Rough Sets: Extension and a Case Study

171

27. Pattaraintakorn, P., Cercone, N., Naruedomkul, K.: Selecting Attributes for SoftComputing Analysis in Hybrid Intelligent Systems. In D. Slezak et al. (Eds.): Rough Sets, Fuzzy Sets, Data Mining, and Granular Computing 10th International Conference (RSFDGrC), Proceedings, Part II Series: Lecture Notes in Computer Science, Subseries: LNAI 3642, Springer-Verlag, Berlin, Heidelberg (2005) 698–708 28. Pattaraintakorn, P., Cercone, N., Naruedomkul, K.: Rule Analysis with Rough Sets Theory, The IEEE International Conference on Granular Computing, Atlanta, USA (2006) 29. Elisa, L.T., John, W.W.: Statistical methods for survival data analysis, 3rd edn. New York: John Wiley and Sons (2003) 30. Klein, J.P., Moeschberger, M.L.: Survival analysis: techniques for censored and truncated data, 2nd edn. Berlin: Springer (2003) 31. Newman, D.J., Hettich, S., Blake, C.L. and Merz, C.J.: UCI Repository of machine learning databases. University of California, Irvine, Department of Information and Computer Seiences (1998) http://www.ics.uci.edu/∼mlearn/MLRepository.html

The Impact of Rough Set Research in China: In Commemoration of Professor Zdzislaw Pawlak Qing Liu1,2 and Hui Sun1 1

Department of Computer Science & Technology Nanchang Institute of Technology, Nanchang 330099, China qliu− [email protected] 2 Department of Computer Science & Technology Nanchang University, Nanchang 330029, China

This article is dedicated to the creative genius Zdzislaw Pawlak for his contribution to the theoretical development of science and technology in China. His distinguished discovery of Rough Set Theory is a formal theory which is well suited for uncertainty computing to analyze imprecise, uncertain or incomplete information of data. Inspired by his work scientists and engineers in China has developed many theories and applications in various science and technology ﬁelds. For instance, J.H.Dai studied the theories of Rough Algebras and Axiom Problem of Rough 3-Valued Algebras [1, 2]. G.L.Liu studied the Rough Sets over Fuzzy Lattices [3, 4]. D.W.Pei studied the Generalized Model of Fuzzy Rough Sets [5]. W.Z.Wu Studied the On Random Rough Sets [6]. D.Q.Miao studied the Rough Group and Their Properties [7]. These are part of their recent research results related to rough set theory. As a matter of fact, there are still many researchers working in the ﬁeld of rough sets in China, who have proposed many creative results for last few years. These results are not listed one by one in this short commemorative article. We will try to review all the ”Rough Set” researchers and their research results in the appeared next article. In this article, we present only a recent partial research results of the authors. Based on Rough Logic and Decision Logic deﬁned by Pawlak [8, 9], ﬁrst author Liu has proposed a rough logic in a given information system [10]. Inﬂuenced by the concept of granular language proposed by Skowron [11], the granular logic deﬁned by Polkowski [12], and the work of Lin, Yao in [13, 14], we also have deﬁned a granular logic by applying the semantics of rough logical formulas in a given information system, and have created the deductive systems as well as have discussed many properties in [15, 16]. The proposed granular logic is a set which consists of granular formulas of form m(F ), where F is the rough logical formula in the given information system. It is used as individual variable of semantic function symbol m, so, we call it a paradigm of higher order logic. Truth values of granular formula of form m(F ) in the logic have two types. One is the function value, which is the meaning of rough logical formula, a subset in U; Another is the truth value of a degree, which is equal to a degree of meaning of the formula to close to universe U of objects. Pawlak introduced the concept of rough truth in 1987, assuming that a formula is roughly true in a J.F. Peters et al. (Eds.): Transactions on Rough Sets VI, LNCS 4374, pp. 172–175, 2007. c Springer-Verlag Berlin Heidelberg 2007

The Impact of Rough Set Research in China

173

given information system if and only if the upper approximation of meaning of the formula is equal to the whole universe. So, our approach extends Pawlak’s approach [9]. Skolem clause form, resolution principles, λ-resolution strategies and deductive reasoning of the granular logic are also discussed in the next article. These logic systems should be an extension of Rough Logic proposed by Pawlak in 1987 [9]. The practicability of the higher order logic will oﬀer the new idea for studying classical logic. It could also be a theoretical tool for studying granular computing. Based on reference [23], we further propose to use rough set theory to ﬁnd out minimal approximate solution set in the approximate solution space of diﬀerential equations and functional variation problems in mechanics. This could be a new studying style in rough set applications. The signiﬁcance and future development direction of the proposed Rough Sets are described. Any undeﬁnable subset on the universe of a given information system is constructed into precise deﬁnable lower and upper approximations via indiscernibility relation. Hence, complex and diﬃcult problems on undeﬁnable sets are resolved or transformed into precise deﬁnable lower and upper approximations [8, 17]. This is one of a great contribution of Pawlak’s Rough Sets Theory. Successful applications of rough sets in many ﬁelds oﬀer a lot of new idea of studying granular computing, which also promote the development of granular computing. Founder of predicate logic, G. Frege proposed the vague boundary in 1904, that is, how to compute the number of elements on vague boundary [18, 19]. Many mathematicians and computer scientists have made hard eﬀorts on the question. L. A. Zadeh proposed the Fuzzy Sets (FS) in 1965. He attempted to solve the computability of Frege’s vague boundary by Fuzzy Set concept [20]. Unfortunately, Fuzzy Sets are not mechanically computable, that is, the formula of exact describing for the fuzzy concept hasn’t been given. Therefore, the number of elements on the vague boundary could not be computed by exact formula. For example, the membership μ in Fuzzy Sets and fuzzy operator λ in operator fuzzy logic [21], could not be computed exactly. Z.Pawlak proposed the Rough Sets (RS) in 1982 for computing Frege’s vague boundary [8, 17], and the number of elements on the vague boundary could be exactly computed by it. Rough Set Theory is a new tool to deal with incomplete and uncertainty problems. In the ﬁeld of computer applications nowadays, this theory is no doubt a challenge to other uncertainty theories. Since rough set theory is one of the most important, newest theories and with the rapid development, it is also very important in artiﬁcial intelligence and cognitive science. Especially methods,which are based on rough set theory alone or are in combination with other approaches, have been used over a wide range of applications in many areas. More and more people in the region of China are attracted by them. In this article, we also present state of art of RS in China. Since May, 2001 Professor Pawlak being invited to China made his keynote speech in a conference and gave many invited talks among universities, the research of RS have been

174

Q. Liu and H. Sun

rapidly developed in China. Rough Set Theory has been used for important applications in human life, such as, data reduction, approximate classiﬁcation of data, management systems of business information, computing the average value and the standard deviation in quality test of products of statistics and so on. We developed successfully the Business Information Management Systems with rough set approach, which is a management for price of houses of some region [22]. We deﬁned an indiscernibility relation on numerical interval [a, b] by using ancient mathematical Golden Cut method and created rough sets by the deﬁned relation. Based on the rough set approach, we developed a ”Diagnosis Software of Blood Viscosity Syndrome on Hemorheology”, which is used to test Blood Viscosity chroma of patients. The systems has been applied in the clinic for many years. The medicine experts review that the diagnosis software is precursive, creative, scientiﬁc and practical [24]. In this article, we present the state of art of primary rough set research results and their applications in practice in China during last few years. Especially, in 2001, Pawlak was invited to China, his keynote speech had made a big inﬂuence on the development of science and technology of China. We would like to thank Pawlak for his fundamental and signiﬁcant contribution to the development of rough set research in China. We would like to thank the editor-in-chief Professor James F. Peters, Professor Andrej Skowron and Professor Ewa Orlowska for their kindness to let us publish article in this historical event to commemorate the great scientist Zdzislaw Pawlak for his contribution to the science and technology world. Thanks are also to the support of Natural Science Fund of China (NSFC-60173054). At last we would like to thank Dr. James Kuodo Huang (who is a IEEE member and have taught in the universities of USA for over 20 years) for his kind suggestions of English in this article. Still we would like to take the whole responsibility for any further errors made in this article.

References 1. Dai,J.H., Axiom Problem of Rough 3-Valued Algebras, The Proceedings of International Forum on Theory of GrC from Rough Set Perspective (IFTGrCRSP2006), Nanchang, China, Journal of Nanchang Institute of Technology, Vol.25, No.2, 2006, 48-51. 2. Dai,J.H., On Rough Algebras, Journal of software 16 (2005),1197-1204. 3. Liu,G.L., Rough Set Theory Over Fuzzy Lattices, The Proceedings of International Forum on Theory of GrC from Rough Set Perspective (IFTGrCRSP2006), Nanchang, China, Journal of Nanchang Institute of Technology, Vol.25, No.2, 2006, 51-51. 4. Liu,G.L., The Topological Structure of Rough Sets over Fuzzy Lattices, 2005 IEEE International Conference on Granualr Computing, Vol.I, Proceedings, Beijing, China, July 25-27,2005, 535-538. 5. Pei,D.W., A Generalized Model of Fuzzy Rough Sets, Int. J. General Systems 34 (5)2005, 603-613.

The Impact of Rough Set Research in China

175

6. Wu,W.Z., On Random Rough Sets, The Proceedings of International Forum on Theory of GrC from Rough Set Perspective (IFTGrCRSP2006), Nanchang, China, Journal of Nanchang Institute of Technology, Vol.25, No.2, 2006, 66-69. 7. Miao,D.Q., Rough Group, Rough Subgroup and their Properties, LNAI 3641, 10th International Conference, RSFDGrC2005, Regina, Canada, August/September 2005, Proceedings, Part I, 104-113. 8. Pawlak, Z., Rough Sets: Theoretical Aspects of Reasoning about Data, Kluwer Academic Publishers, Dordrecht , 1991. 9. Pawlak, Z., Rough Logic, Bulletin of the Polish Academy of Sciences, Technical Sciences, Vol.35, No.5-6, 1987, 253-259. 10. Liu, Q., Liu, S.H. and Zheng, F., Rough Logic and Applications in Data Reduction, Journal of Software, Vol.12, No.3, March 2001,415-419. (In Chinese). 11. Skowron, A., Toward Intelligent Systems: Calculi of Information Granules, Bulletin of International Rough Set Society, Vol.5, No.1/2, Japan, 2001,9-30. 12. Polkowski,L., A Calculus on Granules from Rough Inclusions in Information Systems, The Proceedings of International Forum on Theory of GrC from Rough Set Perspective (IFTGrCRSP2006), Nanchang, China, Journal of Nanchang Institute of Technology, Vol.25, No.2, 2006, 22-27. 13. Lin,T.Y., Liu,Q., First Order Rough Logic I: Approximate Reasoning via Rough Sets, Fundamenta Informaticae 2-3, 1996, 137-153. 14. Yao,Y.Y., Liu,Q., A Generalized Decision Logic in Interval-Set-Valued Information table, Lecture Notes in AI 1711, Springer-Verlag, Berlin, 1999, 285-294. 15. Liu,Q.,Sun,H., Theoretical Study of Granular Computing, LNAI 4062, The Proceedings of RSKT2006, by Springer, China, July 2006, 93-102. 16. Liu,Q. and Huang,Z.H., G-Logic and Its Resolution Reasoning, Chinese Journal of Computer, Vol.27,No.7, 2004, 865-873. (In Chinese). 17. Pawlak,Z., Rough Sets, Int. J. Inform. Comp. Sci., 11(1982), 341-356. 18. Frege,G., Grundgesetze der Arithmentic, In: Geach and Black (eds.) , Selection from the philosophical Writings of Gotlob Frege, Blackwei, Oxford 1970 19. Pawlak,Z., Rough Sets present State and Further Prospects, The Proceedings of Third International Workshop on Rough Sets and Soft Computing, Nov. 1012,1994, 72-76. 20. Zadeh,L.A., Fuzzy Sets, Information and Control, No.,8, 1965, 338-353. 21. Liu,X.H., Fuzzy Logic and Fuzzy Reasoning [M], Press. of Jilin University, Jilin, 1989,(In Chinese). 22. Liu,Q.,Sun H., Studying Direction of Granular Computing from Rough Set Perspective, Journal of Nanchang Institute of TechnologyVol. 25, No.3, 2006, 1-5. 23. Sun,H.,Liu,Q., The Research of Rough Sets in Normed Linear Space, LNAI 4259The Proceedings of RSCTC2006, by Springer, Japan, 8-11 Nov., 2006, 91-98. 24. Liu,Q., Jiang,F. and Deng,D.Y., Design and Implement for the Diagnosis Software of Blood Viscosity Syndrome Based on Hemorheology on GrC., Lecture Notes in Artiﬁcial Intelligence 2639, Springer-Verlag, 2003,413-420.

A Four-Valued Logic for Rough Set-Like Approximate Reasoning Jan Małuszy´nski1, Andrzej Szałas1,2 , and Aida Vit´oria3 1

Link¨oping University, Department of Computer and Information Science 581 83 Link¨oping, Sweden [email protected] 2 The University of Economics and Computer Science Olsztyn, Poland [email protected] 3 Dept. of Science and Technology, Link¨oping University S 601 74 Norrk¨oping, Sweden [email protected]

Abstract. This paper extends the basic rough set formalism introduced by Pawlak [1] to a rule-based knowledge representation language, called Rough Datalog, where rough sets are represented by predicates and described by finite sets of rules. The rules allow us to express background knowledge involving rough concepts and to reason in such a knowledge base. The semantics of the new language is based on a four-valued logic, where in addition to the usual values T RUE and FALSE, we also have the values B OUNDARY, representing uncertainty, and U NKNOWN corresponding to the lack of information. The semantics of our language is based on a truth ordering different from the one used in the well-known Belnap logic [2, 3] and we show why Belnap logic does not properly reflect natural intuitions related to our approach. The declarative semantics and operational semantics of the language are described. Finally, the paper outlines a query language for reasoning about rough concepts.

1 Introduction The seminal ideas of Pawlak [1, 4, 5, 6] on the treatment of imprecise and incomplete data opened a new area of research, where the notion of rough sets is used in theoretical studies as well as practical applications. Rough sets are constructed by means of approximations obtained by using elementary sets which partition a universe of considered objects. The assumption as to partitioning of the universe has been relaxed in many papers (see, e.g., [7, 8, 9, 10, 11, 12, 13, 14, 15]), however the Pawlak’s idea of approximations has remained the same. This paper extends the basic rough set formalism to a rule-based language, where rough sets are represented by predicates and are described by finite sets of rules. The rules allow one to express background knowledge concerning rough concepts and to reason in such a knowledge base. The new language is different from that proposed in [14, 15], where the rules described rough sets by combining their regions (lower approximation, upper approximation and boundary region). In contrast to the language J.F. Peters et al. (Eds.): Transactions on Rough Sets VI, LNCS 4374, pp. 176–190, 2007. c Springer-Verlag Berlin Heidelberg 2007

A Four-Valued Logic for Rough Set-Like Approximate Reasoning

177

described in this paper, the rules expressed in the language presented in [14, 15] refer explicitly to different regions of a rough set. Lifting the level of description makes the definitions easier to understand, also for the people not familiar with the technicalities of rough sets. The semantics of the new language is based on a four-valued logic, where in addition to the usual values T RUE and FALSE we have the values B OUNDARY representing uncertain/inconsistent information and U NKNOWN corresponding to the lack of information. As discussed in Section 3.2, the well-known four-valued Belnap logic [3, 2] does not properly reflect the natural intuitions related to our approach. We propose instead a slightly different truth ordering and use it, together with the standard knowledge ordering, for defining a declarative semantics of our language. By using the four-valued logic we propose, we are then able to deal with some important issues. First of all, we are able to provide a natural semantics for Datalog-like rules where negation can be used freely, both in the bodies and in the heads of rules. This, in previous approaches to various variants of negation, has always been problematic either due to the high computational complexity of queries or to a nonstandard semantics of negation, often leading to counterintuitive results (for an overview of different approaches to negation see, e.g., [16]). Our semantics reflects intuitions of fusing information from various independent sources. If all sources claim that a given fact is true (respectively, false) then we have an agreement and attach T RUE (respectively FALSE) to that fact. If information sources disagree in judgement of a fact, we attach to it the value B OUNDARY . If no source provides an information about a given fact, we then make it U NKNOWN. Second, we are able to import knowledge systems based on the classical logic without any changes and make them work directly within the rough framework. In such cases these systems would act as single information sources providing answers T RUE, FALSE, when queried about facts. Possible conflicting claims of different systems would then be solved by the same, uniform four-valued approach we propose. This might be useful in combining low level data sources, like classifiers as well as higher level expert systems. Third, one can import rough set-based systems, or systems supporting approximate reasoning, like for example, those described in [14, 15], or [17, 18]. In the latter threevalued logics are used (identifying B OUNDARY and U NKNOWN). The paper is structured as follows. First, in Section 2, we recall basic definitions related to rough sets and approximations. Next, in Section 3, we discuss our choice of four-valued logic. In Section 4 we introduce Rough Datalog and provide its semantics. Section 5 outlines a query language and discusses its implementation in logic programming. Finally, Section 6 concludes the paper.

2 Rough Sets According to Pawlak’s definition (see, e.g., [19]), a rough set S over a universe U is characterized by two subsets of U :

178

J. Małuszy´nski, A. Szałas, and A. Vit´oria Table 1. Test results considered in Example 1 car a a b c d

station safe s1 yes s2 no s2 no s1 yes s1 yes

– the set S, of all objects which can be certainly classified as belonging to S, called the lower approximation of S, and – the set S, of all objects which can be possibly classified as belonging to S, called the upper approximation of S. The set difference between the upper approximation and the lower approximation, denoted by S, is called the boundary region. In practice, in order to describe a given reality, one chooses a set of attributes and the elements of the underlying universe are described by tuples of attribute values. Rough sets are then defined by decision tables associating membership decisions with attribute values. The decisions are not exclusive: a given tuple of attribute values may be associated with the decision “yes”, with the decision “no”, with both or with none, if the tuple does not appear. Example 1. Consider a universe consisting of cars. If a car passed a test then it may be classified as safe (and as not safe, if it failed the test). Tests may be done independently at two test stations. The upper approximation of the rough set of safe cars would then include cars which passed at least one test. The lower approximation of the set would include the cars which passed all tests (and therefore, they did not fail at any test). The boundary region consists of the cars which passed one test and failed at one test. Notice that there are two other categories of cars, namely those which were not tested and those which failed all tests. As an example consider the situation described in Table 1, where the first column consists of cars, the second column consists of test stations and the third one contains test results. Denote by “Safe” the set of safe cars. Then: – the upper approximation of Safe consists of cars for which there is a decision “yes”, i.e., Safe = {a, c, d} – the lower approximation of Safe consists of cars for which all decisions are “yes”, i.e., Safe = {c, d} – the boundary region of Safe consists of cars for which there are both decisions “yes” and “no”, i.e., Safe = {a}. A decision table, representing a concept t, may be represented as a finite set of literals of the form t(y) or ¬t(x), where y ranges over the tuples of attribute values associated with the decision “yes” and x ranges over the tuples of attribute values associated with the decision “no”.

A Four-Valued Logic for Rough Set-Like Approximate Reasoning

179

Example 2. For the Example 1 with the universe of cars {a, b, c, d, e} and with two test stations, we may have the decision table, shown in Table 1, encoded as {safe(a), ¬safe(a), ¬safe(b), safe(c), safe(d)} . Notice that the literal safe(a) indicates that car a has passed a safety test in one of the stations while literal ¬safe(a) states that the same car as failed a safety test in another test station. In this case the rough set Safe has the approximations Safe = {a, c, d} and Safe = {c, d}. The rough set ¬Safe, describing those cars that have failed some test, has the approximations ¬Safe = {a, b} and ¬Safe = {b}. Note that it is totally unknown what is the status of car e. We notice that a decision table T of this kind defines two rough sets, T and ¬T , with a common boundary region which is the intersection of the upper approximations of both sets, i.e. T ∩ ¬T . As rough sets are usually defined by decision tables, we then adopt the following definition (used also in [20, 14, 15]). Definition 1. A rough set S over a universe U is a pair S, ¬S of subsets of U .

Intuitively, the rough set S describes those elements of U having certain property. The set S is the upper approximation of S, and consists of the elements of U for which there is an indication of having the given property. On the other hand, the set ¬S consists of the elements for which there is an indication of not having the property. In Example 2, Safe = {a, c, d} and ¬Safe = {a, b}. Remark 1 1. Observe that Definition 1 differs from the understanding of rough sets as defined by Pawlak. In fact, the definition of Pawlak requires the underlying elementary sets used in approximations to be based on equivalence relations, while Definition 1 relaxes this requirement. Such differences are examined and discussed in depth in [12]. 2. Since relations are sets of tuples, we further on also use the term rough relation to mean a rough set of tuples.

3 A Four-Valued Logic for Rough Sets 3.1 The Truth Values for Rough Membership Our objective is to define a logical language for rough set reasoning. The vocabulary of the language includes predicates to be interpreted as rough relations and constants to be used for representing attribute values. Consider an atomic formula of the form p(t1 , · · · , tn ), where p is a predicate, denoting a rough set P , and t1 , . . . , tn (with n > 0) are constants. We now want to define the truth value represented by an atom p(t1 , · · · , tn ). Let v = t1 , . . . , tn and “−” denote the set difference operation. Then, the following cases are possible:

180

J. Małuszy´nski, A. Szałas, and A. Vit´oria

– v ∈ P − ¬P : intuitively, we only have evidence that the element of the universe described by the attributes v has property P . Thus, the truth value of p(v) is defined to be T RUE . – v ∈ ¬P − P : intuitively, we only have evidence that the element of the universe described by the attributes v does not have property P . Thus, the truth value of p(v) is defined to be FALSE. – v ∈ P ∩ ¬P : in this case, we have contradictory evidences, i.e. an evidence that an element of the universe described by the attributes v has property P and an evidence that it does not have the property P . This is an uncertain information and we use the additional truth value B OUNDARY to denote it. – v ∈ P ∪ ¬P : in this case, we have no evidence whether the element of the universe described by the attributes v has property P . We then use another truth value called U NKNOWN. 3.2 Is Belnap Logic Suitable for Rough Reasoning? The truth values emerging from our discussion have been studied in the literature outside of the rough set context for defining four-valued logic. A standard reference is the well-known Belnap’s logic [2]. We now recall its basic principles and we discuss whether it is suitable for rough set reasoning. The Belnap logic is defined by considering a distributive bilattice of truth values and introducing logical connectives corresponding to the operations in the bilattice. Bilattices have been introduced in [21, 22]. They generalize the notion of Kripke structures (see, e.g., [23]). A bilattice is a structure B = U, ≤t , ≤k such that U is a non-empty set, ≤t and ≤k are partial orderings each making set U a lattice. Moreover, there is usually a useful connection between both orderings. We follow the usual convention that ∧t and ∨t stand respectively for the meet and join, with respect to ≤t . The symbols ∧k and ∨k stand respectively for the meet and join, with respect ≤k . Operations ∧t and ∨t are also called the conjunction and disjunction, and ∧k and ∨k are often designated as the consensus and accept all operators, respectively. The bilattice used in Belnap’s logic is shown in Fig 1. In the knowledge ordering, ≤k , U NKNOWN is the least value, reflecting total lack of knowledge. Each of the values T RUE and FALSE provide more information than U NKNOWN. Finally, the I NCONSISTENT value corresponds to the situation when there is evidence for both T RUE and FALSE.1 The truth ordering ≤t (see Fig 1) has T RUE as its largest element, and FALSE as its smallest element. Example 3. Assume that a family owns two cars: a and e. We want to check if the family has a safe car. This corresponds to the logical value of the expression safe(a) ∨t safe(e) .

1

(1)

Observe that I NCONSISTENT is replaced in our approach by B OUNDARY, which is closer to intuitions from rough set theory.

A Four-Valued Logic for Rough Set-Like Approximate Reasoning

181

≤k

6 I NCONSISTENT

FALSE

T RUE

U NKNOWN

- ≤t Fig. 1. The bilattice corresponding to Belnap’s logic

The truth values of safe(a) and safe(e) are determined by the results of the tests, as specified in Example 2. Thus safe(a) has the value B OUNDARY and safe(e) has the value U NKNOWN. If the join operation ∨t is defined by Belnap’s logic, as shown in Fig 1, then I NCONSISTENT ∨t U NKNOWN = T RUE . This contradicts our intuitions. We know that the safety of car a is unclear, since the results of both safety tests are contradictory, and we know nothing about safety of car e. Asking instead if all cars of the family are safe, safe(a) ∧t safe(e) ,

(2)

would in Belnap’s logic result in the answer FALSE. However, we really do not know whether both cars are safe because we do not have any information about the safety of car e. In contrast to the answer obtained in the Belnap’s logic, U NKNOWN seems to be a more intuitive answer in this case. The example above shows that the truth ordering of Fig 1, and consequently Belnap’s logic are not suitable for rough set-based reasoning. On the other hand, the knowledge ordering of Fig. 1 is adequate for our purposes. Indeed, the values T RUE and FALSE show that only one kind of evidence, either positive or negative, is known while the value B OUNDARY indicates existence of contradictory evidence, both positive and negative. 3.3 A Four-Valued Logic for Rough Set Reasoning We now define a four-valued logic suitable for rough set-based reasoning by modifying the bilattice of Fig.1. As discussed in Section 3.2, only the truth ordering is to be changed. We will use the new truth ordering to define conjunction (∧t ) as the greatest lower bound in this ordering. The ordering should preserve the usual meaning of conjunction for the truth values T RUE and FALSE. Intuitively, the value U NKNOWN represents the lack of information. Thus, the result of its conjunction with any other

182

J. Małuszy´nski, A. Szałas, and A. Vit´oria

truth value is accepted here to be U NKNOWN. A new information may arrive, replacing U NKNOWN by either T RUE, or FALSE, or B OUNDARY , providing in each case a different result. On the other hand, B OUNDARY represents existence of contradictory information. Its conjunction with T RUE would not remove this contradiction. Thus, we define the result of such a conjunction to be B OUNDARY . It also seems natural, that the conjunction of FALSE with T RUE or B OUNDARY gives FALSE. Consequently the truth ordering, ≤t , is redefined in our framework as U NKNOWN ≤t FALSE ≤t B OUNDARY ≤t T RUE .

(3)

The new structure R = U, ≤t , ≤k , where U is the universe of objects of interest, ≤t is the truth ordering defined in (3), and ≤k is the knowledge ordering as in the Belnap’s logic, gives the meaning of the logical connectives and is used in our approach. Example 4. Referring to Example 3, we then compute the logical values associated with the queries (1) and (2) by considering the new truth ordering above. The first query, (1) of Example 3, B OUNDARY ∨t U NKNOWN , returns the logical B OUNDARY which better corresponds to the intuitions. For the second query, (2) of Example 3, we have that B OUNDARY ∧t U NKNOWN = U NKNOWN . In contrast to Belnap’s logic, it is not excluded that some cars of the family of Example 3 are safe, but to be sure we need to obtain some information about the safety of car e. So, the answer U NKNOWN adequately reflects our intuitions. The proposition below shows that there is a connection between the knowledge ordering and the truth ordering. In this sense, the structure R can then be seen as a bilattice. Proposition 1. Consider the bilattice R = U, ≤t , ≤k and that x, y ∈ U . The operation ∧t is monotonic with respect to ≤k on both arguments, i.e. if x ≤k y then, for every z ∈ U , we have (z ∧t x) ≤k (z ∧t y) and (x ∧t z) ≤k (y ∧t z). Proof. Table 2 shows the result. Operation ∧t is obviously commutative.

We now define formally the logic underlying our work, called Rough Logic. Definition 2. Consider the following negation operation ¬. def

def

¬T RUE = FALSE, ¬FALSE = T RUE , def

def

¬B OUNDARY = B OUNDARY , ¬U NKNOWN = U NKNOWN. The propositional four-valued logic defined by the bilattice R together with negation ¬ is called the Rough Logic.

A Four-Valued Logic for Rough Set-Like Approximate Reasoning

183

Table 2. The table considered in the proof of Proposition 1 z B OUNDARY B OUNDARY B OUNDARY B OUNDARY B OUNDARY z FALSE FALSE FALSE FALSE FALSE z T RUE T RUE T RUE T RUE T RUE

x U NKNOWN U NKNOWN FALSE T RUE U NKNOWN

x U NKNOWN U NKNOWN FALSE T RUE U NKNOWN x U NKNOWN U NKNOWN FALSE T RUE U NKNOWN

z U NKNOWN U NKNOWN U NKNOWN U NKNOWN U NKNOWN

y T RUE FALSE B OUNDARY B OUNDARY B OUNDARY

y T RUE FALSE B OUNDARY B OUNDARY B OUNDARY y T RUE FALSE B OUNDARY B OUNDARY B OUNDARY

x U NKNOWN U NKNOWN FALSE T RUE U NKNOWN

z ∧t x U NKNOWN U NKNOWN FALSE B OUNDARY U NKNOWN

z ∧t x U NKNOWN U NKNOWN FALSE FALSE U NKNOWN z ∧t x U NKNOWN U NKNOWN FALSE T RUE U NKNOWN

y T RUE FALSE B OUNDARY B OUNDARY B OUNDARY

z ∧t y B OUNDARY FALSE B OUNDARY B OUNDARY B OUNDARY

z ∧t y FALSE FALSE FALSE FALSE FALSE z ∧t y T RUE FALSE B OUNDARY B OUNDARY B OUNDARY

z ∧t x U NKNOWN U NKNOWN U NKNOWN U NKNOWN U NKNOWN

z ∧t y U NKNOWN U NKNOWN U NKNOWN U NKNOWN U NKNOWN

4 Rough Datalog Language We now define a rule language, called Rough Datalog, such that its semantics is based on the Rough Logic. Intuitively, Rough Datalog corresponds to the usual logic programming language Datalog. While predicates in the latter denote crisp relations, in Rough Datalog a predicate p denotes a rough relation P . Thus, Rough Datalog caters for uncertainty in the knowledge. A rough literal is any expression of the form p(t1 , . . . , tn ) or ¬p(t1 , . . . , tn ). In Rough Datalog, knowledge is represented in the form of rough clauses, H:- B1 , . . . , Bn . where H and every Bi (0 ≤ i ≤ n) is a rough literal. A rough clause with the empty body (i.e. n = 0) is called a rough fact. A rough program P is a finite set of rough clauses. Rough clauses are used to specify rough relations as explained next. Intuitively, a rough clause is to be understood as the knowledge inequality ≤k stating that the truth value of the body is less than or equal to the truth value of the head. The comma

184

J. Małuszy´nski, A. Szałas, and A. Vit´oria

symbol “,” is interpreted as the meet in the truth ordering ≤t . Notice that the arguments of ≤k are the truth values U NKNOWN, B OUNDARY , T RUE, or FALSE but the logical value associated with a rough clause is either T RUE or FALSE. Information obtained from different rough clauses with heads referring to the same rough relation P (i.e. p or ¬p occurs in the head) is combined using the knowledge join operation ∨k . Example 5. The following rough clauses belong to an exemplary rough program P. (1) ¬useful(a) :- red(a), squared(a). “Object a is not useful if it is red and squared.” (2) squared(a) :- useful(a). —“Object a is squared if it is useful.” (3) ¬squared(a). —“Object a is not squared.”

4.1 Semantics of Rough Datalog Programs We now define notions of four-valued interpretation and model, extend the knowledge ordering to interpretations and show that each rough program has the least model in this ordering. Let P be a rough program and L be the set of all constant symbols occurring in P. Then, the Herbrand base HP is the set of all literals whose predicate symbols occur in P and whose arguments belong to L. A four-valued interpretation I of a rough program P associates with each atom a occurring in P a logical value I(a) ∈ {U NKNOWN, T RUE, FALSE, B OUNDARY } and ¬I(a) = I(¬a). The notion of interpretation extends naturally to conjunction (disjunction) of literals. Let l1 , . . . , ln , with n > 0, be rough literals. I(l1 ∧t · · · ∧t ln ) = I(l1 ) ∧t · · · ∧t I(ln ) . Definition 3. An interpretation I of a rough program P is any subset of the Herbrand base HP . Moreover, the rough relation I(p) is defined as I(p) = I(p), ¬I(p) = {t | p(t) ∈ I}, {t | ¬p(t) ∈ I} .

Intuitively, an interpretation associates each predicate p occurring in a program P with a rough set. Notice that ¬I(p) = I(¬p). Moreover, we have that – I(p(t)) = U NKNOWN, if t ∈ I(p) ∪ ¬I(p). – I(p(t)) = FALSE, if t ∈ ¬I(p). – I(p(t)) = T RUE, if t ∈ I(p). – I(p(t)) = B OUNDARY , if t ∈ I(p). Notice that we only consider variable-free rough programs. However, the results presented below can be also extended to rough programs with variables.

A Four-Valued Logic for Rough Set-Like Approximate Reasoning

185

An interpretation I of a rough program P satisfies a rough clause H:- B. ∈ P if I(B) ≤k I(H). A model M of P is any interpretation that satisfies every rough clause belonging to P. Notice also that the Herbrand base HP is a model of any rough program P. In this model the truth value of every literal is B OUNDARY . However, usually a program has more models. For comparing them we introduce a partial order on interpretations based on the knowledge ordering relation, ≤k . Definition 4. Let I1 ⊆ HP and I2 ⊆ HP be two interpretations. Then, I1 ≤k I2 , if and only if I1 (l) ≤k I2 (l), for every literal l ∈ HP . It can be checked that the knowledge ordering on interpretations corresponds to set inclusion. Proposition 2. I1 ≤k I2 if and only if I1 ⊆ I2 .

We show now that there is the least model for every rough program. Proposition 3. Let P be a rough program. Then, P has the least model with respect to ≤k . Proof. To prove that P has a least model with respect to ≤k , we show that the intersection of all models of P is also a model of P. n Let M = i Mi , where {M1 , . . . , Mn } (n ≥ 1) is the set of all models of P. Notice that, by Proposition 2, M ≤k Mi , with Mi ∈ {M1 , . . . , Mn }. We prove that M is a model of P. For this we have to show that, for any clause H :- B. ∈ P, we have M(H) ≥k M(B). We prove this by cases, considering possible truth values of the body of a clause. (a) If M(B) = U NKNOWN then M satisfies the rough clause, since U NKNOWN is the least element in the knowledge ordering. (b) If M(B) = T RUE then W (B) ≥t B OUNDARY , for every model W of P. Hence, W (H) ≥k T RUE, for every model W of P. Consequently, M(H) ≥k T RUE because the literal occurring in the head belongs to every model W . We conclude then that M satisfies the rough clause. (c) If M(B) = FALSE then B includes a literal l that is FALSE in some model of P and l is either FALSE or B OUNDARY in the other models. Obviously, no literal occurring in B can be U NKNOWN in any model. Consequently, M(H) ≥k FALSE because ¬H belongs to every model W . We conclude then that M satisfies the rough clause. (d) If M(B) = B OUNDARY then W (B) = B OUNDARY , for every model W of P. Notice that if I(B) = B OUNDARY , for some interpretation I of P, then we have that either I(l) = T RUE or I(l) = B OUNDARY , for every literal l in the body B. Hence, W (H) = B OUNDARY , for every model W of P. Consequently, M(H) = B OUNDARY because {H, ¬H} ⊆ W , for every model W . We conclude then that M satisfies the rough clause. The semantics of a rough program P is captured by its least model, with respect to ≤k .

186

J. Małuszy´nski, A. Szałas, and A. Vit´oria

Example 6. Consider again the rough program of Example 5. Its least model is M = {¬squared(a)}. Hence, useful(a) and red(a) are U NKNOWN, while squared(a) is FALSE. 4.2 A Fixpoint Characterization of the Least Model We now give a fixpoint characterization of the least model which makes it possible to compute the semantics of a program. We define an operator on interpretations, considered as sets of literals. We show that the operator is monotonic with respect to set inclusion. Thus, it has the least fixpoint (with respect to set inclusion) which can be obtained by iterations of the operator starting with the empty interpretation. We also show that the least fixpoint is a model. Taking into account Proposition 2, we can then conclude that the least fixpoint is also the least model of the program with respect to knowledge ordering. In the following definition if l is a negative literal of the form ¬a, then ¬l denotes a. Definition 5. Let P be a rough program. A total function TP mapping interpretations into interpretations is defined as follows: TP (I) = {l | l:- B. ∈ P and I(B) = T RUE} ∪ ∪ {¬l | l:- B. ∈ P and I(B) = FALSE} {l, ¬l | l:- B. ∈ P and I(B) = B OUNDARY } .

Thus, the set TP (I) consists of the heads of the rough clauses whose bodies are T RUE or B OUNDARY in I and, the negated heads of the rules whose bodies are FALSE or B OUNDARY in I. Such a way to gather heads of rules corresponds to defining the result by the disjunction of heads w.r.t. knowledge ordering ≤k . Proposition 4. Given a rough program P, the operator TP is monotonic with respect to set inclusion. Proof. The bodies of the program clauses are conjunctions of atoms. By Proposition 1 the conjunction is monotonic with respect to knowledge ordering. Hence by Proposition 2, it is also monotonic with respect to set inclusion of the interpretations. Thus, I ⊆ TP (I), for every interpretation I. The proposition above guarantees that TP has a least fixpoint (with respect to set inclusion), denoted as LFP(TP ). Proposition 5. Given a rough program P, the LFP(TP ) coincides with the least model of P. Proof. It is easy to see that the interpretation I = LFP(TP ) is a model of P. Assume the contrary. Then, there exists a clause H:- B. such that I(H) 0 , (18) μINCL(A,B) (x) = 0 otherwise . By assuming that μIncl(A,B) (x) = 0, for μA (x) = 0, we take into account only the support of the set A. For the sake of simplicity of the computational algorithm, it is not necessary to consider inclusion for all elements of the universe. Furthermore, we can require that the degree of inclusion with respect to x should be equal to 1, if the inequality μA (x) ≤ μB (x) for that x is satisﬁed I(μA (x), μB (x)) = 1,

if μA (x) ≤ μB (x) .

(19)

The requirement (19) is always satisﬁed by residual implicators. In order to deﬁne a suitable fuzzy counterpart of the rough inclusion function (12), we apply the notions of α-cut, power (cardinality) and support of a fuzzy set. Given a fuzzy subset A of the universe U , the α-cut of A, denoted by Aα , is a crisp set deﬁned as follows Aα = {x ∈ U : μA (x) ≥ α}

for

α ∈ [0, 1] .

(20)

For a ﬁnite fuzzy set A with n elements, power(A) and support(A) are given by power(A) =

n i=1

μA (xi ) ,

support(A) = {x : μA (xi ) > 0} .

(21)

198

A. Mieszkowicz-Rolka and L. Rolka

Using the above notions, we deﬁne the fuzzy rough inclusion function on the Cartesian product of the families F(U ) of all fuzzy subsets of the universe U να : F(U ) × F(U ) → [0, 1] .

(22)

Definition 5. The fuzzy rough α-inclusion function να (A, B) of any nonempty fuzzy set A in a fuzzy set B is defined as follows, for any α ∈ (0, 1] να (A, B) =

power(A ∩ INCL(A, B)α ) , power(A)

(23)

The value να (A, B) expresses how many elements of the nonempty fuzzy set A belong, at least to the degree α, to the fuzzy set B. First, we prove monotonicity of the proposed fuzzy rough inclusion function. Theorem 1. Implication-based fuzzy rough inclusion function να is monotonic with respect to the second parameter, for any α ∈ (0, 1] να (X, Y ) ≤ να (X, Z)

for any Y ⊆ Z,

where X, Y, Z ⊆ F(U ) .

Proof. According to the deﬁnition of a fuzzy subset [14], for Y ⊆ Z, we have μY (x) ≤ μZ (x), ∀ x ∈ U . Since every R-implicator, S-implicator and QL-implicator is right monotonic [25], it holds that: μI(X,Y ) (x) ≤ μI(X,Z) (x), ∀ x ∈ U . Thus, using the deﬁnition (18), we get μINCL(X,Y ) (x) ≤ μINCL(X,Z) (x),

∀x ∈ U .

Finally, for any α ∈ (0, 1], we can easy show that power(X ∩ INCL(X, Y )α ) power(X ∩ INCL(X, Z)α ) ≤ . power(X) power(X) Hence να (X, Y ) ≤ να (X, Z).

Furthermore, we can show that the rough inclusion function used in formulae (10) and (11) is a special case of the fuzzy rough inclusion function (23), when we use the implication-based inclusion set. Theorem 2. For any nonempty crisp set A, any crisp set B, and for α ∈ (0, 1], the implication-based inclusion function να (A, B) is equal to the inclusion degree incl(A, B). Proof. We show that for any crisp sets A and B, the inclusion set Incl(A, B) is equal to the crisp intersection A ∩ B. The membership function of any crisp set X is given by 1 for x ∈ X (24) μX (x) = 0 for x ∈ /X. Every implicator I satisﬁes the conditions: I(1, 1) = I(0, 1) = I(0, 0) = 1, and I(1, 0) = 0.

On Representation and Analysis of Crisp and Fuzzy Information Systems

Thus, applying the deﬁnition (18), we get 1 μIncl(A,B) (x) = μA∩B (x) = 0

if x ∈ A and x ∈ B otherwise .

199

(25)

For any ﬁnite crisp set X, and any α ∈ (0, 1], by formulae and (20), (21) and (24) we get: power(X) = card(X), and Xα = X. Consequently, applying (25), we ﬁnally have card(A ∩ B) power(A ∩ Incl(A, B)α ) = . power(A) card(A) Hence, we proved that να (A, B) = incl(A, B), for any α ∈ (0, 1].

We want to formulate the fuzzy rough approximation in a general way. Therefore, we introduce a function called res, deﬁned on the Cartesian product P(U )×F(U ), where P(U ) denotes the powerset of the universe U , and F(U ) the family of all fuzzy subsets of the universe U , respectively res : P(U ) × F(U ) → [0, 1] .

(26)

We require that res(∅, Y ) = 0 , res(X, Y ) ∈ {0, 1}, if Y is a crisp set , res(X, Y ) ≤ res(X, Z) for any Y ⊆ Z, where X ∈ P(U ), and Y, Z ∈ F(U ) . The form of the function res can be chosen depending on requirements of a considered application. For a given crisp set X and fuzzy set Y , the value of function res(X, Y ) should express the resulting membership degree in the set Y , taking into account not all elements of the universe, but only the elements of the set X. When we apply the limit-based approach, according to Dubois and Prade, we obtain the following form of the function res res(X, Y ) = inf μY (x) . x∈X

(27)

In the deﬁnition (27) of the function res, only one (limit) value of membership degree of elements in the set Y is taken into account. However, this means that we disregard the character (shape) of the membership function. Basing on a single value of membership degree is not always acceptable, especially in the case of large information systems. Hence, we can use the opportunity of giving another deﬁnitions of res, in which many values of membership degree are considered. Now, we introduce the notion of generalized fuzzy rough ε-approximation. Definition 6. For ε ∈ (0, 1], the ε-approximation Φε (A) of a fuzzy set A, by a fuzzy partition Φ = {F1 , F2 , . . . , Fn }, is a fuzzy set on the domain Φ with membership function expressed as μΦε (A) (Fi ) = res(Sε (Fi , A), INCL(Fi , A)) ,

(28)

200

A. Mieszkowicz-Rolka and L. Rolka

where Sε (Fi , A) = support(Fi ∩ INCL(Fi , A)αε ) , αε = sup{α ∈ [0, 1] : να (Fi , A) ≥ ε} . The set Sε (Fi , A) is equal to support of the intersection of the class Fi with the part of INCL(Fi , A), which contains those elements of the approximating class Fi which are included in A at least to the degree αε . The resulting membership μΦε (A) (Fi ) is determined using only the elements from Sε (Fi , A) instead of the whole class Fi . This is accomplished by applying the function res. It can be easy shown that applying the deﬁnition (27) of the function res leads to a simple form of the ε-approximation (28) μΦε (A) (Fi ) = sup{α ∈ [0, 1] : να (Fi , A) ≥ ε} .

(29)

In contrast to the approximations (16) and (17), which use two diﬀerent fuzzy connectives, we have a single uniﬁed deﬁnition of fuzzy rough approximation. In this way we obtain a consistent variable precision fuzzy rough set model. Thus, we are able to compare approximations determined for various values of the parameter ε. 3.3

Analysis of Fuzzy Decision Tables

In the analysis of fuzzy decision tables, two fuzzy partitions are generated with the help of a suitable similarity relation. The partition obtained with respect to condition attributes is used for approximation of fuzzy similarity classes obtained with respect to decision attributes. It is necessary to address the problem of comparing objects described by fuzzy sets. This issue has been widely studied in the literature, see, for example, [3,7,8]. In our considerations, we apply a symmetric T-transitive fuzzy similarity relation [3], which is deﬁned by means of the distance between the compared elements. In the following, we only give formulae for condition attributes. We apply the notation given in section 2. If we need to compare any two elements x and y of the universe U with respect to the condition attribute ci , i = 1, 2, . . . , n, then the similarity between x and y could be expressed using a T-similarity relation based on the Łukasiewicz T-norm [7]. (30) Sci (x, y) = 1 − max |μVik (x) − μVik (y)| . k=1,ni

In order to evaluate the similarity SC (x, y), with respect to all condition attributes C, we must aggregate the results obtained for particular attributes ci , i = 1, 2, . . . , n. This can be done by using the T-norm operator min as follows SC (x, y) = min Sci (x, y) = min (1 − max |μVik (x) − μVik (y)|) . i=1,n

i=1,n

k=1,ni

(31)

By calculating the similarity for all pairs of elements of the universe U , we obtain a symmetric similarity matrix. If the value of similarity between the elements x and y is equal to 1, they belong to the same similarity class. In that case two rows

On Representation and Analysis of Crisp and Fuzzy Information Systems

201

of the similarity matrix should be merged into one fuzzy set with membership degrees equal to 1 for x and y. In consequence, we get a family of fuzzy similarity = {C1 , C2 , . . . , Cn }, for condition attributes C and a family of fuzzy classes C = {D1 , D2 , . . . , Dm }, for decision attributes D. similarity classes D In the next step, we determine fuzzy rough approximations of elements of the by the family C, using the parameterized fuzzy rough set model. family D To determine the consistency of fuzzy decision tables and signiﬁcance of attributes, we apply a generalized measure of ε-approximation quality [17]. For = {C1 , C2 , . . . , Cn } the = {D1 , D2 , . . . , Dm } and the family C the family D ε-approximation quality of D by C is deﬁned as follows = γCε (D) where

= PosCε (D)

power(PosCε (D)) card(U )

,

(32)

ε (Dj )) ∩ Dj . ω(C

Dj ∈D

into the domain The fuzzy extension ω denotes a mapping from the domain C of the universe U , which is expressed for any fuzzy set X by μω(X) (x) = μX (Ci ),

if μCi (x) = 1 .

(33)

The deﬁnition (32) is based on the generalized notion of positive region. For any fuzzy set X and a similarity relation R, the positive region of X is deﬁned as follows (34) PosRε (X) = X ∩ ω(Rε (X)) . In the deﬁnition of the positive region (34), we take into account only those elements of the ε-approximation, for which there is no contradiction between the set X and the approximating similarity classes.

4

Fuzzy Flow Graphs

In addition to the VPFRS model, we want to introduce fuzzy ﬂow graphs as a second tool for analysis of fuzzy information systems. The idea of applying ﬂow graphs in the framework of crisp rough sets, for discovering the statistical properties of decision algorithms, was proposed by Pawlak [20,21,22]. We should start with recalling the basic notions of the crisp ﬂow graph approach. A ﬂow graph is given in the form of directed acyclic ﬁnal graph G = (N , B, ϕ), where N is a set of nodes, B ⊆ N × N is a set of directed branches, ϕ: B → R+ is a ﬂow function with values in the set of non-negative reals R+ . For any (X, Y ) ∈ B, X is an input of Y and Y is an output of X. The quantity ϕ(X, Y ) is called the throughﬂow from X to Y . I(X) and O(X) denote an input and an output of X, respectively. The input I(G) and output O(G) of a graph G are deﬁned by I(G) = {X ∈ N : I(X) = ∅} ,

O(G) = {X ∈ N : O(X) = ∅} .

(35)

202

A. Mieszkowicz-Rolka and L. Rolka

Every node X ∈ N of a ﬂow graph G is characterized by its inﬂow ϕ(Y, X) , ϕ+ (X) =

(36)

Y ∈I(X)

and by its outﬂow ϕ− (X) =

(37)

ϕ(X, Y ) .

Y ∈O(X)

For any internal node X, the equality ϕ+ (X) = ϕ− (X) = ϕ(X) is satisﬁed. The quantity ϕ(X) is called the ﬂow of the node X. The ﬂow for the whole graph G is deﬁned by ϕ− (X) = ϕ+ (X) . (38) ϕ(G) = x∈I(G)

x∈O(G)

By using the ﬂow ϕ(G), the normalized throughﬂow σ(X, Y ) and the normalized ﬂow σ(X) are determined as follows σ(X, Y ) =

ϕ(X, Y ) , ϕ(G)

σ(X) =

ϕ(X) . ϕ(G)

(39)

For every branch of a ﬂow graph G the certainty factor is deﬁned by cer(X, Y ) =

σ(X, Y ) . σ(X)

(40)

The coverage factor for every branch of a ﬂow graph G is deﬁned by cov(X, Y ) =

σ(X, Y ) . σ(Y )

The certainty and coverage factors satisfy the following properties cer(X, Y ) = 1 , cov(X, Y ) = 1 . Y ∈O(X)

(41)

(42)

X∈I(Y )

The measures of certainty (40) and coverage (41) are useful for analysis of decision algorithms [10]. Now, we consider the issue of applying ﬂow graphs to representation and analysis of fuzzy decision algorithms. We use decision tables with fuzzy values of attributes, presented in section 2. All possible decision rules, generated by the Cartesian product of sets of linguistic values of the attributes, will m be examined. n According to notation used in section 2, we obtain r = i=1 ni j=1 mj possible rules. The k-th decision rule, denoted by Rk , is expressed as follows Rk : IF c1 is V1k AND c2 is V2k . . . AND cn is Vnk k THEN d1 is W1k AND d2 is W2k . . . AND dm is Wm

(43)

where k = 1, 2, . . . , r , Vik ∈ Vi , i = 1, 2, . . . n , Wjk ∈ Wj , j = 1, 2, . . . , m.

On Representation and Analysis of Crisp and Fuzzy Information Systems

203

When we use the fuzzy Cartesian products C k = V1k × V2k × . . . × Vnk and k D = W1k × W2k × . . . × Wm , the k-th decision rule can be expressed in the form of a fuzzy implication, denoted here by C k → Dk . It is necessary to select a subset of decision rules which are relevant to the considered decision process. This can be done by determining to what degree any element x ∈ U , corresponding to a single row of the decision table, conﬁrms particular decision rules. We calculate the truth value of the decision rule’s antecedent and the truth value of the decision rule’s consequent, by determining the conjunction of the respective membership degrees of x in the linguistic values of attributes. If we take a decision table with crisp attributes, a decision rule can be conﬁrmed for some x, if the result of conjunction is equal to 1, both for the rule’s premise and the rule’s conclusion. Otherwise, the element x does not conﬁrm the considered decision rule. The set of elements x ∈ U , which conﬁrm a decision rule, is called the support of the decision rule. To determine the conﬁrmation degree of fuzzy decision rules, a T-norm operator need to be applied. By cd(x, k), we denote the conﬁrmation degree of the k-th decision rule by the element x ∈ U k

cd(x, k) = T(cda(x, k), cdc(x, k)) ,

(44)

where cda(x, k) denotes the conﬁrmation degree of the decision rule’s antecedent cda(x, k) = T(μV1k (x), μV2k (x), . . . , μVnk (x)) ,

(45)

and cdc(x, k) the conﬁrmation degree of the decision rule’s consequent cdc(x, k) = T(μW1k (x), μW2k (x), . . . , μWm k (x)) .

(46)

Through determining the conﬁrmation degrees (45), (46) and (44), we generate the following fuzzy sets on the domain U : the support of the decision rule’s antecedent support(cda(x, k)) = {cda(x1 , k)/x1 , cda(x2 , k)/x2 , . . . , cda(xN , k)/xN }, (47) the support of the decision rule’s consequent support(cdc(x, k)) = {cdc(x1 , k)/x1 , cdc(x2 , k)/x2 , . . . , cda(xN , k)/xN } , (48) and the support of the decision rule Rk , respectively support(Rk ) = {cd(x1 , k)/x1 , cd(x2 , k)/x2 , . . . , cd(xN , k)/xN } .

(49)

The introduced notions (47), (48) and (49) will be used for deﬁning strength, certainty, and coverage factors of a decision rule. Now, let us explain the way of constructing fuzzy ﬂow graphs on the basis of a decision table with fuzzy attributes.

204

A. Mieszkowicz-Rolka and L. Rolka

Every fuzzy attribute is represented by a layer of nodes. The nodes of a layer correspond to linguistic values of a given attribute. a fuzzy set on the universe U , which describes membership We denote by X degree of particular elements x ∈ U in the linguistic value represented by X. The can be found in a respective column of membership degrees of all x in the set X the considered decision table. Let us pick out such two attributes, which are represented by two consecutive layers of the ﬂow graph. We denote by X a linguistic value of the ﬁrst attribute, and by Y a linguistic value of the second attribute. In the case of crisp ﬂow graphs, the ﬂow between nodes X and Y is equal to the number of elements of the universe U , which are characterized by the combination of attribute values X and Y . In consequence, a particular element x ∈ U can only be assigned to a unique path in the ﬂow graph. In a fuzzy information system, however, every element of the universe can belong to several linguistic values, and it can be assigned to several paths in the ﬂow graph. It is possible to determine the ﬂow distribution in the crisp ﬂow graph by using the operations of set intersection and set cardinality. To obtain the ﬂow ϕ(X, Y ) for the branch (X, Y ) of a fuzzy ﬂow graph, we have to calculate power of the in and Y . Many deﬁnitions of fuzzy intersection (T-norm tersection of fuzzy sets X operator) are known. In order to satisfy the ﬂow conservation equations, it is necessary to use the T-norm operator prod for determining the intersection of sets. Furthermore, we should assume that the linguistic values of attributes satisfy the requirement (3). We conclude the above discussion with the following theorem. Theorem 3. Let S be a fuzzy information systems with the linguistic values of attributes satisfying the requirement (3), and let ∩ denote a fuzzy intersection operator based on the T-norm prod. The following properties are satisfied for the flow graph, which represents the information system S : (G1) the inflow for any output or internal layer node X is given by = ∩ Y ) , ϕ(Y, X) = power(X ϕ+ (X) = power(X) Y ∈I(X)

Y ∈I(X)

(G2) the outflow for any input or internal layer node X is given by = ∩ Y ) , ϕ− (X) = power(X) ϕ(X, Y ) = power(X Y ∈O(X)

(50)

(51)

Y ∈O(X)

(G3) for any internal layer node X, it holds that ϕ+ (X) = ϕ− (X) .

(52)

The properties (G1), (G2) and (G3) do not hold in general, if we use another Tnorm operator, e.g. min. In the special case of crisp decision tables, the formulae (50) and (51) become equivalent to (36) and (37). The layers corresponding to condition attributes can be merged into a single layer, which contains nodes representing all possible combinations of linguistic

On Representation and Analysis of Crisp and Fuzzy Information Systems

205

values of the condition attributes. We can also merge all the layers corresponding to decision attributes. Let us denote by X ∗ , a node of the resulting layer obtained for condition attributes and by Y ∗ , a node of the resulting layer obtained for decision attributes. The node X ∗ corresponds to antecedent of some decision rule Rk . Support of the antecedent of the decision rule Rk is determined with the help of formula (47). The decision rule Rk is represented by the branch (X ∗ , Y ∗ ). Power of the support of the rule Rk is equal to the ﬂow between the nodes X ∗ and Y ∗ , which is obtained using formula (49) ϕ(X ∗ , Y ∗ ) = power(support(Rk )) .

(53)

By applying the formulae (47), (48) and (49), we can determine, for every decision rule Rk , the certainty factor cer(X ∗ , Y ∗ ), the coverage factor cov(X ∗ , Y ∗ ), and the strength of the rule σ(X ∗ , Y ∗ ) cer(X ∗ , Y ∗ ) = cer(Rk ) =

power(support(Rk )) , power(support(cda(x, k)))

(54)

cov(X ∗ , Y ∗ ) = cov(Rk ) =

power(support(Rk )) , power(support(cdc(x, k)))

(55)

σ(X ∗ , Y ∗ ) = strength(Rk ) =

power(support(Rk )) . card(U )

(56)

It is possible to represent any decision rule by a sequence of nodes [X1 . . . Xn ], namely by a path from the 1-th to the n-th layer of the ﬂow graph G. For a given path [X1 . . . Xn ], the resulting certainty and strength can be deﬁned. In contrast to the deﬁnitions presented in [20,21,22], in which the statistical properties of ﬂow are taken into account, we propose a diﬀerent form of the path’s certainty and strength n−1 cer[X1 . . . Xn ] = cer(X1 . . . Xi , Xi+1 ) , (57) i=1

σ[X1 . . . Xn ] = σ(X1 ) cer[X1 . . . Xn ] , where cer(X1 . . . Xi , Xi+1 ) =

2 ∩ . . . ∩ X i+1 ) 1 ∩ X power(X . i ) power(X1 ∩ X2 ∩ . . . ∩ X

(58)

(59)

The resulting certainty (57) of the path [X1 . . . Xn ], expresses what part of the ﬂow of the starting node X1 reaches the ﬁnal node Xn , passing through all nodes of the path.

5

Examples

Let us analyze a fuzzy decision table (Table 1) with condition attributes c1 and c2 and one decision attribute d. All attributes have three linguistic values.

206

A. Mieszkowicz-Rolka and L. Rolka Table 1. Decision table with fuzzy attributes c1

x1 x2 x3 x4 x5 x6 x7 x8 x9 x10

c2

d

V11

V12

V13

V21

V22

V23

W11

W12

W13

0.1 0.8 0.0 0.1 0.0 0.8 0.1 0.0 0.0 0.1

0.9 0.2 0.2 0.9 0.8 0.2 0.9 0.1 0.2 0.9

0.0 0.0 0.8 0.0 0.2 0.0 0.0 0.9 0.8 0.0

0.0 1.0 0.0 0.0 0.8 0.0 0.0 0.8 0.0 0.1

0.9 0.0 0.2 0.9 0.2 0.2 0.9 0.2 0.2 0.9

0.1 0.0 0.8 0.1 0.0 0.8 0.1 0.0 0.8 0.0

0.0 0.0 0.9 0.0 0.0 1.0 0.1 0.0 0.9 0.0

1.0 0.1 0.1 1.0 0.1 0.0 0.9 0.0 0.1 0.9

0.0 0.9 0.0 0.0 0.9 0.0 0.0 1.0 0.0 0.1

First, we apply the variable precision fuzzy rough set approach. Using similarity relation in the form (31), we determine similarity matrices with respect to condition and decision attributes. By merging identical rows of the similarity matrix, we get 9 condition similarity classes and and 6 decision similarity classes. We calculate ε-approximation quality using the Łukasiewicz implication operator. The results are presented in table 2. Table 2. ε-approximation quality for diﬀerent values of parameter ε

Method

Ł-inf

γCε (D)

Removed attribute

ε=1

ε = 0.9

ε = 0.85

ε = 0.8

none c1 c2

0.830 0.820 0.250

0.900 0.880 0.250

0.900 0.880 0.410

0.910 0.910 0.450

We can state that the considered information system has a high consistency. The condition attribute c1 can be omitted from the decision table without a signiﬁcant decrease of the ε-approximation quality. In the next step, the ﬂow graph method will be applied. We use the same labels for both the linguistic values of the attributes and the corresponding nodes of the ﬂow graph. As stated in previous section, the T-norm operator prod should be used in our calculations. The obtained fuzzy ﬂow graph has a very simple form, because there is only one condition attribute c2 and one decision attribute d. Values of the normalized ﬂow between nodes of the condition layer and nodes of the decision layer are given in Table 3.

On Representation and Analysis of Crisp and Fuzzy Information Systems

207

Table 3. Normalized ﬂow between nodes of condition and decision layers σ(V2i , W1j )

V21 V22 V23 Σ

W11

W12

W13

Σ

0.000 0.065 0.225 0.290

0.027 0.348 0.045 0.420

0.243 0.047 0.000 0.290

0.270 0.460 0.270 1.000

We see that the ﬂow conservation equations (50) and (51), are satisﬁed, for example, 3 power(V21 ) σ− (V21 ) = = σ(V21 , W1i ) = 0.270 , card(U ) i=1 11 ) power(W = σ(V2i , W11 ) = 0.290 . card(U ) i=1 3

σ+ (W11 ) =

Let us determine the certainty and coverage factors for branches between the layers according to formulae (54), (55). The results are given in Tables 4 and 5. Table 4. Certainty factor for branches between condition and decision layers cer(V2i , W1j )

V21 V22 V23

W11

W12

W13

Σ

0.0000 0.1413 0.8333

0.1000 0.7565 0.1667

0.9000 0.1022 0.0000

1.0000 1.0000 1.0000

Table 5. Coverage factor for branches between condition and decision layers cov(V2i , W1j )

V21 V22 V23 Σ

W11

W12

W13

0.0000 0.2241 0.7759 1.0000

0.0643 0.8286 0.1071 1.0000

0.8379 0.1621 0.0000 1.0000

Fuzzy decision rules with the largest values of certainty factor (Table 6) can be included in the ﬁnal fuzzy inference system. The respective values of coverage factor are useful for explaining the selected decision rules. Only 3 decision rules

208

A. Mieszkowicz-Rolka and L. Rolka Table 6. Decision rules with the largest value of certainty factor decision rule

certainty

coverage

V21 → W13 V22 → W12 V23 → W11

0.9000 0.7565 0.8333

0.8379 0.8286 0.7759

strength [%] 24.30 34.80 22.50

could be generated from our decision table. Owing to the application of the VPFRS approach, we got a simple fuzzy ﬂow graph. Let us construct a ﬂow graph without a prior reduction of attributes. We merge the layers corresponding to condition attributes c1 and c2 to a resulting layer, which represents all possible linguistic values in the antecedences of decision rules. We determine the degrees of satisfaction of the rules’ antecedences for particular elements x ∈ U . For the antecedence represented by V12 V22 , we get: V 12 V22 = V12 ∩ V22 = { 0.81/x1 , 0.00/x2 , 0.04/x3 , 0.81/x4 , 0.16/x5 , 0.04/x6 , 0.81/x7 , 0.02/x8, 0.04/x9 , 0.81/x10 }, ϕ(V12 , V22 ) = power(V 12 V22 ) = 3.54, σ(V12 , V22 ) =

ϕ(V12 ,V22 ) cardU

= 0.354.

Table 7. Decision rules with the largest certainty factor (full information system) decision rule V11 V21 V11 V23 V12 V21 V12 V22 V13 V21 V13 V23

→ W13 → W11 → W13 → W12 → W13 → W11

certainty

coverage

0.8901 0.9567 0.8366 0.8763 0.9818 0.9000

0.2486 0.2210 0.2914 0.7386 0.2979 0.3972

strength [%] 7.21 6.41 8.45 31.02 8.64 11.52

Finally, we determine the normalized throughﬂow, certainty and coverage factors for branches between of the resulting condition and decision layers. Decision rules with the largest value of certainty factor are given in Table 7. We can observe that the attribute c1 is superﬂuous in the obtained decision rules.

6

Conclusions

Information systems with crisp and fuzzy attributes can be eﬀectively analyzed by a hybrid approach which combines the variable precision fuzzy rough set (VPFRS) model with fuzzy ﬂow graphs. The VPFRS model can be deﬁned in a uniﬁed way with the help of a single notion of ε-approximation. This allows to

On Representation and Analysis of Crisp and Fuzzy Information Systems

209

avoid the inconsistency of the VPFRS model caused by diﬀerent forms of fuzzy connectives. The proposed fuzzy ﬂow graph method is suitable for representing and analyzing decision tables with fuzzy attributes. Every fuzzy attribute can be represented by a layer of a ﬂow graph. All nodes of a layer correspond to linguistic values of an attribute. A fuzzy decision table can be reduced by applying the VPFRS approach prior to using the fuzzy ﬂow graph method for determining a system of fuzzy decision rules.

References 1. Bandler, W., Kohout, L.: Fuzzy Power Sets and Fuzzy Implication Operators. Fuzzy Sets and Systems 4 (1980) 13–30 2. Burillo, P., Frago, N., Fuentes, R.: Inclusion Grade and Fuzzy Implication Operators. Fuzzy Sets and Systems 114 (2000) 417–429 3. Chen, S.M., Yeh, M.S., Hsiao, P.Y.: A Comparison of Similarity Measures of Fuzzy Values. Fuzzy Sets and Systems 72 (1995) 79–89 4. Cornelis, C., Van der Donck, C., Kerre, E.: Sinha-Dougherty Approach to the Fuzziﬁcation of Set Inclusion Revisited. Fuzzy Sets and Systems 134 (2003) 283–295 5. De Baets, B., De Meyer, H., Naessens, H.: On Rational Cardinality-based Inclusion Measures. Fuzzy Sets and Systems 128 (2002) 169–183 6. Dubois, D., Prade, H.: Putting Rough Sets and Fuzzy Sets Together. [30] 203–232 7. Fernández Salido, J.M., Murakami, S.: Rough Set Analysis of a General Type of Fuzzy Data Using Transitive Aggregations of Fuzzy Similarity Relations. Fuzzy Sets and Systems 139 (2003) 635–660 8. Greco, S., Matarazzo, B., Słowiński, R.: Rough Set Processing of Vague Information Using Fuzzy Similarity Relations. In: Calude, C.S., Paun, G., (eds.): Finite Versus Inﬁnite — Contributions to an Eternal Dilemma. Springer-Verlag, Berlin Heidelberg New York (2000) 149–173 9. Greco, S., Pawlak, Z., Słowiński, R.: Generalized Decision Algorithms, Rough Inference Rules, and Flow Graphs. In: Alpigini, J., Peters, J.F., Skowron, A., Zhong, N., (eds.): Rough Sets and Current Trends in Computing. Lecture Notes in Artiﬁcial Intelligence, Vol. 2475. Springer-Verlag, Berlin Heidelberg New York (2002) 93–104 10. Greco, S., Pawlak, Z., Słowiński, R.: Bayesian Conﬁrmation Measures within Rough Set Approach. [31] 264–273 11. Greco, S., Matarazzo, B., Słowiński, R.: Rough Membership and Bayesian Conﬁrmation Measures for Parameterized Rough Sets. [28] 314–324 12. Inuiguchi, M.: Generalizations of Rough Sets: From Crisp to Fuzzy Cases. [31] 26–37 13. Katzberg, J.D., Ziarko, W.: Variable Precision Extension of Rough Sets. Fundamenta Informaticae 27 (1996) 155–168 14. Klir, G.J., Folger, T.A.: Fuzzy Sets, Uncertainty, and Information. Prentice Hall, Englewood, New Jersey (1988) 15. Lin, T.Y.: Coping with Imprecision Information — Fuzzy Logic. Downsizing Expo, Santa Clara Convention Center (1993) 16. Mieszkowicz-Rolka, A., Rolka, L.: Variable Precision Rough Sets: Evaluation of Human Operator’s Decision Model. In: Sołdek, J., Drobiazgiewicz, L., (eds.): Artiﬁcial Intelligence and Security in Computing Systems. Kluwer Academic Publishers, Boston Dordrecht London (2003) 33–40

210

A. Mieszkowicz-Rolka and L. Rolka

17. Mieszkowicz-Rolka, A., Rolka, L.: Variable Precision Fuzzy Rough Sets Model in the Analysis of Process Data. [28] 354–363 18. Nakamura, A.: Application of Fuzzy-Rough Classiﬁcations to Logics. [30] 233–250 19. Pawlak, Z.: Rough Sets: Theoretical Aspects of Reasoning about Data. Kluwer Academic Publishers, Boston Dordrecht London (1991) 20. Pawlak, Z.: Decision Algorithms, Bayes’ Theorem and Flow Graphs. In: Rutkowski, L., Kacprzyk, J., (eds.): Advances in Soft Computing. Physica-Verlag, Heidelberg (2003) 18–24 21. Pawlak, Z.: Flow Graphs and Data Mining. [23] 1–36 22. Pawlak, Z.: Rough Sets and Flow Graphs. [28] 1–11 23. Peters, J.F., et al., (eds.): Transactions on Rough Sets III. Lecture Notes in Computer Science (Journal Subline), Vol. 3400. Springer-Verlag, Berlin Heidelberg New York (2005) 24. Polkowski, L.: Toward Rough Set Foundations. Mereological Approach. [31] 8–25 25. Radzikowska, A.M., Kerre, E.E.: A Comparative Study of Fuzzy Rough Sets. Fuzzy Sets and Systems 126 (2002) 137–155 26. Skowron, A., Stepaniuk, J.: Tolerance Approximation Spaces. Fundamenta Informaticae 27 (1996) 245–253 27. Ślęzak, D., Ziarko, W.: Variable Precision Bayesian Rough Set Model. In: Wang, G., Liu, Q., Yao, Y., Skowron, A., (eds.): Rough Sets, Fuzzy Sets, Data Mining, and Granular Computing. Lecture Notes in Artiﬁcial Intelligence, Vol. 2639. SpringerVerlag, Berlin Heidelberg New York (2003) 312–315 28. Ślęzak, D., et al., (eds.): Rough Sets and Current Trends in Computing. Lecture Notes in Artiﬁcial Intelligence, Vol. 3641. Springer-Verlag, Berlin Heidelberg New York (2005) 29. Ślęzak, D.: Rough Sets and Bayes Factor. [23] 202–229 30. Słowiński, R., (ed.): Intelligent Decision Support: Handbook of Applications and Advances of the Rough Sets Theory. Kluwer Academic Publishers, Boston Dordrecht London (1992) 31. Tsumoto, S., et al., (eds.): Rough Sets and Current Trends in Computing. Lecture Notes in Artiﬁcial Intelligence, Vol. 3066. Springer-Verlag, Berlin Heidelberg New York (2004) 32. Yager, R.R., Filev, D.P.: Essentials of Fuzzy Modelling and Control. John Wiley & Sons, Inc., New York (1994) 33. Liu, W.N., Yao, J., Yao, Y.: Rough Approximations under Level Fuzzy Sets. [31] 78–83 34. Zadeh, L.: Fuzzy Sets. Information and Control 8 (1965) 338–353 35. Ziarko, W.: Variable Precision Rough Sets Model. Journal of Computer and System Sciences 46 (1993) 39–59 36. Ziarko, W.: Probabilistic Rough Sets. [28] 283–293

On Partial Covers, Reducts and Decision Rules with Weights Mikhail Ju. Moshkov1, Marcin Piliszczuk2 , and Beata Zielosko3 1

2

Institute of Computer Science, University of Silesia 39, B¸edzi´ nska St., Sosnowiec, 41-200, Poland [email protected] ´ aski S.A., 34, Sokolska St., Katowice, 40-086, Poland ING Bank Sl¸ [email protected] 3 Institute of Computer Science, University of Silesia 39, B¸edzi´ nska St., Sosnowiec, 41-200, Poland [email protected]

Abstract. In the paper the accuracy of greedy algorithms with weights for construction of partial covers, reducts and decision rules is considered. Bounds on minimal weight of partial covers, reducts and decision rules based on an information on greedy algorithm work are studied. Results of experiments with software implementation of greedy algorithms are described. Keywords: partial covers, partial reducts, partial decision rules, weights, greedy algorithms.

1

Introduction

The paper is devoted to consideration of partial decision-relative reducts (we will omit often words ”decision-relative”) and partial decision rules for decision tables on the basis of partial cover investigation. Rough set theory [11,17] often deals with decision tables containing noisy data. In this case exact reducts and rules can be ”overlearned” i.e. depend essentially on noise. If we see constructed reducts and rules as a way of knowledge representation [16] then instead of large exact reducts and rules it is more appropriate to work with relatively small partial ones. In [12] Zdzislaw Pawlak wrote that ”the idea of an approximate reduct can be useful in cases when a smaller number of condition attributes is preferred over accuracy of classiﬁcation”. Last years in rough set theory partial reducts, partial decision rules and partial covers are studied intensively [6,7,8,9,10,13,19,20,21,22,23,24,27]. Approximate reducts are investigated also in extensions of rough set model such as VPRS (variable precision rough sets) [26] and α-RST (alpha rough set theory) [14]. We study the case where each subset, used for covering, has its own weight, and we must minimize the total weight of subsets in partial cover. The same situation is with partial reducts and decision rules: each conditional attribute has its own weight, and we must minimize the total weight of attributes in partial J.F. Peters et al. (Eds.): Transactions on Rough Sets VI, LNCS 4374, pp. 211–246, 2007. c Springer-Verlag Berlin Heidelberg 2007

212

M.Ju. Moshkov, M. Piliszczuk, and B. Zielosko

reduct or decision rule. If weight of each attribute characterizes time complexity of attribute value computation then we try to minimize total time complexity of computation of attributes from partial reduct or partial decision rule. If weight characterizes a risk of attribute value computation (as in medical or technical diagnosis) then we try to minimize total risk, etc. In rough set theory various problems can be represented as set cover problems with weights: – problem of construction of a reduct [16] or partial reduct with minimal total weight of attributes for an information system; – problem of construction of a decision-relative reduct [16] or partial decisionrelative reduct with minimal total weight of attributes for a decision table; – problem of construction of a decision rule or partial decision rule with minimal total weight of attributes for a row of a decision table (note that this problem is closely connected with the problem of construction of a local reduct [16] or partial local reduct with minimal total weight of attributes); – problem of construction of a subsystem of a given system of decision rules which ”covers” the same set of rows and has minimal total weight of rules (in the capacity of a rule weight we can consider its length). So the study of covers and partial covers is of some interest for rough set theory. In this paper we list some known results on set cover problems which can be useful in applications and obtain certain new results. From results obtained in [20,22] it follows that the problem of construction of partial cover with minimal weight is N P -hard. Therefore we must consider polynomial approximate algorithms for minimization of weight of partial covers. In [18] a greedy algorithm with weights for partial cover construction was investigated. This algorithm is a generalization of well known greedy algorithm with weights for exact cover construction [2]. The algorithm from [18] is a greedy algorithm with one threshold which gives the exactness of constructed partial cover. Using results from [9] (based on results from [3,15] and technique created in [20,22]) on precision of polynomial approximate algorithms for construction of partial cover with minimal cardinality and results from [18] on precision of greedy algorithm with one threshold we show that under some natural assumptions on the class N P the greedy algorithm with one threshold is close to best polynomial approximate algorithms for construction of partial cover with minimal weight. However we can try to improve results of the work of greedy algorithm with one threshold for some part of set cover problems with weight. We generalize greedy algorithm with one threshold [18], and consider greedy algorithm with two thresholds. First threshold gives the exactness of constructed partial cover, and the second one is an interior parameter of the considered algorithm. We prove that for the most part of set cover problems there exist a weight function and values of thresholds such that the weight of partial cover constructed by greedy algorithm with two thresholds is less than the weight of partial cover constructed by greedy algorithm with one threshold.

On Partial Covers, Reducts and Decision Rules with Weights

213

We describe two polynomial algorithms which always construct partial covers that are not worse than the one constructed by greedy algorithm with one threshold, and for the most part of set cover problems there exists a weight function and a value of ﬁrst threshold such that the weight of partial covers constructed by the considered two algorithms is less than the weight of partial cover constructed by greedy algorithm with one threshold. Information on greedy algorithm work can be used for obtaining lower bounds on minimal cardinality of partial covers [9]. We ﬁx some kind of information on greedy algorithm work, and ﬁnd unimprovable lower bound on minimal weight of partial cover depending on this information. Obtained results show that this bound is not trivial and can be useful for investigation of set cover problems. There exist bounds on precision of greedy algorithm without weights for partial cover construction which do not depend on the cardinality of covered set [1,6,7,8]. We obtain similar bound for the case of weight. The most part of the results obtained for partial covers is generalized on the case of partial decision-relative reducts and partial decision rules for decision tables which, in general case, are inconsistent (a decision table is inconsistent if it has equal rows with diﬀerent decisions). In particular, we show that – Under some natural assumptions on the class N P greedy algorithms with weights are close to best polynomial approximate algorithms for minimization of total weight of attributes in partial reducts and partial decision rules. – Based on information receiving during greedy algorithm work it is possible to obtain nontrivial lower bounds on minimal total weight of attributes in partial reducts and partial decision rules. – There exist polynomial modiﬁcations of greedy algorithms which for a part of decision tables give better results than usual greedy algorithms. Obtained results will further to more wide use of greedy algorithms with weighs and their modiﬁcations in rough set theory and applications. This paper is, in some sense, an extension of [9] on the case of weights which are not equal to 1. However, problems considered in this paper (and proofs of results) are more complicated than the ones considered in [9]. Bounds obtained in this paper are sometimes more weak than the corresponding bounds from [9]. We must note also that even if all weights are equal to 1 then results of the work of greedy algorithms considered in this paper can be diﬀerent from the results of the work of greedy algorithms considered in [9]. For example, for case of reducts the number of chosen attributes is the same, but last attributes can diﬀer. The paper consists of ﬁve sections. In Sect. 2 partial covers are studied. In Sect. 3 partial tests (partial superreducts) and partial reducts are investigated. In Sect. 4 partial decision rules are considered. Sect. 5 contains short conclusions.

2 2.1

Partial Covers Main Notions

Let A = {a1 , . . . , an } be a nonempty ﬁnite set. Elements of A are enumerated by numbers 1, . . . , n (in fact we ﬁx a linear order on A). Let S = {Bi }i∈{1,...,m} =

214

M.Ju. Moshkov, M. Piliszczuk, and B. Zielosko

{B1 , . . . , Bm } be a family of subsets of A such that B1 ∪ . . . ∪ Bm = A. We will assume that S can contain equal subsets of A. The pair (A, S) will be called a set cover problem. Let w be a weight function which corresponds to each Bi ∈ S a natural number w(Bi ). The triple (A, S, w) will be called a set cover problem with weights. Note that in fact weight function w is given on the set of indexes {1, . . . , m}. But, for simplicity, we are writing w(Bi ) instead of w(i). Let I be a subset of {1, . . . , m}. The family P = {Bi }i∈I will be called a subfamily of S. The number |P | = |I| will be called the cardinality of P . Let P = {Bi }i∈I and Q = {Bi }i∈J be subfamilies of S. The notation P ⊆ Q will mean that I ⊆ J. Let us denote P ∪ Q = {Bi }i∈I∪J , P ∩ Q = {Bi }i∈I∩J , and P \ Q = {Bi }i∈I\J . A subfamily Q = {Bi1 , . . . , Bit } of the family S will be called a partial cover for (A, S). Let α be a real number such that 0 ≤ α < 1. The subfamily Q will be called an α-cover for (A, S) if |Bi1 ∪ . . . ∪ Bit | ≥ (1 − α)|A|. For example, 0.01-cover means that we must cover at least 99% of elements from A. Note that t a 0-cover is usual (exact) cover. The number w(Q) = j=1 w(Bij ) will be called the weight of the partial cover Q. Let us denote by Cmin (α) = Cmin (α, A, S, w) the minimal weight of α-cover for (A, S). Let α and γ be real numbers such that 0 ≤ γ ≤ α < 1. Let us describe a greedy algorithm with two thresholds α and γ. Let us denote N = |A|(1 − γ) and M = |A|(1 − α) . Let we make i ≥ 0 steps and choose subsets Bj1 , . . . , Bji . Let us describe the step number i + 1. Let us denote D = Bj1 ∪ . . . ∪ Bji (if i = 0 then D = ∅). If |D| ≥ M then we ﬁnish the work of the algorithm. The family {Bj1 , . . . , Bji } is the constructed α-cover. Let |D| < M . Then we choose a subset Bji+1 from S with minimal number ji+1 for which Bji+1 \ D = ∅ and the value w(Bji+1 ) min{|Bji+1 \ D|, N − |D|} is minimal. Pass to the step number i + 2. γ γ Let us denote by Cgreedy (α) = Cgreedy (α, A, S, w) the weight of α-cover constructed by the considered algorithm for the set cover problem with weights (A, S, w). Note that greedy algorithm with two thresholds α and α coincides with the greedy algorithm with one threshold α considered in [18]. 2.2

Some Known Results

In this subsection we assume that the weight function has values from the set of positive real numbers. For natural m denote H(m) = 1 + . . . + 1/m. It is known that ln m ≤ H(m) ≤ ln m + 1 .

On Partial Covers, Reducts and Decision Rules with Weights

215

Consider some results for the case of exact covers where α = 0. In this case γ = 0. First results belong to Chv´ atal. Theorem 1. (Chv´ atal [2]) For any set cover problem with weights (A, S, w) the 0 inequality Cgreedy (0) ≤ Cmin (0)H(|A|) holds. Theorem 2. (Chv´ atal [2]) For any set cover problem with weights (A, S, w) the 0 (0) ≤ Cmin (0)H (maxBi ∈S |Bi |) holds. inequality Cgreedy Chv´ atal proved in [2] that the bounds from Theorems 1 and 2 are almost unimprovable. Consider now some results for the case where α ≥ 0 and γ = α. First upper α (α) was obtained by Kearns. bound on Cgreedy Theorem 3. (Kearns [5]) For any set cover problem with weights (A, S, w) and α any α, 0 ≤ α < 1, the inequality Cgreedy (α) ≤ Cmin (α)(2H(|A|) + 3) holds. This bound was improved by Slav´ık. Theorem 4. (Slav´ık [18]) For any set cover problem with weights (A, S, w) and α (α) ≤ Cmin (α)H ((1 − α)|A| ) holds. any α, 0 ≤ α < 1, the inequality Cgreedy Theorem 5. (Slav´ık [18])) For any set cover problem with weights (A, S, w) and α any α, 0 ≤ α < 1, the inequality Cgreedy (α) ≤ Cmin (α)H (maxBi ∈S |Bi |) holds. Slav´ık proved in [18] that the bounds from Theorems 4 and 5 are unimprovable. 2.3

On Polynomial Approximate Algorithms for Minimization of Partial Cover Weight

In this subsection we consider three theorems which follow immediately from Theorems 13–15 [9]. Let 0 ≤ α < 1. Consider the following problem: for given set cover problem with weights (A, S, w) it is required to ﬁnd an α-cover for (A, S) with minimal weight. Theorem 6. Let 0 ≤ α < 1. Then the problem of construction of α-cover with minimal weight is N P -hard. From this theorem it follows that we must consider polynomial approximate algorithms for minimization of α-cover weight. Theorem 7. Let α ∈ IR and 0 ≤ α < 1. If N P ⊆ DT IM E(nO(log log n) ) then for any ε, 0 < ε < 1, there is no polynomial algorithm that for a given set cover problem with weights (A, S, w) constructs an α-cover for (A, S) which weight is at most (1 − ε)Cmin (α, A, S, w) ln |A|. Theorem 8. Let α be a real number such that 0 ≤ α < 1. If P = N P then there exists δ > 0 such that there is no polynomial algorithm that for a given set cover problem with weights (A, S, w) constructs an α-cover for (A, S) which weight is at most δCmin (α, A, S, w) ln |A|.

216

M.Ju. Moshkov, M. Piliszczuk, and B. Zielosko

α From Theorem 4 it follows that Cgreedy (α) ≤ Cmin (α)(1 + ln |A|). From this inequality and from Theorem 7 it follows that under the assumption N P ⊆ DT IM E(nO(log log n) ) greedy algorithm with two thresholds α and α (in fact greedy algorithm with one threshold α from [18]) is close to best polynomial approximate algorithms for minimization of partial cover weight. From the considered inequality and from Theorem 8 it follows that under the assumption P = N P greedy algorithm with two thresholds α and α is not far from best polynomial approximate algorithms for minimization of partial cover weight. However we can try to improve the results of the work of greedy algorithm with two thresholds α and α for some part of set cover problems with weights.

2.4

Comparison of Greedy Algorithms with One and Two Thresholds

The following example shows that if for greedy algorithm with two thresholds α and γ we will use γ such that γ < α we can obtain sometimes better results than in the case γ = α. Example 1. Consider a set cover problem (A, S, w) such that A = {1, 2, 3, 4, 5, 6}, S = {B1 , B2 }, B1 = {1}, B2 = {2, 3, 4, 5, 6}, w(B1 ) = 1 and w(B2 ) = 4. Let α = 0.5. It means that we must cover at least M = (1 − α)|A| = 3 elements from A. If γ = α = 0.5 then the result of the work of greedy algorithm with thresholds α and γ is the 0.5-cover {B1 , B2 } which weight is equal to 5. If γ = 0 < α then the result of the work of greedy algorithm with thresholds α and γ is the 0.5-cover {B2 } which weight is equal to 4. In this subsection we show that under some assumptions on |A| and |S| for the most part of set cover problems (A, S) there exist a weight function w γ (α, A, S, w) < and real numbers α, γ such that 0 ≤ γ < α < 1 and Cgreedy α Cgreedy (α, A, S, w). First, we consider criterion of existence of such w, α and γ (see Theorem 9). First part of the proof of this criterion is based on a construction similar to considered in Example 1. Let A be a ﬁnite nonempty set and S = {B1 , . . . , Bm } be a family of subsets of A. We will say that the family S is 1-uniform if there exists a natural number k such that |Bi | = k or |Bi | = k + 1 for any nonempty subset Bi from S. We will say that S is strongly 1-uniform if S is 1-uniform and for any subsets Bl1 , . . . , Blt from S the family {B1 \ U, . . . , Bm \ U } is 1-uniform where U = Bl1 ∪ . . . ∪ Blt . Theorem 9. Let (A, S) be a set cover problem. Then the following two statements are equivalent: 1. The family S is not strongly 1-uniform. 2. There exist a weight function w and real numbers α and γ such that 0 ≤ γ < γ α (α, A, S, w) < Cgreedy (α, A, S, w). α < 1 and Cgreedy Proof. Let S = {B1 , . . . , Bm }. Let the family S be not strongly 1-uniform. Let us choose minimal number of subsets Bl1 , . . . , Blt from the family S (it is possible

On Partial Covers, Reducts and Decision Rules with Weights

217

that t = 0) such that the family {B1 \ U, . . . , Bm \ U } is not 1-uniform where U = Bl1 ∪ . . . ∪ Blt (if t = 0 then U = ∅). Since {B1 \ U, . . . , Bm \ U } is not 1-uniform, there exist two subsets Bi and Bj from S such that |Bi \ U | > 0 and |Bj \ U | ≥ |Bi \ U | + 2. Let us choose real α and γ such that M = |A|(1 − α) = |U |+|Bi \U |+1 and N = |A|(1 − γ) = |U |+|Bi \U |+2. It is clear that 0 ≤ γ < α < 1. Let us deﬁne a weight function w as follows: w(Bl1 ) = . . . = w(Blt ) = 1, w(Bi ) = |A| · 2|Bi \ U |, w(Bj ) = |A|(2|Bi \ U | + 3) and w(Br ) = |A|(3|Bi \ U | + 6) for any Br from S such that r ∈ / {i, j, l1 , . . . , lt }. Let us consider the work of greedy algorithm with two thresholds α and α. One can show that during ﬁrst t steps the greedy algorithm will choose subsets Bl1 , . . . , Blt (may be in an another order). It is clear that |U | < M . Therefore the greedy algorithm must make the step number t + 1. During this step the greedy algorithm will choose a subset Bk from S with minimal number k for w(Bk ) k) which Bk \ U = ∅ and the value p(k) = min{|Bkw(B \U|,M−|U|} = min{|Bk \U|,|Bi \U|+1} is minimal. 1 It is clear that p(i) = 2|A|, p(j) = (2 + |Bi \U|+1 )|A| and p(k) > 3|A| for any subset Bk from S such that Bk \ U = ∅ and k ∈ / {i, j, l1 , . . . , lt }. Therefore during the step number t + 1 the greedy algorithm will choose the subset Bi . Since |U | + |Bi \ U | = M − 1, the greedy algorithm will make the step number t + 2 and will choose a subset from S which is diﬀerent from Bl1 , . . . , Blt , Bi . As α (α, A, S, w) ≥ t + |A| · 2|Bi \ U | + |A|(2|Bi \ U | + 3). the result we obtain Cgreedy Let us consider the work of greedy algorithm with two thresholds α and γ. One can show that during ﬁrst t steps the greedy algorithm will choose subsets Bl1 , . . . , Blt (may be in an another order). It is clear that |U | < M . Therefore the greedy algorithm must make the step number t + 1. During this step the greedy algorithm will choose a subset Bk from S with minimal number k for w(Bk ) k) which Bk \ U = ∅ and the value q(k) = min{|Bw(B = min{|Bk \U|,|B i \U|+2} k \U|,N −|U|} is minimal. 1 )|A| and q(k) ≥ 3|A| for It is clear that q(i) = 2|A|, q(j) = (2 − |Bi \U|+2 any subset Bk from S such that Bk \ U = ∅ and k ∈ / {i, j, l1 , . . . , lt }. Therefore during the step number t + 1 the greedy algorithm will choose the subset Bj . Since |U | + |Bj \ U | > M , the α-cover constructed by greedy algorithm will γ be equal to {Bl1 , . . . , Blt , Bj }. As the result we obtain Cgreedy (α, A, S, w) = α t+|A|(2|Bi \U |+3). Since Cgreedy (α, A, S, w) ≥ t+|A|·2|Bi \U |+|A|(2|Bi \U |+3) γ α (α, A, S, w) > Cgreedy (α, A, S, w). and |Bi \ U | > 0, we conclude that Cgreedy Let the family S be strongly 1-uniform. Consider arbitrary weight function w for S and real numbers α and γ such that 0 ≤ γ < α < 1. Let us show that γ α (α, A, S, w) ≥ Cgreedy (α, A, S, w). Let us denote M = |A|(1 − α) and Cgreedy γ α N = |A|(1 − γ) . If M = N then Cgreedy (α, A, S, w) = Cgreedy (α, A, S, w). Let N > M. Let us apply the greedy algorithm with thresholds α and α to the set cover problem with weights (A, S, w). Let during the construction of α-cover this algorithm choose sequentially subsets Bg1 , . . . , Bgt . Let us apply now the greedy algorithm with thresholds α and γ to the set cover problem with weights (A, S, w). If during the construction of α-cover this algorithm chooses sequentially subsets

218

M.Ju. Moshkov, M. Piliszczuk, and B. Zielosko

γ α Bg1 , . . . , Bgt then Cgreedy (α, A, S, w) = Cgreedy (α, A, S, w). Let there exist a nonnegative integer r, 0 ≤ r ≤ t − 1, such that during ﬁrst r steps the considered algorithm chooses subsets Bg1 , . . . , Bgr , but at the step number r + 1 the algorithm chooses a subset Bk such that k = gr+1 . Let us denote Bg0 = ∅, D = Bg0 ∪. . .∪Bgr and J = {i : i ∈ {1, . . . , m}, Bi \D = ∅}. It is clear that gr+1 , k ∈ J. w(Bi ) i) For any i ∈ J denote p(i) = min{|Biw(B \D|,M−|D|} , q(i) = min{|Bi \D|,N −|D|} . Since k = gr+1 , we conclude that there exists i ∈ J such that p(i) = q(i). Therefore |Bi \ D| > M − |D|. Since S is strongly 1-uniform family, we have |Bj \ D| ≥ M − |D| for any j ∈ J. From here it follows, in particular, that r + 1 = t, and {Bg1 , . . . , Bgt−1 , Bk } is an α-cover for (A, S). It is clear that p(gt ) ≤ p(k). Since |Bk \D| ≥ M −|D| and |Bgt \D| ≥ M −|D|, w(Bgt ) w(Bk ) we have p(k) = M−|D| , p(gt ) = M−|D| . Therefore w(Bgt ) ≤ w(Bk ). γ Taking into account that Cgreedy (α, A, S, w) = w(Bg1 ) + . . . + w(Bgt−1 ) + α w(Bk ) and Cgreedy (α, A, S, w) = w(Bg1 ) + . . . + w(Bgt−1 ) + w(Bgt ) we obtain γ α (α, A, S, w).

Cgreedy (α, A, S, w) ≥ Cgreedy

Let us show that under some assumptions on |A| and |S| the most part of set cover problems (A, S) is not 1-uniform, and therefore is not strongly 1-uniform. There is one-to-one correspondence between set cover problems and tables ﬁlled by numbers from {0, 1} and having no rows ﬁlled by 0 only. Let A = {a1 , . . . , an } and S = {B1 , . . . , Bm }. Then the problem (A, S) corresponds to the table with n rows and m columns which for i = 1, . . . , n and j = 1, . . . , m has 1 at the intersection of i-th row and j-th column if and only if ai ∈ Bj . A table ﬁlled by numbers from {0, 1} will be called SC-table if this table has no rows ﬁlled by 0 only. For completeness of the presentation we consider here a statement from [9] with proof. Lemma 1. The number of SC-tables with n rows and m columns is at least 2mn − 2mn−m+log2 n . Proof. Let i ∈ {1, . . . , n}. The number of tables in which the i-th row is ﬁlled by 0 only is equal to 2mn−m . Therefore the number of tables which are not SCtables is at most n2mn−m = 2mn−m+log2 n . Thus, the number of SC-tables is at least 2mn − 2mn−m+log2 n .

n/2

Lemma 2. Let n ∈ IN, n ≥ 4 and k ∈ {0, . . . , n}. Then Cnk ≤ Cn

q(li+1 ). Using inequality p(gi+1 ) ≥ q(gi+1 ) we obtain p(gi+1 ) > q(li+1 ) and p(li+1 ) > q(li+1 ). From last inequality it follows that |Bli+1 \ D| > M − |D|. Let gi+1 > li+1 . In this case we have p(gi+1 ) < p(li+1 ) and q(gi+1 ) ≥ q(li+1 ). Using inequality p(gi+1 ) ≥ q(gi+1 ) we obtain p(gi+1 ) ≥ q(li+1 ) and p(li+1 ) > q(li+1 ). From last inequality it follows that |Bli+1 \ D| > M − |D|. So in any case we have |Bli+1 \ D| > M − |D|. From this inequality it follows that after the step number i+1 the algorithm with thresholds α and γ must ﬁnish

the work. Thus, k = i + 1, k ≤ t, (l1 , . . . , lk−1 ) = (g1 , . . . , gk−1 ) and lk = gk .

2.6

Lower Bound on Cmin (α)

In this subsection we ﬁx some information about the work of greedy algorithm with two thresholds and ﬁnd the best lower bound on the value Cmin (α) depending on this information. Let (A, S, w) be a set cover problem with weights and α, γ be real numbers such that 0 ≤ γ ≤ α < 1. Let us apply the greedy algorithm with thresholds α and γ to the set cover problem with weights (A, S, w). Let during the construction of α-cover the greedy algorithm choose sequentially subsets Bg1 , . . . , Bgt . Let us denote Bg0 = ∅ and δ0 = 0. For i = 1, . . . , t denote δi = |Bgi \ (Bg0 ∪ . . . ∪ Bgi−1 )| and wi = w(Bgi ).

On Partial Covers, Reducts and Decision Rules with Weights

221

As information on the greedy algorithm work we will use numbers MC = MC (α, γ, A, S, w) = |A|(1 − α) and NC = NC (α, γ, A, S, w) = |A|(1 − γ) , and tuples ΔC = ΔC (α, γ, A, S, w) = (δ1 , . . . , δt ) and WC = WC (α, γ, A, S, w) = (w1 , . . . , wt ). For i = 0, . . . , t − 1 denote wi+1 (MC − (δ0 + . . . + δi )) ρi = . min{δi+1 , NC − (δ0 + . . . + δi )} Let us deﬁne parameter ρC (α, γ) = ρC (α, γ, A, S, w) as follows: ρC (α, γ) = max {ρi : i = 0, . . . , t − 1} . We will prove that ρC (α, γ) is the best lower bound on Cmin (α) depending on MC , NC , ΔC and WC . This lower bound is based on a generalization of the following simple reasoning: if we must cover M elementsand the maximal subsets. cardinality of a subset from S is δ then we must use at least M δ Theorem 11. For any set cover problem with weights (A, S, w) and any real numbers α, γ, 0 ≤ γ ≤ α < 1, the inequality Cmin (α, A, S, w) ≥ ρC (α, γ, A, S, w) holds, and there exists a set cover problem with weights (A , S , w ) such that MC (α, γ, A , S , w ) = MC (α, γ, A, S, w), NC (α, γ, A , S , w ) = NC (α, γ, A, S, w) ΔC (α, γ, A , S , w ) = ΔC (α, γ, A, S, w), WC (α, γ, A , S , w ) = WC (α, γ, A, S, w) ρC (α, γ, A , S , w ) = ρC (α, γ, A, S, w), Cmin (α, A , S , w ) = ρC (α, γ, A , S , w ) . Proof. Let (A, S, w) be a set cover problem with weights, S = {B1 , . . . , Bm }, and α, γ be real numbers such that 0 ≤ γ ≤ α < 1. Let us denote M = MC (α, γ, A, S, w) = |A|(1 − α) and N = NC (α, γ, A, S, w) = |A|(1 − γ) . Let {Bl1 , . . . , Blk } be an optimal α-cover for (A, S, w), i.e. w(Bl1 ) + . . . + w(Blk ) = Cmin (α, A, S, w) = Cmin (α) and |Bl1 ∪ . . . ∪ Blk | ≥ M . Let us apply the greedy algorithm with thresholds α and γ to (A, S, w). Let during the construction of α-cover the greedy algorithm choose sequentially subsets Bg1 , . . . , Bgt . Let us denote Bg0 = ∅. Let i ∈ {0, . . . , t − 1}. Let us denote D = Bg0 ∪ . . . ∪ Bgi . It is clear that after i steps of greedy algorithm work in the set Bl1 ∪ . . . ∪ Blk at least |Bl1 ∪ . . .∪ Blk | − |Bg0 ∪ . . . ∪ Bgi | ≥ M − |D| > 0 elements remained uncovered. After i-th step p1 = |Bl1 \ D| elements remained uncovered in the set Bl1 , ..., and pk = |Blk \ D| elements remained uncovered in the set Blk . We know that p1 + . . . + pk ≥ M − |D| > 0. Let, for the deﬁniteness, p1 > 0, . . . , pr > 0, pr+1 = . . . = pk = 0. For j = 1, . . . , r denote qj = min{pj , N −|D|}. It is clear that N −|D| ≥ M −|D|. w(B ) w(B ) Therefore q1 + . . . + qr ≥ M − |D|. Let us consider numbers q1l1 , . . . , qrlr . w(Bl1 )+...+w(Blr ) . q1 +...+qr w(Blr )qr +...+ > qr

Let us show that at least one of these numbers is at most β = w(B )q

1 l1 Assume the contrary. Then w(Bl1 ) + . . . + w(Blr ) = q1 (q1 + . . . + qr )β = w(Bl1 ) + . . . + w(Blr ) which is impossible.

222

M.Ju. Moshkov, M. Piliszczuk, and B. Zielosko

We know that q1 + . . . + qr ≥ M − |D| and w(Bl1 ) + . . . + w(Blr ) ≤ Cmin (α). min (α) , and there exists j ∈ {1, . . . , k} such that Blj \ D = ∅ Therefore β ≤ CM−|D| w(Blj ) w(Bgi+1 ) min{|Blj \D|,N −|D|} ≤ β. Hence min{|Bgi+1 \D|,N −|D|} w(Bg )(M−|D|) Cmin (α) ≥ min{|Bgi+1 \D|,N −|D|} . i+1

and

≤ β ≤

Cmin (α) M−|D|

and

into account

that Cmin (α) is a natural number we obtain Cmin (α) ≥ Taking w(Bgi+1 )(M−|D|) min{|Bg \D|,N −|D|} = ρi . Since last inequality holds for any i ∈ {0, . . . , t − 1} i+1

and ρC (α, γ) = ρC (α, γ, A, S, w) = max {ρi : i = 0, . . . , t − 1}, we conclude that Cmin (α) ≥ ρC (α, γ). Let us show that this bound is unimprovable depending on MC , NC , ΔC and WC . Let us consider a set cover problem with weights (A , S , w ) where A = A, S = {B1 , . . . , Bm , Bm+1 }, |Bm+1 | = M , Bg1 ∪ . . . ∪ Bgt−1 ⊆ Bm+1 ⊆ Bg1 ∪ . . . ∪ Bgt , w (B1 ) = w(B1 ), . . . , w (Bm ) = w(Bm ) and w (Bm+1 ) = ρC (α, γ). It is clear that MC (α, γ, A , S , w ) = MC (α, γ, A, S, w) = M and NC (α, γ, A , S , w ) = NC (α, γ, A, S, w) = N . We show ΔC (α, γ, A , S , w ) = ΔC (α, γ, A, S, w) and WC (α, γ, A , S , w ) = WC (α, γ, A, S, w). Let us show by induction on i ∈ {1, . . . , t} that for the set cover problem with weights (A , S , w ) at the step number i the greedy algorithm with two thresholds α and γ will choose the subset Bgi . Let us consider the ﬁrst step. (Bm+1 ) ρC (α,γ) Let us denote D = ∅. It is clear that min{|Bwm+1 \D|,N −|D|} = M−|D| . From the deﬁnition of ρC (α, γ) it follows that

w (Bg1 ) min{|Bg1 \D|,N −|D|}

=

w(Bg1 ) min{|Bg1 \D|,N −|D|}

≤

ρC (α,γ) M−|D| .

Using this fact and the inequality g1 < m + 1 it is not diﬃcult to prove that at the ﬁrst step greedy algorithm will choose the subset Bg1 . Let i ∈ {1, . . . , t − 1}. Let us assume that the greedy algorithm made i steps for (A , S , w ) and chose subsets Bg1 , . . . , Bgi . Let us show that at the step i + 1 the subset Bgi+1 will be chosen. Let us denote D = Bg1 ∪ . . . ∪ Bgi . Since Bg1 ∪ . . . ∪ Bgi ⊆ Bm+1 and |Bm+1 | = M , we have |Bm+1 \ D| = M − |D|. (Bm+1 ) ρC (α,γ) Therefore min{|Bwm+1 \D|,N −|D|} = M−|D| . From the deﬁnition of the parameter w (Bg

)

w(Bg

)

ρC (α,γ) i+1 i+1 ρC (α, γ) it follows that min{|Bg \D|,N −|D|} = min{|Bgi+1 \D|,N −|D|} ≤ M−|D| . i+1 Using this fact and the inequality gi+1 < m + 1 it is not diﬃcult to prove that at the step number i + 1 greedy algorithm will choose the subset Bgi+1 . Thus, ΔC (α, γ, A , S , w ) = ΔC (α, γ, A, S, w) and WC (α, γ, A , S , w ) = WC (α, γ, A, S, w). Therefore ρC (α, γ, A , S , w ) = ρC (α, γ, A, S, w) = ρC (α, γ). From been proven it follows that Cmin (α, A , S , w ) ≥ ρC (α, γ, A , S , w ). It is clear that {Bm+1 } is an α-cover for (A , S ) and the weight of {Bm+1 } is equal to ρC (α, γ, A , S , w ). Hence Cmin (α, A , S , w ) = ρC (α, γ, A , S , w ).

Let us consider a property of the parameter ρC (α, γ) which is important for practical use of the bound from Theorem 11. Proposition 2. Let (A, S, w) be a set cover problem with weights and α, γ be real numbers such that 0 ≤ γ ≤ α < 1. Then ρC (α, α, A, S, w) ≥ ρC (α, γ, A, S, w).

On Partial Covers, Reducts and Decision Rules with Weights

223

Proof. Let S = {B1 , . . . , Bm }, M = |A|(1 − α) , N = |A|(1 − γ) , ρC (α, α) = ρC (α, α, A, S, w) and ρC (α, γ) = ρC (α, γ, A, S, w). Let us apply the greedy algorithm with thresholds α and α to (A, S, w). Let during the construction of α-cover this algorithm choose sequentially subsets Bg1 , . . . , Bgt . Let us denote B g0 = ∅. For j = 0, . . . , t − 1 denote Dj = Bg0 ∪ . . . ∪ Bgj and ρC (α, α, j) =

w(Bgj+1 )(M−|Dj |) min{|Bgj+1 \Dj |,M−|Dj |}

. Then ρC (α, α) =

max{ρC (α, α, j) : j = 0, . . . , t − 1}. Apply the greedy algorithm with thresholds α and γ to (A, S, w). Let during the construction of α-cover this algorithm choose sequentially subsets Bl1 , . . . , Blk . From Proposition 1 it follows that either k = t and (l1 , . . . , lk ) = (g1 , . . . , gt ) or k ≤ t, (l1 , . . . , lk−1 ) = (g1 , . . . , gk−1 ) and lk = gk . Let us consider these two cases separately. Let k = t and (l1 , . . . , lk ) = (g1 , . . . , gt ). For j = 0, . . . , t − 1 denote ρC (α, γ, j) =

w(Bgj+1 )(M−|Dj |) min{|Bgj+1 \Dj |,N −|Dj |}

j = 0, . . . , t − 1}. Since N ≥ M , we 0, . . . , t − 1. Hence ρC (α, γ) ≤ ρC (α, α). and lk = gk . Let us denote ρC (α, γ, k

w(B )(M−|Dj |) g ρC (α, γ, j) = min{|Bg j+1\Dj |,N −|Dj |} j+1

. Then ρC (α, γ) = max{ρC (α, γ, j) :

have ρC (α, γ, j) ≤ ρC (α, α, j) for j = Let k ≤ t, (l1 , . . . , lk−1 ) = (g1 , . . . , gk−1 ) w(B k )(M−|Dk−1 |) and − 1) = min{|Bl l\D k−1 |,N −|Dk−1 |} k

for j = 0, . . . , k − 2. Then ρC (α, γ) =

max{ρC (α, γ, j) : j = 0, . . . , k − 1}. Since N ≥ M , we have ρC (α, γ, j) ≤ w(B ) ρC (α, α, j) for j = 0, . . . , k − 2. It is clear that min{|Bl \Dk−1lk|,N −|Dk−1 |} ≤ k

w(Bgk ) min{|Bgk \Dk−1 |,N −|Dk−1 |}

w(B

)

k ≤ min{|Bg \Dk−1g|,M−|D . Thus, ρC (α, γ, k − 1) ≤ k−1 |} k

ρC (α, α, k − 1) and ρC (α, γ) ≤ ρC (α, α).

2.7

γ (α) Upper Bounds on Cgreedy

In this subsection we study some properties of parameter ρC (α, γ) and obtain γ two upper bounds on the value Cgreedy (α) which do not depend directly on cardinality of the set A and cardinalities of subsets Bi from S. Theorem 12. Let (A, S, w) be a set cover problem with weights and α, γ be real numbers such that 0 ≤ γ < α < 1. Then 1−γ γ Cgreedy (α, A, S, w) < ρC (γ, γ, A, S, w) ln +1 . α−γ Proof. Let S = {B1 , . . . , Bm }. Let us denote M = |A|(1 − α) and N = |A|(1 − γ) . Let us apply the greedy algorithm with thresholds γ and γ to (A, S, w). Let during the construction of γ-cover the greedy algorithm choose sequentially subsets Bg1 , . . . , Bgt . Let us denote Bg0 = ∅, for i = 0, . . . , t − 1 denote Di = Bg0 ∪ . . . ∪ Bgi , and denote ρ = ρC (γ, γ, A, S, w). Immediately from the deﬁnition of the parameter ρ it follows that for i = 0, . . . , t − 1 w(Bgi+1 ) ρ ≤ . min{|Bgi+1 \ Di |, N − |Di |} N − |Di |

(1)

224

M.Ju. Moshkov, M. Piliszczuk, and B. Zielosko

Note that min{|Bgi+1 \ Di |, N − |Di |} = |Bgi+1 \ Di | for i = 0, . . . , t − 2 since {Bg0 , . . . , Bgi+1 } is not a γ-cover for (A, S). Therefore for i = 0, . . . , t − 2 we have

w(Bgi+1 ) |Bgi+1 \Di |

≤

ρ N −|Di |

and

N −|Di | ρ

≤

|Bgi+1 \Di | w(Bgi+1 ) .

Thus, for i = 1, . . . , t − 1

i−1 | during the step number i the greedy algorithm covers at least N −|D elements ρ on each unit of weight. From (1) it follows that that for i = 0, . . . , t − 1

w(Bgi+1 ) ≤

ρ min{|Bgi+1 \ Di |, N − |Di |} ≤ρ . N − |Di |

(2)

Assume that ρ = 1. Using (2) we obtain w(Bg1 ) = 1. From this equality and (1) it follows that |Bg1 | ≥ N . Therefore {Bg1 } is an α-cover for (A, S), and

γ 1−γ + 1 > 1. Therefore the statement of the (α) = 1. It is clear that ln α−γ Cgreedy theorem holds if ρ = 1. Assume now that ρ ≥ 2. Let |Bg1 | ≥ M . Then {B g1 } is an α-cover for (A, S). γ 1−γ Using (2) we obtain Cgreedy (α) ≤ ρ. Since ln α−γ +1 > 1, we conclude that the statement of the theorem holds if |Bg1 | ≥ M . Let |Bg1 | < M . Then there exists q ∈ {1, . . . , t − 1} such that |Bg1 ∪ . . . ∪ Bgq | < M and |Bg1 ∪ . . . ∪ Bgq+1 | ≥ M . Taking into account that for i = 1, . . . , q during the step number i the greedy i−1 | algorithm covers at least N −|D elements on each unit of weight we obtain ρ w(Bg1 )+...+w(Bgq ) . Let us denote k = w(Bg1 ) + N − |Bg1 ∪ . . . ∪ Bgq | ≤ N 1 − 1ρ k . . . + w(Bgq ). Then N − N 1 − 1ρ ≤ |Bg1 ∪ . . . ∪ Bgq | ≤ M − 1. Therefore k k |A|(1 − γ) − |A|(1 − γ) 1 − ρ1 < |A|(1 − α), 1 − γ − 1 + α < (1 − γ) ρ−1 , ρ k k ρ 1−γ 1−γ 1 . To obtain last inequality < α−γ , 1 + ρ−1 < 1−γ , and kρ < ln α−γ ρ−1

α−γ 1 1 we use known inequality ln 1 + r > r+1 which holds for any natural r. It is γ clear that Cgreedy (α) = k + w(Bq+1 ). Using (2) we conclude that w(Bq+1 ) ≤ ρ. γ 1−γ + ρ.

Therefore Cgreedy (α) < ρ ln α−γ

Corollary 1. Let ε be a real number, and 0 < ε < 1. Then for any α such that ε ≤ α < 1 the following inequalities hold: 1 α−ε ρC (α, α) ≤ Cmin (α) ≤ Cgreedy (α) < ρC (α − ε, α − ε) ln + 1 . ε For example, if ε = 0.01 and 0.01 ≤ α < 1 then ρC (α, α) ≤ Cmin (α) ≤ α−0.01 Cgreedy (α) < 5.61ρC (α − 0.01, α − 0.01), and if ε = 0.1 and 0.1 ≤ α < 1 α−0.1 then ρC (α, α) ≤ Cmin (α) ≤ Cgreedy (α) < 3.31ρC (α − 0.1, α − 0.1). The obtained results show that the lower bound Cmin (α) ≥ ρC (α, α) is nontrivial. Theorem 13. Let (A, S, w) be a set cover problem with weights and α, γ be real numbers such that 0 ≤ γ < α < 1. Then 1−γ γ Cgreedy (α, A, S, w) < Cmin (γ, A, S, w) ln +1 . α−γ

On Partial Covers, Reducts and Decision Rules with Weights

225

γ Proof. From Theorem 12 it follows that Cgreedy (α, A, S, w) < ρC (γ, γ, A, S, w) · 1−γ ln α−γ + 1 . The inequality ρC (γ, γ, A, S, w) ≤ Cmin (γ, A, S, w) follows from Theorem 11.

0 0.001 Corollary 2. Cgreedy (0.001) < 7.91Cmin(0), Cgreedy (0.01) < 5.71Cmin(0.001), 0.1 0.3 Cgreedy (0.2) < 3.20Cmin(0.1), Cgreedy (0.5) < 2.26Cmin (0.3).

0 Corollary 3. Let 0 < α < 1. Then Cgreedy (α) < Cmin (0) ln α1 + 1 . Corollary 4. Let ε be a real number, and 0 < ε < 1. Then for any α such that α−ε ε ≤ α < 1 the inequalities Cmin (α) ≤ Cgreedy (α) < Cmin (α − ε) ln 1ε + 1 hold.

3 3.1

Partial Tests and Reducts Main Notions

Let T be a table with n rows labeled by nonnegative integers (decisions) and m columns labeled by attributes (names of attributes) f1 , . . . , fm . This table is ﬁlled by nonnegative integers (values of attributes). The table T is called a decision table. Let w be a weight function for T which corresponds to each attribute fi a natural number w(fi ). Let us denote by P (T ) the set of unordered pairs of diﬀerent rows of T with diﬀerent decisions. We will say that an attribute fi separates a pair of rows (r1 , r2 ) ∈ P (T ) if rows r1 and r2 have diﬀerent numbers at the intersection with the column fi . For i = 1, . . . , m denote by P (T, fj ) the set of pairs from P (T ) which the attribute fi separates. Let α be a real number such that 0 ≤ α < 1. A set of attributes Q ⊆ {f1 , . . . , fm } will be called an α-test for T if attributes from Q separate at least (1 − α)|P (T )| pairs from the set P (T ). An α-test is called an α-reduct if each proper subset of the considered α-test is not α-test. If P (T ) = ∅ then each subset of {f1 , . . . , fm } is an α-test, and only empty set is an α-reduct. For example, 0.01-test means that we must separate at least 99% of pairs from P (T ). Note that 0-reduct is usual (exact) reduct. It must be noted also that each α-test contains at least one α-reduct as a subset. The number w(Q) = fi ∈Q w(fi ) will be called the weight of the set Q. If Q = ∅ then w(Q) = 0. Let us denote by Rmin (α) = Rmin (α, T, w) the minimal weight of α-reduct for T . It is clear that Rmin (α, T, w) coincides with the minimal weight of α-test for T . Let α, γ be real numbers such that 0 ≤ γ ≤ α < 1. Let us describe a greedy algorithm with thresholds α and γ which constructs an α-test for given decision table T and weight function w. If P (T ) = ∅ then the constructed α-test is empty set. Let P (T ) = ∅. Let us denote M = |P (T )|(1 − α) and N = |P (T )|(1 − γ) . Let we make i ≥ 0

226

M.Ju. Moshkov, M. Piliszczuk, and B. Zielosko

steps and construct a set Q containing i attributes (if i = 0 then Q = ∅). Let us describe the step number i + 1. Let us denote by D the set of pairs from P (T ) separated by attributes from Q (if i = 0 then D = ∅). If |D| ≥ M then we ﬁnish the work of the algorithm. The set of attributes Q is the constructed α-test. Let |D| < M . Then we choose an attribute fj with minimal number j for which P (T, fj ) \ D = ∅ and the value w(fj ) min{|P (T, fj ) \ D|, N − |D|} is minimal. Add the attribute fj to the set Q. Pass to the step number i + 2. γ γ Let us denote by Rgreedy (α) = Rgreedy (α, T, w) the weight of α-test constructed by greedy algorithm with thresholds α and γ for given decision table T and weight function w. 3.2

Relationships Between Partial Covers and Partial Tests

Let (A, S, w) be a set cover problem with weights and α, γ be real numbers such that 0 ≤ γ ≤ α < 1. Let us apply the greedy algorithm with thresholds α and γ to (A, S, w). Let during the construction of α-cover the greedy algorithm choose sequentially subsets Bj1 , . . . , Bjt from the family S. Let us denote OC (α, γ, A, S, w) = (j1 , . . . , jt ). Let T be a decision table with m columns labeled by attributes f1 , . . . , fm , and with a nonempty set P (T ). Let w be a weight function for T . We correspond a set cover problem with weights (A(T ), S(T ), uw ) to the considered decision table T and weight function w in the following way: A(T ) = P (T ), S(T ) = {B1 (T ), . . . , Bm (T )} where B1 (T ) = P (T, f1 ), . . . , Bm (T ) = P (T, fm ), uw (B1 (T )) = w(f1 ), . . . , uw (Bm (T )) = w(fm ). Let α, γ be real numbers such that 0 ≤ γ ≤ α < 1. Let us apply the greedy algorithm with thresholds α and γ to decision table T and weight function w. Let during the construction of α-test the greedy algorithm choose sequentially attributes fj1 , . . . , fjt . Let us denote OR (α, γ, T, w) = (j1 , . . . , jt ). Let us denote P (T, fj0 ) = ∅. For i = 1, . . . , t denote wi = w(fji ) and δi = |P (T, fji ) \ (P (T, fj0 ) ∪ . . . ∪ P (T, fji−1 ))| . Let us denote MR (α, γ, T, w) = |P (T )|(1 − α) , NR (α, γ, T, w) = |P (T )|(1 − γ) , ΔR (α, γ, T, w) = (δ1 , . . . , δt ) and WR (α, γ, T, w) = (w1 , . . . , wt ). It is not diﬃcult to prove the following statement. Proposition 3. Let T be a decision table with m columns labeled by attributes f1 , . . . , fm , P (T ) = ∅, w be a weight function for T , and α, γ be real numbers such that 0 ≤ γ ≤ α < 1. Then |P (T )| = |A(T )| , |P (T, fi )| = |Bi (T )|, i = 1, . . . , m , OR (α, γ, T, w) = OC (α, γ, A(T ), S(T ), uw ) ,

On Partial Covers, Reducts and Decision Rules with Weights

227

MR (α, γ, T, w) = MC (α, γ, A(T ), S(T ), uw ) , NR (α, γ, T, w) = NC (α, γ, A(T ), S(T ), uw ) , ΔR (α, γ, T, w) = ΔC (α, γ, A(T ), S(T ), uw ) , WR (α, γ, T, w) = WC (α, γ, A(T ), S(T ), uw ) , Rmin (α, T, w) = Cmin (α, A(T ), S(T ), uw ) , γ γ (α, T, w) = Cgreedy (α, A(T ), S(T ), uw ) . Rgreedy Let (A, S, w) be a set cover problem with weights where A = {a1 , . . . , an } and S = {B1 , . . . , Bm }. We correspond a decision table T (A, S) and a weight function vw for T (A, S) to the set cover problem with weights (A, S, w) in the following way. The table T (A, S) contains m columns labeled by attributes f1 , . . . , fm and n+ 1 rows ﬁlled by numbers from {0, 1}. For i = 1, . . . , n and j = 1, . . . , m at the intersection of i-th row and j-th column the number 1 stays if and only if ai ∈ Bj . The row number n + 1 is ﬁlled by 0. First n rows are labeled by the decision 0. Last row is labeled by the decision 1. Let vw (f1 ) = w(B1 ), . . . , vw (fm ) = w(Bm ). For i = {1, . . . , n + 1} denote by ri the i-th row. It is not diﬃcult to see that P (T (A, S)) = {(r1 , rn+1 ), . . . , (rn , rn+1 )}. Let i ∈ {1, . . . , n} and j ∈ {1, . . . , m}. One can show that the attribute fj separates the pair (ri , rn+1 ) if and only if ai ∈ Bj . It is not diﬃcult to prove the following statement. Proposition 4. Let (A, S, w) be a set cover problem with weights and α, γ be real numbers such that 0 ≤ γ ≤ α < 1. Then |P (T (A, S))| = |A| , OR (α, γ, T (A, S), vw ) = OC (α, γ, A, S, w) , MR (α, γ, T (A, S), vw ) = MC (α, γ, A, S, w) , NR (α, γ, T (A, S), vw ) = NC (α, γ, A, S, w) , ΔR (α, γ, T (A, S), vw ) = ΔC (α, γ, A, S, w) , WR (α, γ, T (A, S), vw ) = WC (α, γ, A, S, w) , Rmin (α, T (A, S), vw ) = Cmin (α, A, S, w) , γ γ Rgreedy (α, T (A, S), vw ) = Cgreedy (α, A, S, w) .

3.3

On Precision of Greedy Algorithm with Thresholds α and α

The following two statements are simple corollaries of results of Slav´ık (see Theorems 4 and 5) and Proposition 3. Theorem 14. Let T be a decision table, P (T ) = ∅, w be a weight function for α T , α ∈ IR and 0 ≤ α < 1. Then Rgreedy (α) ≤ Rmin (α)H ((1 − α)|P (T )| ). Theorem 15. Let T be a decision table with m columns labeled by attributes f1 , . . . , fm , P (T ) = ∅, w be a weight function for T , and α be a real number α such that 0 ≤ α < 1. Then Rgreedy (α) ≤ Rmin (α)H maxi∈{1,...,m} |P (T, fi )| .

228

M.Ju. Moshkov, M. Piliszczuk, and B. Zielosko

3.4

On Polynomial Approximate Algorithms

In this subsection we consider three theorems which follows immediately from Theorems 26–28 [9]. Let 0 ≤ α < 1. Let us consider the following problem: for given decision table T and weight function w for T it is required to ﬁnd an α-test (α-reduct) for T with minimal weight. Theorem 16. Let 0 ≤ α < 1. Then the problem of construction of α-test (αreduct) with minimal weight is N P -hard. So we must consider polynomial approximate algorithms for minimization of α-test (α-reduct) weight. Theorem 17. Let α ∈ IR and 0 ≤ α < 1. If N P ⊆ DT IM E(nO(log log n) ) then for any ε, 0 < ε < 1, there is no polynomial algorithm that for given decision table T with P (T ) = ∅ and weight function w for T constructs an α-test for T which weight is at most (1 − ε)Rmin (α, T, w) ln |P (T )|. Theorem 18. Let α be a real number such that 0 ≤ α < 1. If P = N P then there exists δ > 0 such that there is no polynomial algorithm that for given decision table T with P (T ) = ∅ and weight function w for T constructs an αtest for T which weight is at most δRmin (α, T, w) ln |P (T )|. α (α) ≤ Rmin (α)(1 + ln |P (T )|). From this From Theorem 14 it follows that Rgreedy inequality and from Theorem 17 it follows that under the assumption N P ⊆ DT IM E(nO(log log n) ) greedy algorithm with two thresholds α and α is close to best polynomial approximate algorithms for minimization of partial test weight. From the considered inequality and from Theorem 18 it follows that under the assumption P = N P greedy algorithm with two thresholds α and α is not far from best polynomial approximate algorithms for minimization of partial test weight. However we can try to improve the results of the work of greedy algorithm with two thresholds α and α for some part of decision tables.

3.5

Two Modiﬁcations of Greedy Algorithm

First, we consider binary diagnostic decision tables and prove that under some assumptions on the number of attributes and rows for the most part of tables there exist weight function w and numbers α, γ such that the weight of α-test constructed by greedy algorithm with thresholds α and γ is less than the weight of α-test constructed by greedy algorithm with thresholds α and α. Binary means that the table is ﬁlled by numbers from the set {0, 1} (all attributes have values from {0, 1}). Diagnostic means that rows of the table are labeled by pairwise diﬀerent numbers (decisions). Let T be a binary diagnostic decision table with m columns labeled by attributes f1 , . . . , fm and with n rows. We will assume that rows of T with numbers 1, . . . , n are labeled by decisions 1, . . . , n respectively. Therefore the number of considered tables is equal to 2mn . Decision table will be called simple if it has no equal rows.

On Partial Covers, Reducts and Decision Rules with Weights

229

Theorem 19. Let us consider binary diagnostic decision tables with m columns labeled by attributes f1 , . . . , fm and n ≥ 4 rows labeled by decisions 1, . . . , n. The fraction of decision tables T for each of which there exist a weight function w and γ α (α, T, w) < Rgreedy (α, T, w) is numbers α, γ such that 0 ≤ γ < α < 1 and Rgreedy at least 1 −

3m n

m −1 2

−

n2 2m .

Proof. We will say that a decision table T is not 1-uniform if there exist two attributes fi and fj of T such that |P (T, fi )| > 0 and |P (T, fj )| ≥ |P (T, fi )| + 2. Otherwise, we will say that T is 1-uniform. Using Theorem 9 and Proposition 3 we conclude that if T is not 1-uniform then there exist a weight function w and γ α (α, T, w) < Rgreedy (α, T, w). numbers α, γ such that 0 ≤ γ < α < 1 and Rgreedy We evaluate the number of simple decision tables which are 1-uniform. Let us consider a simple decision table T which is 1-uniform. Let fi be an attribute of T . It is clear that |P (T, fi )| = 0 if and only if the number of units in the column fi is equal to 0 or n. Let k, l be natural numbers such that k, k + l ∈ {1, . . . , n − 1}, and i, j ∈ {1, . . . , m}, i = j. Let the decision table T have k units in the column fi and k + l units in the column fj . Then |P (T, fi )| = k(n − k) = kn − k 2 and |P (T, fj )| = (k + l)(n − k − l) = kn − k 2 + l(n − 2k − l). Since T is 1-uniform, we have l(n − 2k − l) ∈ {0, 1, −1}. Let l(n − 2k − l) = 0. Then n − 2k − l = 0 and l = n − 2k. Since l is a natural number, we have k < n/2. Let l(n − 2k − l) = 1. Since l, n and k are natural numbers, we have l = 1 and n − 2k − 1 = 1. Therefore k = n2 − 1. Since k is a natural number, we have n is even. Let l(n − 2k − l) = −1. Since l, n and k are natural numbers, we have l = 1 and n − 2k − 1 = −1. Therefore k = n2 . Since k is a natural number, we have n is even. Let n be odd. Then there exists natural k such that 1 ≤ k < n2 and the number of units in each column of T belongs to the set {0, n, k, n − k}. Therefore the n/2 number of considered tables is at most k=1 (Cnk + Cnn−k + 2)m . Since n ≥ 4, n/2 we have 2 ≤ Cn . Using Lemma 2 we concludethat the number of 1-uniform m n/2 n/2 m n √ simple tables is at most k=1 3Cn < n 3·2 . n Let n be even. Then there exists natural k such that 1 ≤ k < n2 − 1 and the number of units in each column of T belongs to the set {0, n, k, n − k}, or the number of units in each column belongs to the set {0, n, n2 − 1, n2 , n2 + 1}. n/2−2 k (Cn + Cnn−k + Therefore the number of considered tables is at most k=1 n/2−1 n/2 n/2+1 m m 2) + (Cn + Cn + Cn + 2) . It is well known (see, for example, [25], n/2 r page 178) that Cn < Cn for anyr ∈ {1,. . . , n} \ {n/2}. Therefore the number n/2

of 1-uniform tables is at most n 3Cn

m

. Using Lemma 2 we conclude that

(as of odd n) the number of 1-uniform simple tables is less than m case mn inn the 3·2 2 3m n √n = m2 −1 . The number of tables which are not simple is at most n

n2 2mn−m . Therefore the number of tables which are not 1-uniform is at least

230

M.Ju. Moshkov, M. Piliszczuk, and B. Zielosko mn m 3 m −1 2

2mn − 2

n

− n2 2mn−m . Thus, the fraction, considered in the statement of the

theorem, is at least 1 −

3m n

m −1 2

−

n2 2m .

m

So if m ≥ 4 and n, 2n2 are large enough then for the most part of binary diagnostic decision tables there exist weight function w and numbers α, γ such that the weight of α-test constructed by greedy algorithm with thresholds α and γ is less than the weight of α-test constructed by greedy algorithm with thresholds α and α. The obtained results show that the greedy algorithm with two thresholds α and γ is of some interest. Now we consider two polynomial modiﬁcations of greedy algorithm which allow to use advantages of greedy algorithm with two thresholds α and γ. Let T be a decision table, P (T ) = ∅, w be a weight function for T and α be a real number such that 0 ≤ α < 1. 1. It is impossible to consider eﬀectively all γ such that 0 ≤ γ ≤ α. Instead of this we can consider all natural N such that M ≤ N ≤ |P (T )| where M = |P (T )|(1 − α) (see the description of greedy algorithm with two thresholds). For each N ∈ {M, . . . , |P (T )|} we apply greedy algorithm with parameters M and N to T and w and after that choose an α-test with minimal weight among constructed α-tests. 2. There exists also an another way to construct an α-test which is not worse than the one obtained under consideration of all N such that M ≤ N ≤ |P (T )|. Let us apply greedy algorithm with thresholds α and α to T and w. Let the algorithm choose sequentially attributes fj1 , . . . , fjt . For each i ∈ {0, . . . , t − 1} we ﬁnd (if it is possible) an attribute fli of T with minimal weight w(fli ) such that the set {fj1 , . . . , fji , fli } is an α-test for T (if i = 0 then it will be the set {fl0 }). After that among constructed α-tests {fj1 , . . . , fjt }, ..., {fj1 , . . . , fji , fli }, ... we choose an α-test with minimal weight. From Proposition 5 it follows that the constructed α-test is not worse than the one constructed under consideration of all γ, 0 ≤ γ ≤ α, or (which is the same) all N , M ≤ N ≤ |P (T )|. Next statement follows immediately from Propositions 1 and 3. Proposition 5. Let T be a decision table, P (T ) = ∅, w be a weight function for T and α, γ be real numbers such that 0 ≤ γ < α < 1. Let the greedy algorithm with two thresholds α and α, which is applied to T and w, choose sequentially attributes fg1 , . . . , fgt . Let the greedy algorithm with two thresholds α and γ, which is applied to T and w, choose sequentially attributes fl1 , . . . , flk . Then either k = t and (l1 , . . . , lk ) = (g1 , . . . , gt ) or k ≤ t, (l1 , . . . , lk−1 ) = (g1 , . . . , gk−1 ) and lk = gk . 3.6

Bounds on Rmin (α) and Rγgreedy (α)

First, we ﬁx some information about the work of greedy algorithm with two thresholds and ﬁnd the best lower bound on the value Rmin (α) depending on this information.

On Partial Covers, Reducts and Decision Rules with Weights

231

Let T be a decision table such that P (T ) = ∅, w be a weight function for T , and α, γ be real numbers such that 0 ≤ γ ≤ α < 1. Let us apply the greedy algorithm with thresholds α and γ to the decision table T and the weight function w. Let during the construction of α-test the greedy algorithm choose sequentially attributes fg1 , . . . , fgt . Let us denote P (T, fg0 ) = ∅ and δ0 = 0. For i = 1, . . . , t denote δi = |P (T, fgi ) \ (P (T, fg0 ) ∪ . . . ∪ P (T, fgi−1 ))| and wi = w(fgi ). As information on the greedy algorithm work we will use numbers MR = MR (α, γ, T, w) = |P (T )|(1 − α) and NR = NR (α, γ, T, w) = |P (T )|(1 − γ) , and tuples ΔR = ΔR (α, γ, T, w) = (δ1 , . . . , δt ) and WR = WR (α, γ, T, w) = (w1 , . . . , wt ). For i = 0, . . . , t − 1 denote wi+1 (MR − (δ0 + . . . + δi )) ρi = . min{δi+1 , NR − (δ0 + . . . + δi )} Let us deﬁne parameter ρR (α, γ) = ρR (α, γ, T, w) as follows: ρR (α, γ) = max {ρi : i = 0, . . . , t − 1} . We will show that ρR (α, γ) is the best lower bound on Rmin (α) depending on MR , NR , ΔR and WR . Next statement follows from Theorem 11 and Propositions 3 and 4. Theorem 20. For any decision table T with P (T ) = ∅, any weight function w for T , and any real numbers α, γ, 0 ≤ γ ≤ α < 1, the inequality Rmin (α, T, w) ≥ ρR (α, γ, T, w) holds, and there exist a decision table T and a weight function w for T such that MR (α, γ, T , w ) = MR (α, γ, T, w), NR (α, γ, T , w ) = NR (α, γ, T, w) , ΔR (α, γ, T , w ) = ΔR (α, γ, T, w), WR (α, γ, T , w ) = WR (α, γ, T, w) , ρR (α, γ, T , w ) = ρR (α, γ, T, w), Rmin (α, T , w ) = ρR (α, γ, T , w ) . Let us consider a property of the parameter ρR (α, γ) which is important for practical use of the bound from Theorem 20. Next statement follows from Propositions 2 and 3. Proposition 6. Let T be a decision table with P (T ) = ∅, w be a weight function for T , α, γ ∈ IR and 0 ≤ γ ≤ α < 1. Then ρR (α, α, T, w) ≥ ρR (α, γ, T, w). Now we study some properties of parameter ρR (α, γ) and obtain two upper γ bounds on the value Rgreedy (α) which do not depend directly on cardinality of the set P (T ) and cardinalities of subsets P (T, fi ). Next statement follows from Theorem 12 and Proposition 3. Theorem 21. Let T be a decision table with P (T ) = ∅, w be a weight function for T and α, γ be real numbers such that 0 ≤ γ < α < 1. Then 1−γ γ Rgreedy (α, T, w) < ρR (γ, γ, T, w) ln +1 . α−γ

232

M.Ju. Moshkov, M. Piliszczuk, and B. Zielosko

Corollary 5. Let ε ∈ IR and 0 < ε < 1. Then for any α, ε ≤ α < 1, the α−ε inequalities ρC (α, α) ≤ Rmin (α) ≤ Rgreedy (α) < ρR (α − ε, α − ε) ln 1ε + 1 hold.

1 1 For example, ln 0.01 + 1 < 5.61 and ln 0.1 + 1 < 3.31. The obtained results show that the lower bound Rmin (α) ≥ ρR (α, α) is nontrivial. Next statement follows from Theorem 13 and Proposition 3. Theorem 22. Let T be a decision table with P (T ) = ∅, w be a weight function for T and α, γ be real numbers such that 0 ≤ γ < α < 1. Then 1−γ γ Rgreedy (α, T, w) < Rmin (γ, T, w) ln +1 . α−γ 0 0.001 (0.001) < 7.91Rmin(0), Rgreedy (0.01) < 5.71Rmin(0.001), Corollary 6. Rgreedy 0.1 0.3 Rgreedy (0.2) < 3.20Cmin(0.1), Rgreedy (0.5) < 2.26Rmin(0.3).

0 Corollary 7. Let 0 < α < 1. Then Rgreedy (α) < Rmin (0) ln α1 + 1 .

Corollary 8. Let ε be a real number, and 0 < ε < 1. Then for any α such that α−ε (α) < Rmin (α − ε) ln 1ε + 1 hold. ε ≤ α < 1 the inequalities Rmin (α) ≤ Rgreedy 3.7

Results of Experiments for α-Tests and α-Reducts

In this subsection we will consider only binary decision tables with binary decision attributes. First Group of Experiments. First group of experiments is connected with study of quality of greedy algorithm with one threshold (where γ = α or, which is the same, N = M ), and comparison of quality of greedy algorithm with one threshold and ﬁrst modiﬁcation of greedy algorithm (where for each N ∈ {M, . . . , |P (T )|} we apply greedy algorithm with parameters M and N to decision table and weight function and after that choose an α-test with minimal weight among constructed α-tests). We generate randomly 1000 decision tables T and weight functions w such that T contains 10 rows and 10 conditional attributes f1 , . . . , f10 , and 1 ≤ w(fi ) ≤ 1000 for i = 1, . . . , 10. For each α ∈ {0.0, 0.1, . . . , 0.9} we ﬁnd the number of pairs (T, w) for which greedy algorithm with one threshold constructs an α-test with minimal weight α (α, T, w) = Rmin (α, T, w). This number is con(an optimal α-test), i.e. Rgreedy tained in the row of Table 1 labeled by ”Opt”. We ﬁnd the number of pairs (T, w) for which ﬁrst modiﬁcation of greedy algorithm constructs an α-test which weight is less than the weight of α-test constructed by greedy algorithm with one threshold, i.e. there exists γ such that γ α (α, T, w) < Rgreedy (α, T, w). This number is contained in 0 ≤ γ < α and Rgreedy the row of Table 1 labeled by ”Impr”. Also we ﬁnd the number of pairs (T, w) for which ﬁrst modiﬁcation of greedy algorithm constructs an optimal α-test which weight is less than the weight of

On Partial Covers, Reducts and Decision Rules with Weights

233

α-test constructed by greedy algorithm with one threshold, i.e. there exists γ γ α such that 0 ≤ γ < α and Rgreedy (α, T, w) = Rmin (α, T, w) < Rgreedy (α, T, w). This number is contained in the row of Table 1 labeled by ”Opt+”. Table 1. Results of ﬁrst group of experiments with α-tests α 0.0 Opt 409 Impr 0 Opt+ 0

0.1 575 42 22

0.2 625 47 28

0.3 826 33 24

0.4 808 24 22

0.5 818 8 5

0.6 950 6 6

0.7 981 5 5

0.8 0.9 992 1000 2 0 2 0

The obtained results show that the percentage of pairs for which greedy algorithm with one threshold ﬁnds an optimal α-test grows almost monotonically (with local minimum near to 0.4–0.5) from 40.9% up to 100%. The percentage of problems for which ﬁrst modiﬁcation of greedy algorithm can improve the result of the work of greedy algorithm with one threshold is less than 5%. However, sometimes (for example, if α = 0.3 or α = 0.7) the considered improvement is noticeable. Second Group of Experiments. Second group of experiments is connected with comparison of quality of greedy algorithm with one threshold and ﬁrst modiﬁcation of greedy algorithm. We make 25 experiments (row ”Nr” in Table 2 contains the number of experiment). Each experiment includes the work with three randomly generated families of pairs (T, w) (1000 pairs in each family) such that T contains n rows and m conditional attributes, and w has values from the set {1, . . . , v}. If the column ”n” contains one number, for example ”40”, it means that n = 40. If this row contains two numbers, for example ”30–120”, it means that for each of 1000 pairs we choose the number n randomly from the set {30, . . . , 120}. The same situation is for the column ”m”. If the column ”α” contains one number, for example ”0.1”, it means that α = 0.1. If this column contains two numbers, for example ”0.2–0.4”, it means that we choose randomly the value of α such that 0.2 ≤ α ≤ 0.4. For each of the considered pairs (T, w) and number α we apply greedy algorithm with one threshold and ﬁrst modiﬁcation of greedy algorithm. Column ”#i”, i = 1, 2, 3, contains the number of pairs (T, w) from the family number i for each of which the weight of α-test, constructed by ﬁrst modiﬁcation of greedy algorithm, is less than the weight of α-test constructed by greedy algorithm with one threshold. In other words, in column ”#i” we have the number of pairs (T, w) from the family number i such that there exists γ for which γ α 0 ≤ γ < α and Rgreedy (α, T, w) < Rgreedy (α, T, w). The column ”avg” contains #1+#2+#3 . the number 3 In experiments 1–3 we consider the case where the parameter v increases. In experiments 4–8 the parameter α increases. In experiments 9–12 the parameter m increases. In experiments 13–16 the parameter n increases. In experiments

234

M.Ju. Moshkov, M. Piliszczuk, and B. Zielosko Table 2. Results of second group of experiments with α-tests Nr n 1 1–50 2 1–50 3 1–50 4 1–50 5 1–50 6 1–50 7 1–50 8 1–50 9 50 10 50 11 50 12 50 13 1–20 14 20–40 15 40–60 16 60–80 17 10 18 10 19 10 20 10 21 10 22 10 23 10 24 10 25 10

m 1–50 1–50 1–50 1–50 1–50 1–50 1–50 1–50 1–20 20–40 40–60 60–80 30 30 30 30 10 10 10 10 10 10 10 10 10

v 1–10 1–100 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000

α 0–1 0–1 0–1 0–0.2 0.2–0.4 0.4–0.6 0.6–0.8 0.8–1 0–0.2 0–0.2 0–0.2 0–0.2 0–0.2 0–0.2 0–0.2 0–0.2 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9

#1 1 5 10 16 23 7 3 1 19 26 21 13 27 34 22 19 36 33 43 30 10 12 3 5 0

#2 2 6 8 20 8 6 5 0 11 24 18 18 26 37 26 14 42 53 25 18 10 13 13 2 0

#3 3 13 11 32 12 5 3 0 22 24 23 22 39 35 23 14 50 46 45 19 13 7 6 6 0

avg 2.00 8.00 9.67 22.67 14.33 6.00 3.67 0.33 17.33 24.67 20.67 17.67 30.67 35.33 23.67 15.67 42.67 44.00 37.67 22.33 11.00 10.67 7.33 4.33 0

17–25 the parameter α increases. The results of experiments show that the value of #i can change from 0 to 53. It means that the percentage of pairs for which ﬁrst modiﬁcation of greedy algorithm is better than the greedy algorithm with one threshold can change from 0% to 5.3%. Third Group of Experiments. Third group of experiments is connected with investigation of quality of lower bound Rmin (α) ≥ ρR (α, α). We choose natural n, m, v and real α, 0 ≤ α < 1. For each chosen tuple (n, m, v, α) we generate randomly 30 pairs (T, w) such that T contains n rows and m conditional attributes, and w has values from the set {1, ..., v}. After α (α, T, w) and ρR (α, α, T, w) for each of generated that we ﬁnd values of Rgreedy α 30 pairs. Note that ρR (α, α, T, w) ≤ Rmin (α, T, w) ≤ Rgreedy (α, T, w). Finally, α we ﬁnd mean values of Rgreedy (α, T, w) and ρR (α, α, T, w) for generated 30 pairs. Results of experiments can be found in Figs. 1 and 2. In these ﬁgures mean values of ρR (α, α, T, w) are called ”average lower bound” and mean values of α Rgreedy (α, T, w) are called ”average upper bound”.

On Partial Covers, Reducts and Decision Rules with Weights

235

In Fig. 1 (left-hand side) one can see the case when n ∈ {1000, 2000, . . . , 5000}, m = 30, v = 1000 and α = 0.01. In Fig. 1 (right-hand side) one can see the case when n = 1000, m ∈ {10, 20, . . . , 100}, v = 1000 and α = 0.01. 1000

2500 average upper bound average lower bound

average upper bound average lower bound

900 2000 800

700

1500

600 1000

500

400 500 300

200 1000

0 1500

2000

2500

3000 number of rows

3500

4000

4500

5000

10

20

30

40

50 60 number of attributes

70

80

90

100

Fig. 1. Results of third group of experiments with α-tests (n and m are changing)

1000

6000 average upper bound average lower bound

average upper bound average lower bound

900 5000 800 700 4000 600 500

3000

400 2000 300 200 1000 100 0 100

0 200

300

400

500 600 range of weight

700

800

900

1000

0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

alpha

Fig. 2. Results of third group of experiments with α-tests (v and α are changing)

In Fig. 2 (left-hand side) one can see the case when n = 1000, m = 30, v ∈ {100, 200, . . . , 1000} and α = 0.01. In Fig. 2 (right-hand side) one can see the case when n = 1000, m = 30, v = 1000 and α ∈ {0.0, 0.1, . . . , 0.9}. Results of experiments show that the considered lower bound is nontrivial and can be useful in investigations.

4

Partial Decision Rules

In this section we omit reasoning on relationships between partial covers and partial decision rules including reductions of one problem to another (description of such reductions can be found in [9]) and two propositions similar to Propositions 3 and 4.

236

4.1

M.Ju. Moshkov, M. Piliszczuk, and B. Zielosko

Main Notions

Let T be a table with n rows labeled by nonnegative integers (decisions) and m columns labeled by attributes (names of attributes) f1 , . . . , fm . This table is ﬁlled by nonnegative integers (values of attributes). The table T is called a decision table. Let w be a weight function for T which corresponds to each attribute fi a natural number w(fi ). Let r = (b1 , . . . , bm ) be a row of T labeled by a decision d. Let us denote by U (T, r) the set of rows from T which are diﬀerent from r and are labeled by decisions diﬀerent from d. We will say that an attribute fi separates rows r and r ∈ U (T, r) if rows r and r have diﬀerent numbers at the intersection with the column fi . For i = 1, . . . , m denote by U (T, r, fi ) the set of rows from U (T, r) which attribute fi separates from the row r. Let α be a real number such that 0 ≤ α < 1. A decision rule fi1 = bi1 ∧ . . . ∧ fit = bit → d

(3)

. , fit separate from r is called an α-decision rule for T and r if attributes fi1 , . . at least (1 − α)|U (T, r)| rows from U (T, r). The number tj=1 w(fij ) is called the weight of the considered decision rule. If U (T, r) = ∅ then for any fi1 , . . . , fit ∈ {f1 , . . . , fm } the rule (3) is an αdecision rule for T and r. Also, the rule (3) with empty left-hand side (when t = 0) is an α-decision rule for T and r. The weight of this rule is equal to 0. For example, 0.01-decision rule means that we must separate from r at least 99% of rows from U (T, r). Note that 0-rule is usual (exact) rule. Let us denote by Lmin (α) = Lmin (α, T, r, w) the minimal weight of α-decision rule for T and r. Let α, γ be real numbers such that 0 ≤ γ ≤ α < 0. Let us describe a greedy algorithm with thresholds α and γ which constructs an α-decision rule for given T , r and weight function w. Let r = (b1 , . . . , bm ), and r be labeled by the decision d. The right-hand side of constructed α-decision rule is equal to d. If U (T, r) = ∅ then the left-hand side of constructed α-decision rule is empty. Let U (T, r) = ∅. Let us denote M = |U (T, r)|(1 − α) and N = |U (T, r)|(1 − γ) . Let we make i ≥ 0 steps and construct a decision rule R with i conditions (if i = 0 then the left-hand side of R is empty). Let us describe the step number i + 1. Let us denote by D the set of rows from U (T, r) separated from r by attributes belonging to R (if i = 0 then D = ∅). If |D| ≥ M then we ﬁnish the work of the algorithm, and R is the constructed α-decision rule. Let |D| < M . Then we choose an attribute fj with minimal number j for which U (T, r, fj ) \ D = ∅ and the value w(fj ) min{|U (T, r, fj ) \ D|, N − |D|} is minimal. Add the condition fj = bj to R. Pass to the step number i + 2. Let us denote by Lγgreedy (α) = Lγgreedy (α, T, r, w) the weight of α-decision rule constructed by the considered algorithm for given table T , row r and weight function w.

On Partial Covers, Reducts and Decision Rules with Weights

4.2

237

On Precision of Greedy Algorithm with Thresholds α and α

The following two statements are simple corollaries of results of Slav´ık (see Theorems 4 and 5). Theorem 23. Let T be a decision table, r be a row of T , U (T, r) = ∅, w be a weight function for T , and α be a real number such that 0 ≤ α < 1. Then Lα greedy (α) ≤ Lmin (α)H ((1 − α)|U (T, r)| ). Theorem 24. Let T be a decision table with m columns labeled by attributes f1 , . . . , fm , r be a row of T , U (T, r) = ∅, w be

a weight function for T , α ∈ IR and 0 ≤ α < 1. Then Lα greedy (α) ≤ Lmin (α)H maxi∈{1,...,m} |U (T, r, fi )| . 4.3

On Polynomial Approximate Algorithms

In this subsection we consider three theorems which follow immediately from Theorems 39–41 [9]. Let 0 ≤ α < 1. Let us consider the following problem: for given decision table T , row r of T and weight function w for T it is required to ﬁnd an α-decision rule for T and r with minimal weight. Theorem 25. Let 0 ≤ α < 1. Then the problem of construction of α-decision rule with minimal weight is N P -hard. So we must consider polynomial approximate algorithms for minimization of α-decision rule weight. Theorem 26. Let α ∈ IR and 0 ≤ α < 1. If N P ⊆ DT IM E(nO(log log n) ) then for any ε, 0 < ε < 1, there is no polynomial algorithm that for given decision table T , row r of T with U (T, r) = ∅ and weight function w for T constructs α-decision rule for T and r which weight is at most (1−ε)Lmin(α, T, r, w) ln |U (T, r)|. Theorem 27. Let α be a real number such that 0 ≤ α < 1. If P = N P then there exists δ > 0 such that there is no polynomial algorithm that for given decision table T , row r of T with U (T, r) = ∅ and weight function w for T constructs α-decision rule for T and r which weight is at most δLmin (α, T, r, w) ln |U (T, r)|. From Theorem 23 it follows that Lα greedy (α) ≤ Lmin (α)(1 + ln |U (T, r)|). From this inequality and from Theorem 26 it follows that under the assumption N P ⊆ DT IM E(nO(log log n) ) greedy algorithm with two thresholds α and α is close to best polynomial approximate algorithms for minimization of partial decision rule weight. From the considered inequality and from Theorem 27 it follows that under the assumption P = N P greedy algorithm with two thresholds α and α is not far from best polynomial approximate algorithms for minimization of partial decision rule weight.

238

M.Ju. Moshkov, M. Piliszczuk, and B. Zielosko

However we can try to improve the results of the work of greedy algorithm with two thresholds α and α for some part of decision tables. 4.4

Two Modiﬁcations of Greedy Algorithm

First, we consider binary diagnostic decision tables and prove that under some assumptions on the number of attributes and rows for the most part of tables for each row there exist weight function w and numbers α, γ such that the weight of α-decision rule constructed by greedy algorithm with thresholds α and γ is less than the weight of α-decision rule constructed by greedy algorithm with thresholds α and α. Binary means that the table is ﬁlled by numbers from the set {0, 1} (all attributes have values from {0, 1}). Diagnostic means that rows of the table are labeled by pairwise diﬀerent numbers (decisions). Let T be a binary diagnostic decision table with m columns labeled by attributes f1 , . . . , fm and with n rows. We will assume that rows of T with numbers 1, . . . , n are labeled by decisions 1, . . . , n respectively. Therefore the number of considered tables is equal to 2mn . Decision table will be called simple if it has no equal rows. Theorem 28. Let us consider binary diagnostic decision tables with m columns labeled by attributes f1 , . . . , fm and n ≥ 5 rows labeled by decisions 1, . . . , n. The fraction of decision tables T for each of which for each row r of T there exist a weight function w and numbers α, γ such that 0 ≤ γ < α < 1 and Lγgreedy (α, T, r, w) < Lα greedy (α, T, r, w) is at least 1−

n2 n3m − . m 2m (n − 1) 2 −1

Proof. Let T be a decision table and r be a row of T with number s ∈ {1, . . . , n}. We will say that a decision table T is 1-uniform relatively r if there exists natural p such that for any attribute fi of T if |U (T, r, fi )| > 0 then |U (T, r, fi )| ∈ {p, p+1}. Using reasoning similar to the proof of Theorem 9 one can show that if T is not 1-uniform relatively r then there exist a weight function w and numbers α, γ such that 0 ≤ γ < α < 1 and Lγgreedy (α, T, r, w) < Lα greedy (α, T, r, w). We evaluate the number of decision tables which are not 1-uniform relatively each row. Let (δ1 , . . . , δm ) ∈ {0, 1}m. First, we evaluate the number of simple decision tables for which r = (δ1 , . . . , δm ) and which are 1-uniform relatively r. Let us consider such a decision table T . It is clear that there exists p ∈ {1, . . . , n − 2} such that for i = 1, . . . , m the column fi contains exactly 0 or p or p + 1 numbers ¬δi . Thereforethe number of considered decision tables m n−2 p p+1 is at most p=1 Cn−1 + Cn−1 + 1 . Using Lemma 2 we conclude that this m n−1 m mn−m m (n−1)/2 3 √ < (n − 1) 3·2 = 2 number is at most (n − 2) 3Cn−1 m −1 . n−1 2 (n−1)

There are 2m variants for the choice of the tuple (δ1 , . . . , δm ) and n variants for the choice of the number s of row r. Therefore the number of simple decision mn−m m 3 tables which are 1-uniform relatively at least one row is at most n2m 2 m −1 = 2 (n−1)

On Partial Covers, Reducts and Decision Rules with Weights n2mn 3m . m (n−1) 2 −1

239

The number of tables which are not simple is at most n2 2mn−m .

Hence the number of tables which are not 1-uniform for each row is at least mn m 2mn − n2 m23 −1 − n2 2mn−m . Thus, the fraction, considered in the statement of (n−1)

the theorem, is at least 1 −

n3m m (n−1) 2 −1

−

n2 2m .

m

So if m ≥ 6 and n, 2n2 are large enough then for the most part of binary diagnostic decision tables for each row there exist weight function w and numbers α, γ such that the weight of α-decision rule constructed by greedy algorithm with thresholds α and γ is less than the weight of α-decision rule constructed by greedy algorithm with thresholds α and α. The obtained results show that the greedy algorithm with two thresholds α and γ is of some interest. Now we consider two polynomial modiﬁcations of greedy algorithm which allow to use advantages of greedy algorithm with two thresholds α and γ. Let T be a decision table with m columns labeled by attributes f1 , . . . , fm , r = (b1 , . . . , bm ) be a row of T labeled by decision d, U (T, r) = ∅, w be a weight function for T and α be a real number such that 0 ≤ α < 1. 1. It is impossible to consider eﬀectively all γ such that 0 ≤ γ ≤ α. Instead of this we can consider all natural N such that M ≤ N ≤ |U (T, r)| where M = |U (T, r)|(1 − α) (see the description of greedy algorithm with two thresholds). For each N ∈ {M, . . . , |U (T, r)|} we apply greedy algorithm with parameters M and N to T , r and w and after that choose an α-decision rule with minimal weight among constructed α-decision rules. 2. There exists also an another way to construct an α-decision rule which is not worse than the one obtained under consideration of all N such that M ≤ N ≤ |U (T, r)|. Let us apply greedy algorithm with thresholds α and α to T , r and w. Let the algorithm choose sequentially attributes fj1 , . . . , fjt . For each i ∈ {0, . . . , t − 1} we ﬁnd (if it is possible) an attribute fli of T with minimal weight w(fli ) such that the rule fj1 = bj1 ∧ . . . ∧ fji = bji ∧ fli = bli → d is an α-decision rule for T and r (if i = 0 then it will be the rule fl0 = bl0 → d). After that among constructed α-decision rules fj1 = bj1 ∧ . . . ∧ fjt = bjt → d, ..., fj1 = bj1 ∧ . . . ∧ fji = bji ∧ fli = bli → d, ... we choose an α-decision rule with minimal weight. From Proposition 7 it follows that the constructed α-decision rule is not worse than the one constructed under consideration of all γ, 0 ≤ γ ≤ α, or (which is the same) all N , M ≤ N ≤ |U (T, r)|. Using Propositions 1 one can prove the following statement. Proposition 7. Let T be a decision table, r be a row of T , U (T, r) = ∅, w be a weight function for T and α, γ be real numbers such that 0 ≤ γ < α < 1. Let the greedy algorithm with two thresholds α and α, which is applied to T , r and w, choose sequentially attributes fg1 , . . . , fgt . Let the greedy algorithm with two thresholds α and γ, which is applied to T , r and w, choose sequentially attributes fl1 , . . . , flk . Then either k = t and (l1 , . . . , lk ) = (g1 , . . . , gt ) or k ≤ t, (l1 , . . . , lk−1 ) = (g1 , . . . , gk−1 ) and lk = gk .

240

4.5

M.Ju. Moshkov, M. Piliszczuk, and B. Zielosko

Bounds on Lmin (α) and Lγgreedy (α)

First, we ﬁx some information about the work of greedy algorithm with two thresholds and ﬁnd the best lower bound on the value Lmin (α) depending on this information. Let T be a decision table, r be a row of T such that U (T, r) = ∅, w be a weight function for T , and α, γ be real numbers such that 0 ≤ γ ≤ α < 1. Let us apply the greedy algorithm with thresholds α and γ to the decision table T , row r and the weight function w. Let during the construction of α-decision rule the greedy algorithm choose sequentially attributes fg1 , . . . , fgt . Let us denote U (T, r, fg0 ) = ∅ and δ0 = 0. For i = 1, . . . , t denote δi = |U (T, r, fgi ) \ (U (T, r, fg0 ) ∪ . . . ∪ U (T, r, fgi−1 ))| and wi = w(fgi ). As information on the greedy algorithm work we will use numbers ML = ML (α, γ, T, r, w) = |U (T, r)|(1 − α) , NL = NL (α, γ, T, r, w) = |U (T, r)|(1 − γ) and tuples ΔL = ΔL (α, γ, T, r, w) = (δ1 , . . . , δt ), WL = WL (α, γ, T, r, w) = (w1 , . . . , wt ). For i = 0, . . . , t − 1 denote wi+1 (ML − (δ0 + . . . + δi )) ρi = . min{δi+1 , NL − (δ0 + . . . + δi )} Let us deﬁne parameter ρL (α, γ) = ρL (α, γ, T, r, w) as follows: ρL (α, γ) = max {ρi : i = 0, . . . , t − 1} . We will show that ρL (α, γ) is the best lower bound on Lmin (α) depending on ML , NL , ΔL and WL . Using Theorem 11 one can prove the following statement. Theorem 29. For any decision table T , any row r of T with U (T, r) = ∅, any weight function w for T , and any real numbers α, γ, 0 ≤ γ ≤ α < 1, the inequality Lmin (α, T, r, w) ≥ ρL (α, γ, T, r, w) holds, and there exist a decision table T , a row r of T and a weight function w for T such that ML (α, γ, T , r , w ) = ML (α, γ, T, r, w), NL (α, γ, T , r , w ) = NL (α, γ, T, r, w) , ΔL (α, γ, T , r , w ) = ΔL (α, γ, T, r, w), WL (α, γ, T , r , w ) = WL (α, γ, T, r, w) , ρL (α, γ, T , r , w ) = ρL (α, γ, T, r, w), Lmin (α, T , r , w ) = ρL (α, γ, T , r , w ) . Let us consider a property of the parameter ρL (α, γ) which is important for practical use of the bound from Theorem 29. Using Proposition 2 one can prove the following statement. Proposition 8. Let T be a decision table, r be a row of T with U (T, r) = ∅, w be a weight function for T , and α, γ be real numbers such that 0 ≤ γ ≤ α < 1. Then ρL (α, α, T, r, w) ≥ ρL (α, γ, T, r, w). Now we study some properties of parameter ρL (α, γ) and obtain two upper bounds on the value Lγgreedy (α) which do not depend directly on cardinality of the set U (T, r) and cardinalities of subsets U (T, r, fi ). Using Theorem 12 one can prove the following statement.

On Partial Covers, Reducts and Decision Rules with Weights

241

Theorem 30. Let T be a decision table, r be a row of T with U (T, r) = ∅, w be a weight function for T , α, γ ∈ IR and 0 ≤ γ < α < 1. Then Lγgreedy (α, T, r, w) < 1−γ +1 . ρL (γ, γ, T, r, w) ln α−γ Corollary 9. Let ε ∈ IR and 0 < ε < 1. Then for any α, ε ≤ α < 1, the 1 inequalities ρL (α, α) ≤ Lmin (α) ≤ Lα−ε greedy (α) < ρL (α − ε, α − ε) ln ε + 1 hold.

1

1 + 1 < 5.61 and ln 0.1 + 1 < 3.31. The obtained results For example, ln 0.01 show that the lower bound Lmin (α) ≥ ρL (α, α) is nontrivial. Using Theorem 13 one can prove the following statement. Theorem 31. Let T be a decision table, r be a row of T with U (T, r) = ∅, w be a weight function for T , α, γ ∈ IR and 0 ≤ γ < α < 1. Then Lγgreedy (α, T, r, w) < 1−γ +1 . Lmin (γ, T, r, w) ln α−γ Corollary 10. L0greedy (0.001) < 7.91Lmin(0), L0.001 greedy (0.01) < 5.71Lmin (0.001), 0.3 (0.2) < 3.20L (0.1), L (0.5) < 2.26L L0.1 min min(0.3). greedy greedy

Corollary 11. Let 0 < α < 1. Then L0greedy (α) < Lmin(0) ln α1 + 1 . Corollary 12. Let ε be a real number, and 0 < ε < 1. Then for any α such that 1 (α) < L (α − ε) ln + 1 hold. ε ≤ α < 1 the inequalities Lmin (α) ≤ Lα−ε min greedy ε 4.6

Results of Experiments for α-Decision Rules

In this subsection we will consider only binary decision tables T with binary decision attributes. First Group of Experiments. First group of experiments is connected with study of quality of greedy algorithm with one threshold (where γ = α or, which is the same, N = M ), and comparison of quality of greedy algorithm with one threshold and ﬁrst modiﬁcation of greedy algorithm (where for each N ∈ {M, . . . , |U (T, r)|} we apply greedy algorithm with parameters M and N to decision table, row and weight function and after that choose an α-decision rule with minimal weight among constructed α-decision rules). We generate randomly 1000 decision tables T , rows r and weight functions w such that T contains 40 rows and 10 conditional attributes f1 , . . . , f10 , r is the ﬁrst row of T , and 1 ≤ w(fi ) ≤ 1000 for i = 1, . . . , 10. For each α ∈ {0.1, . . . , 0.9} we ﬁnd the number of triples (T, r, w) for which greedy algorithm with one threshold constructs an α-decision rule with minimal weight (an optimal α-decision rule), i.e. Lα greedy (α, T, r, w) = Lmin (α, T, r, w). This number is contained in the row of Table 3 labeled by ”Opt”. We ﬁnd the number of triples (T, r, w) for which ﬁrst modiﬁcation of greedy algorithm constructs an α-decision rule which weight is less than the weight of α-decision rule constructed by greedy algorithm with one threshold, i.e. there

242

M.Ju. Moshkov, M. Piliszczuk, and B. Zielosko

exists γ such that 0 ≤ γ < α and Lγgreedy (α, T, r, w) < Lα greedy (α, T, r, w). This number is contained in the row of Table 3 labeled by ”Impr”. Also we ﬁnd the number of triples (T, r, w) for which ﬁrst modiﬁcation of greedy algorithm constructs an optimal α-decision rule which weight is less than the weight of α-decision rule constructed by greedy algorithm with one threshold, i.e. there exists γ such that 0 ≤ γ < α and Lγgreedy (α, T, r, w) = Lmin(α, T, r, w) < Lα greedy (α, T, r, w). This number is contained in the row of Table 3 labeled by ”Opt+”. Table 3. Results of ﬁrst group of experiments with α-decision rules α 0.0 Opt 434 Impr 0 Opt+ 0

0.1 559 31 16

0.2 672 51 35

0.3 800 36 28

0.4 751 22 17

0.5 733 27 26

0.6 866 30 25

0.7 966 17 13

0.8 0.9 998 1000 1 0 1 0

The obtained results show that the percentage of triples for which greedy algorithm with one threshold ﬁnds an optimal α-decision rule grows almost monotonically (with local minimum near to 0.4–0.5) from 43.4% up to 100%. The percentage of problems for which ﬁrst modiﬁcation of greedy algorithm can improve the result of the work of greedy algorithm with one threshold is less than 6%. However, sometimes (for example, if α = 0.3, α = 0.6 or α = 0.7) the considered improvement is noticeable. Second Group of Experiments. Second group of experiments is connected with comparison of quality of greedy algorithm with one threshold and ﬁrst modiﬁcation of greedy algorithm. We make 25 experiments (row ”Nr” in Table 4 contains the number of experiment). Each experiment includes the work with three randomly generated families of triples (T, r, w) (1000 triples in each family) such that T contains n rows and m conditional attributes, r is the ﬁrst row of T , and w has values from the set {1, . . . , v}. If the column ”n” contains one number, for example ”40”, it means that n = 40. If this row contains two numbers, for example ”30–120”, it means that for each of 1000 triples we choose the number n randomly from the set {30, . . . , 120}. The same situation is for the column ”m”. If the column ”α” contains one number, for example ”0.1”, it means that α = 0.1. If this column contains two numbers, for example ”0.2–0.4”, it means that we choose randomly the value of α such that 0.2 ≤ α ≤ 0.4. For each of the considered triples (T, r, w) and number α we apply greedy algorithm with one threshold and ﬁrst modiﬁcation of greedy algorithm. Column ”#i”, i = 1, 2, 3, contains the number of triples (T, r, w) from the family number i for each of which the weight of α-decision rule, constructed by ﬁrst modiﬁcation of greedy algorithm, is less than the weight of α-decision rule constructed by

On Partial Covers, Reducts and Decision Rules with Weights

243

Table 4. Results of second group of experiments with α-decision rules Nr 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25

n 1–100 1–100 1–100 1–100 1–100 1–100 1–100 1–100 100 100 100 100 1–30 30–60 60–90 90–120 40 40 40 40 40 40 40 40 40

m 1–100 1–100 1–100 1–100 1–100 1–100 1–100 1–100 1–30 30–60 60–90 90–120 30 30 30 30 10 10 10 10 10 10 10 10 10

v 1–10 1–100 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000

α 0–1 0–1 0–1 0–0.2 0.2–0.4 0.4–0.6 0.6–0.8 0.8–1 0–0.2 0–0.2 0–0.2 0–0.2 0–0.2 0–0.2 0–0.2 0–0.2 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9

#1 4 7 19 20 28 22 7 0 35 47 45 37 11 20 30 40 31 37 35 27 32 28 10 0 0

#2 2 14 13 39 29 23 6 1 38 43 51 40 8 22 33 48 39 39 30 20 32 26 12 2 0

#3 4 13 15 22 28 34 4 0 28 31 36 55 9 35 34 38 34 47 37 27 36 24 10 0 0

avg 3.33 11.33 15.67 27.00 28.33 26.33 5.67 0.33 33.67 40.33 44.00 44.00 9.33 25.67 32.33 42.00 34.67 41.00 34.00 24.67 33.33 26.00 10.67 0.67 0

greedy algorithm with one threshold. In other words, in column ”#i” we have the number of triples (T, r, w) from the family number i such that there exists γ for which 0 ≤ γ < α and Lγgreedy (α, T, r, w) < Lα greedy (α, T, r, w). The column #1+#2+#3 . ”avg” contains the number 3 In experiments 1–3 we consider the case where the parameter v increases. In experiments 4–8 the parameter α increases. In experiments 9–12 the parameter m increases. In experiments 13–16 the parameter n increases. In experiments 17–25 the parameter α increases. The results of experiments show that the value of #i can change from 0 to 55. It means that the percentage of triples for which the ﬁrst modiﬁcation of greedy algorithm is better than the greedy algorithm with one threshold can change from 0% to 5.5%. Third Group of Experiments. Third group of experiments is connected with investigation of quality of lower bound Lmin (α) ≥ ρL (α, α). We choose natural n, m, v and real α, 0 ≤ α < 1. For each chosen tuple (n, m, v, α) we generate randomly 30 triples (T, r, w) such that T contains

244

M.Ju. Moshkov, M. Piliszczuk, and B. Zielosko

1000

2500 average upper bound average lower bound

average upper bound average lower bound

900 2000 800

700

1500

600 1000

500

400 500 300

200 1000

0 1500

2000

2500

3000 number of rows

3500

4000

4500

5000

10

20

30

40

50 60 number of attributes

70

80

90

100

Fig. 3. Results of third group of experiments with rules (n and m are changing)

900

1600 average upper bound average lower bound

average upper bound average lower bound

800

1400

700

1200

600 1000 500 800 400 600 300 400

200

200

100

0 100

0 200

300

400

500 600 range of weight

700

800

900

1000

0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

alpha

Fig. 4. Results of third group of experiments with rules (v and α are changing)

n rows and m conditional attributes, r is the ﬁrst row of T , and w has values from the set {1, ..., v}. After that we ﬁnd values of Lα greedy (α, T, r, w) and ρL (α, α, T, r, w) for each of generated 30 triples. Note that ρL (α, α, T, r, w) ≤ Lmin (α, T, r, w) ≤ Lα greedy (α, T, r, w). Finally, for generated 30 triples we ﬁnd mean values of Lα greedy (α, T, r, w) and ρL (α, α, T, r, w). Results of experiments can be found in Figs. 3 and 4. In these ﬁgures mean values of ρL (α, α, T, r, w) are called ”average lower bound” and mean values of Lα greedy (α, T, r, w) are called ”average upper bound”. In Fig. 3 (left-hand side) one can see the case when n ∈ {1000, 2000, . . . , 5000}, m = 30, v = 1000 and α = 0.01. In Fig. 3 (right-hand side) one can see the case when n = 1000, m ∈ {10, 20, . . . , 100}, v = 1000 and α = 0.01. In Fig. 4 (left-hand side) one can see the case when n = 1000, m = 30, v ∈ {100, 200, . . . , 1000} and α = 0.01. In Fig. 4 (right-hand side) one can see the case when n = 1000, m = 30, v = 1000 and α ∈ {0.0, 0.1, . . . , 0.9}. Results of experiments show that the considered lower bound is nontrivial and can be useful in investigations.

On Partial Covers, Reducts and Decision Rules with Weights

5

245

Conclusions

The paper is devoted (mainly) to theoretical and experimental analysis of greedy algorithms with weights and their modiﬁcations for partial cover, reduct and decision rule construction. Obtained results will further to more wide use of such algorithms in rough set theory and its applications. In the further investigations we are planning to generalize the obtained results to the case of decision tables which can contain missing values, continuous attributes, and discrete attributes with large number of values.

Acknowledgments The authors are greatly indebted to Andrzej Skowron for stimulating discussions and to anonymous reviewers for helpful remarks and suggestions.

References 1. Cheriyan, J., Ravi, R.: Lecture Notes on Approximation Algorithms for Network Problems (1998) http://www.math.uwaterloo.ca/˜jcheriya/lecnotes.html 2. Chv´ atal, V.: A greedy heuristic for the set-covering problem. Mathematics of Operations Research 4 (1979) 233–235. 3. Feige, U.: A threshold of ln n for approximating set cover (Preliminary version). Proceedings of 28th Annual ACM Symposium on the Theory of Computing (1996) 314–318. 4. Gavrilov G.P., Sapozhenko, A.A.: Problems and Exercises in Discrete Mathematics (third edition). Fizmatlit, Moscow, 2004 (in Russian). 5. Kearns, M.J.: The Computational Complexity of Machine Learning. MIT Press, Cambridge, Massachussetts, 1990. 6. Moshkov, M.Ju.: Greedy algorithm for set cover in context of knowledge discovery problems. In: Proceedings of the International Workshop on Rough Sets in Knowledge Discovery and Soft Computing (ETAPS 2003 Satellite Event). Warsaw, Poland. Electronic Notes in Theoretical Computer Science 82 (2003). 7. Moshkov, M.Ju.: On greedy algorithm for partial cover construction. In: Proceedings of the Fourteenth International Workshop Design and Complexity of Control Systems. Nizhny Novgorod, Russia (2003) 57 (in Russian). 8. Moshkov, M.Ju., Piliszczuk, M., Zielosko, B.: Greedy algorithm for construction of partial covers. In: Proceedings of the Fourteenth International Conference Problems of Theoretical Cybernetics. Penza, Russia (2005) 103 (in Russian). 9. Moshkov, M.Ju., Piliszczuk, M., Zielosko, B.: On partial covers, reducts and decision rules. LNCS Transactions on Rough Sets, Springer-Verlag (submitted). ´ ezak, D.: Approximate reducts and association rules - correspon10. Nguyen, H.S., Sl¸ dence and complexity results. In: Proceedings of the Seventh International Workshop on Rough Sets, Fuzzy Sets, Data Mining, and Granular-Soft Computing. Yamaguchi, Japan. Lecture Notes in Artiﬁcial Intelligence 1711, Springer-Verlag (1999) 137–145. 11. Pawlak, Z.: Rough Sets – Theoretical Aspects of Reasoning about Data. Kluwer Academic Publishers, Dordrecht, Boston, London, 1991.

246

M.Ju. Moshkov, M. Piliszczuk, and B. Zielosko

12. Pawlak, Z.: Rough set elements. In: Polkowski, L., Skowron, A. (Eds.), Rough Sets in Knowledge Discovery 1. Methodology and Applications (Studies in Fuzziness and Soft Computing 18). Phisica-Verlag. A Springer-Verlag Company (1998) 10–30. 13. Piliszczuk, M.: On greedy algorithm for partial reduct construction. In: Proceedings of Concurrency, Speciﬁcation and Programming Workshop 2. Ruciane-Nida, Poland (2005) 400–411. 14. Quafafou, M.: α-RST: a generalization of rough set theory. Information Sciences 124 (2000) 301–316. 15. Raz, R., Safra, S.: A sub-constant error-probability low-degree test, and subconstant error-probability PCP characterization of NP. In: Proceedings of 29th Annual ACM Symposium on the Theory of Computing (1997) 475–484. 16. Skowron, A.: Rough sets in KDD. In: Proceedings of the 16-th World Computer Congress (IFIP’2000). Beijing, China (2000) 1–14. 17. Skowron, A., Rauszer, C.: The discernibility matrices and functions in information systems. In: Slowinski, R. (Ed.), Intelligent Decision Support. Handbook of Applications and Advances of the Rough Set Theory. Kluwer Academic Publishers, Dordrecht, Boston, London (1992) 331–362. 18. Slav´ık, P.: Approximation algorithms for set cover and related problems. Ph.D. thesis. University of New York at Buﬀalo (1998). ´ ezak, D.: Approximate reducts in decision tables. In: Proceedings of the Congress 19. Sl¸ Information Processing and Management of Uncertainty in Knowledge-based Systems 3. Granada, Spain (1996) 1159–1164. ´ ezak, D.: Normalized decision functions and measures for inconsistent decision 20. Sl¸ tables analysis. Fundamenta Informaticae 44 (2000) 291–319. ´ ezak, D.: Approximate decision reducts. Ph.D. thesis. Warsaw University (2001) 21. Sl¸ (in Polish). ´ ezak, D.: Approximate entropy reducts. Fundamenta Informaticae 53 (2002) 22. Sl¸ 365–390. ´ ezak, D., Wr´ 23. Sl¸ oblewski, J.: Order-based genetic algorithms for the search of approximate entropy reducts. In: Proceedings of the International Conference Rough Sets, Fuzzy Sets, Data Mining, and Granular Computing. Chongqing, China. Lecture Notes in Artiﬁcial Intelligence 2639, Springer-Verlag (2003) 308–311. 24. Wr´ oblewski, J.: Ensembles of classiﬁers based on approximate reducts. Fundamenta Informaticae 47 (2001) 351–360. 25. Yablonskii, S.V.: Introduction into Discrete Mathematics (fourth edition). Vishaya Shkola, Moscow, 2003 (in Russian). 26. Ziarko, W.: Analysis of uncertain information in the framework of variable precision rough sets. Foundations of Computing and Decision Sciences 18 (1993) 381–396. 27. Zielosko, B.: On partial decision rules. In: Proceedings of Concurrency, Speciﬁcation and Programming Workshop 2. Ruciane-Nida, Poland (2005) 598–609.

A Personal View on AI, Rough Set Theory and Professor Pawlak Toshinori Munakata Computer and Information Science Department, Cleveland State University Cleveland, Ohio 44115, U.S.A. t.munakata@csuohio.edu

It is an honor to contribute my short article to this special issue commemorating the life and work of Professor Zdzislaw Pawlak. In this article I would like to discuss my encounters with the ﬁeld of artiﬁcial intelligence (AI) in general, and how I see rough set theory and Professor Zdzislaw Pawlak in this context. I have been fortunate to know some of the greatest scholars in the AI ﬁeld. There are many of them, but if I had to choose the three I admire most, they are: Professors Zdzislaw Pawlak, Lotﬁ Zadeh and Herbert A. Simon. There are common characteristics among all of them. Although they are the most prominent of scholars, all are frank and easy and pleasant to talk with. All are professionally active at ages where ordinary people would have long since retired. I became interested in the ﬁeld of AI in the mid 70s. I have observed many ups and downs of the ﬁeld in terms of the popularity since then - a common phenomena in any ﬁeld. The AAAI (American Association for Artiﬁcial Intelligence) was inaugurated and the ﬁrst issue of the AI Magazine was published in the spring 1980. The timing of the birth of rough set theory was soon after this event. At this time many people in the world were becoming interested in the ﬁeld of AI, while there were only a handful researchers when the ﬁeld started in the 1950s. In the spring of 1986, the ﬁrst issue of the IEEE Expert (now IEEE Intelligent Systems) was inaugurated. I served as an Associate Editor of this magazine for two terms from 1987 to 1991. In terms of public popularity AI was ﬂourishing in this eras. During many years of the 70s and 80s, I observed that despite media hype and claims for break-thorough technologies, most AI techniques were not practical. Here “practical” means “having real-world commercial and industrial applications on an everyday basis.” For example, I could not ﬁnd cases where machine learning techniques discussed in textbooks such as “learning from examples” and “learning from analogy” were actually employed at industrial plants or commercial banks. The same were true for other AI techniques such as blackboard and neural networks. After Minsky’s pessimistic view on the ﬁeld, the U.S. government funding ceased, and only a handful researchers remained active in the ﬁeld. The ﬁeld revived in the mid to late 80s, and became quite popular. However, I could not ﬁnd a single case where neural networks were actually used every day for commercial and industrial applications. For all of these observations I could be wrong because there could have been exceptions I was not aware of, but I was certain that these exceptions were few, if any. J.F. Peters et al. (Eds.): Transactions on Rough Sets VI, LNCS 4374, pp. 247–252, 2007. c Springer-Verlag Berlin Heidelberg 2007

248

T. Munakata

This situation of impracticality of most AI techniques appeared to start to change around 1990. That is, many AI techniques were becoming truly practical and their application domains much broader. Of course, there were practical AI techniques before 1990. Robotics was one. The ﬁrst major industrial fuzzy control was implemented in Denmark in 1982, followed by the famous Hitachi’s Sendai subway control in 1986. There had been practical expert systems in the 80s. However, the repertories of AI techniques and their application domains were becoming much more extensive around 1990. With this background, in March 1993 I made a proposal to the Communications of the ACM (CACM ) for a Special Issue entitled “Commercial and Industrial AI.” The CACM was a primary computer science magazine subscribed by some 85,000 professionals worldwide at that time. Its readers went far beyond the computer science discipline including ﬁelds such as engineering, social science and education. The proposal was accepted and I selected the most practical or promising AI areas with the help of many experts. The Special Issue was published in March 1994 [3] and was divided into four major sections with 11 articles. They are: I. “Knowledge Engineering Systems” with two articles – an article on general expert systems and an article on case-based reasoning. II. “Perception, Understanding, and Action” with three articles on vision, speech and robotics. III. “Fuzzy Systems” with two articles – an overview and a soft computing article by Professor Zadeh. IV. “Models of the Brain and Evolution” with four articles - two articles by Rumelhart, Widrow, et al., an article on neural networks in Japan, and an article on genetic algorithms. There were many behind-the-scene stories and one of them was that my original plan was to have only one article by Rumelhart, et al. After much delay, they had an article twice as long as originally planned, and I suggested splitting the manuscript into two parts. In the Guest Editors Introduction, I wrote: The practical application of artiﬁcial intelligence (AI) has been the center of controversy for many years. Certainly, if we mean AI to be a realization of real human intelligence in the machine, its current state may be considered primitive. In this sense, the name artiﬁcial ”intelligence” can be misleading. However, when AI is looked at as ”advanced computing,” it can be seen as much more. In the past few years, the repertory of AI techniques has evolved and expanded, and applications have been made in everyday commercial and industrial domains. AI applications today span the realms of manufacturing, consumer products, ﬁnance, management, and medicine. Implementation of the correct AI technique in an application is often a must to stay competitive. Truly proﬁtable AI techniques are even kept secret. Many of the statements I wrote here are still basically true today. This Special Issue turned out to be a big hit. The ACM printed 1,000 extra copies of this issue for back orders, but they sold out less than a month. A person from a Japanese company wanted to purchase a box of ﬁfty copies of this issue, but it was too late. The issue became one of the most cited CACM, for not only within

A Personal View on AI, Rough Set Theory and Professor Pawlak

249

computer science but also some unexpected places such as Scientiﬁc American, the Berkeley Law School, the Stanford Philosophy Department, etc. The US Air Force has issued interesting predictions in the past. Around 1947, the Air Force predicted likely technologies for the next 50 years. They included jet rather than propeller powered, and supersonic airplanes. They became the reality by 1997. In 1997, they issued predictions for the next 50 years, i.e., by 2047. Between these major predictions, they published future perspectives for shorter time ranges. My Special Issue was cited in a document within a report entitled “Air Force 2025” that describes 30-year predictions by the US Air Force [1]. When I was preparing my ﬁrst CACM Special Issue, I knew there were other AI areas that were not covered. As soon as the ﬁrst issue was nearly complete, I started working on a follow-up Special Issue entitled “New Horizons in Commercial and Industrial AI.” In the “Editorial Pointers” in the ﬁrst issue, Executive Editor Diane Crawford wrote: “He has assembled some of the foremost minds in AI to author and/or review the 11 articles presented here. If that weren’t enough, he’s already digging into another proposed issue for Communications to appear early next year, where he hopes to address new horizons and applications in other AI-related ﬁelds.” For the ﬁrst Special Issue I received many responses. One of them was a letter from Professor Herbert A. Simon of Carnegie-Mellon University, a prominent scholar in AI with a Turing Award and a Nobel Prize in economics. Basically, he stated: “The ﬁrst issue was well done, although if I were the Guest Editor I would have had less emphasis on neural networks and included an article on machine learning.” He suggested placing an article on machine learning in the second issue. I replied to him saying I had already planned that and asked him to write one, and subsequently he co-authored an article. I was lucky to be able to have close contact with Professor Simon. When IBM’s Deep Blue defeated the human chess champion Garry Kasparov in 1997, he and I co-authored a commentary article on the signiﬁcance of this event on AI [14]. He was a pleasant person to talk with. He was a fulltime professor and active until two weeks before his death in 2001 at age 84. For the second Special Issue I had planned to include certain topics from a very early stage. They included symbolic machine learning, natural language processing (e.g., machine translation) and logic programming. Also, I wanted to include articles addressing the commonsense problem, although I did not expect that this area would have many commercial or industrial applications in the near term. At a later stage of preparation of the second issue, I searched for additional areas appropriate for the issue, and found rough set theory. I was not familiar with this area, but from what I found I thought it was a promising technique, appropriate for the second issue. Perhaps it could complement other AI techniques. I contacted Professor Pawlak and asked him whether he was interested in contributing an article to such a CACM Special Issue. These were my ﬁrst encounters with rough set theory and Professor Pawlak. He kindly accepted my invitation and contributed an article co-authored with Professors Jerzy GrzymalaBusse, Roman Slowinski and Wojciech Ziarko [8]. This was my

250

T. Munakata

ﬁrst acquaintance with the rough set community. As said earlier, CACM has a large number of audience worldwide and its impact is high. I don’t know how much the appearance of this article has inﬂuenced the promotion of this theory, but I think at least it helped to introduce the term “rough sets” worldwide. Incidentally, when I studied practically successful symbolic machine learning techniques for the ﬁrst time, such as ID3, I was a bit disappointed. From the term “learning,” I expected some elements of human-like learning. For example, given a speciﬁc experience the machine would abstract it, generalize it and be able to use it for similar circumstances in the future. I did not see such human-like learning in ID3. Rather, it simply classiﬁes data based on entropy in information theory. The characteristics of the target data seemed to be too simple. Perhaps the term “learning” was misleading, and probably I expected too much on what we could do from this kind of data. Both ID3 and rough sets can learn from data, but probably ID3 had attracted more attention than rough sets in the scientiﬁc community, at least during the 80s and 90s. Why? One reason might be that ID3 appears to have been in the main stream in the machine learning community, and had received more support from its early introduction. Professor Simon was one of it’s supporters, and he was a great scientist as well as a good salesman to promote his beliefs. For example, he called a software system he developed a “general problem solver,” which implied, with a bit of exaggeration, the system would solve every problem on earth. He was also an optimist. In the late 1950s he predicted that a computer would defeat a human chess champion within 10 years. We waited 10 years, another 10 years, and so forth for the next 40 years. In contrast, Professor Pawlak was a humble and modest scientist and perhaps not particularly a good salesman. In my opinion, rough set theory was not as widely recognized in the AI and CS ﬁelds as it should have been. After my ﬁrst encounter with the rough set community through my CACM second special issue, I have been fortunate to be able to work in this ﬁeld together with these people. I attended several rough set related conferences after my ﬁrst encounter [4, 5, 6]. To promote rough sets, I could think of two among many possibilities. One was to have promotional articles in journals of large audience like the CACM. Another area was to have a rough set application with a high social impact. For the latter, rough control might be a good candidate, I thought. Fuzzy set theory became a hot topic after Hitachi successfully applied fuzzy logic to Sendai subway control. I tried to push rough control, and I was Chair of the rough control interest group. The basic idea of rough control is to employ rough sets to automatically generate input-to-output control rules [7, 9]. The idea was not particularly new, but breakthrough applications would place rough set theory in the spotlight. A long time rough set activist Professor T.Y. Lin ﬁnancially supported me for this endeavor. Although we have not observed a major breakthrough yet, I think possibilities are still there. In the communication with Professor Pawlak, he suggested presenting a co-authored conference paper [13]. When I published an AI book from Springer, I included a chapter for rough sets [10]. When I served as Guest Editor for third time for CACM Special Section on knowledge discovery [11], I asked Professor Ziarko to contribute an article.

A Personal View on AI, Rough Set Theory and Professor Pawlak

251

When a young colleague approached me to work on a data mining article, I suggested employing rough sets [2]. I am currently working on another article on a rough set application with a young assistant professor. Although we all saddened by the recent death of Professor Pawlak, I think he was fortunate to observe that his theory has been widely recognized in the scientiﬁc community worldwide. This was not necessarily the case for many great scholars in the past. During my sabbatical in the fall of 2002, I traveled to Poland, visiting Professors Slowinski, Skowron and Pawlak, and received a warm welcome. This was the last time I saw Professor Pawlak. What are the future prospects of rough sets? No one knows, but the following is my speculation. Despite it’s founder’s death, the community will grow – there will be more researchers worldwide and more theoretical and application developments. But, growth in the ﬁeld may level out eventually, unless we achieve major breakthroughs. As in the case of other machine learning techniques and AI in general, we don’t know what, when or if such breakthroughs may come. Targeting to extremely large volumes of data (e.g., terabytes) and/or massively parallel computing alone do not look very promising, as we have observed similar attempts such as the Cyc and the Connection Machine. For knowledge discovery techniques such as rough sets, there may be a limit when we deal only with decision tables. Perhaps we should also look at other formats of data as well as other types of data, for example, non-text, comprehensive types of information, such as symbolic, visual, audio, etc. Also, the use of huge background knowledge, in a manner similar to human thought, would be necessary and eﬀective. Human-computer interactions would also enhance the discovery processes. Other totally diﬀerent domains are non-silicon based new computing paradigms. I am currently working on my fourth Special Section for the Communications of the ACM as a guest editor on this subject [12]. These approaches may lead to a new dimension of information processing in a wide range of application domains including rough sets. As with other scientiﬁc developments in history, such as alchemy and the ﬁrst airplane, a breakthrough may come in a totally unexpected form.

References 1. Clarence E. Carter, et al. The Man In The Chair: Cornerstone Of Global Battlespace Dominance, Air Force 2025, 1996. 2. Brenda Mak and T. Munakata. ”Rule extraction from expert heuristics: A comparative study of rough sets with neural networks and ID3,” European Journal of Operational Research, 136(1), pp. 212-229, 2002. 3. Toshinori Munakata. ”Guest Editor’s Introduction,” in the Special Issue ”Commercial and Industrial AI Applications,” Communications of the ACM, 37(3), pp. 2325, 1994. 4. Toshinori Munakata. ”Commercial and Industrial AI: Where it is now, and where it will be,” an invited talk at the Third International Workshop on Rough Sets and Soft Computing (RSSC’94), San Jose, CA, Nov., 1012, pp. 5155, 1994.

252

T. Munakata

5. Toshinori Munakata. ”Rough Control: A Perspective,” Workshop on Rough Sets and Data Mining, the 1995 ACM Computer Science Conference, Nashville, TN, March 2, 1995. 6. Toshinori Munakata. ”Commercial and Industrial AI and a Future Perspective on Rough Sets,” in Soft Computing: the Third International Workshop on Rough Sets and Soft Computing (RSSC94), T.Y. Lin and A.M. Wildberger (Eds.), the Society of Computer Simulation, pp. 219222, 1995. 7. Toshinori Munakata. ”Rough Control: Basic Ideas and Applications,” Workshop on Rough Set Theory, the Second Annual Joint Conference on Information Sciences, Wrightsville Beach, NC, Sept. 28 Oct. 1, 1995. 8. Toshinori Munakata. ”Guest Editor’s Introduction,” in the Special Issue ”New Horizons for Commercial and Industrial AI,” Communications of the ACM, 38(11), pp. 28-31, 1995. 9. Toshinori Munakata. ”Rough Control: A Perspective,” in T.Y. Lin and N. Cercone (Eds.), Rough Sets and Data Mining: Analysis of Imprecise Data, Kluwer Academic, pp. 7788, 1997. 10. Toshinori Munakata. Fundamentals of the New Artiﬁcial Intelligence: Beyond Traditional Paradigms, Springer-Verlag, 1998, 2007. 11. Toshinori Munakata. ”Guest Editor’s Introduction,” in the Special Section “Knowledge Discovery,” Communications of the ACM, 42(11), pp. 26-29, 1999. 12. Toshinori Munakata. ”Guest Editor’s Introduction,” in the Special Section “Beyond Silicon: New Computing Paradigms,” Communications of the ACM, to appear in 2007. 13. Zdzislaw Pawlak and Toshinori Munakata. ”Rough Control: Application of Rough Set Theory to Control,” EUFIT ’96 the Fourth European Congress on Intelligent Techniques and Soft Computing, Aachen, Germany, Sept. 25, 1996, pp. 209-217. 14. Herbert A. Simon and Toshinori Munakata. ”Kasparov vs. Deep Blue: The Aftermath - AI Lessons,” Communications of the ACM, 40(8), pp. 23-25, 1997.

Formal Topology and Information Systems Piero Pagliani1 and Mihir K. Chakraborty2 1

Research Group on Knowledge and Communication Models Via Imperia, 6. 00161 Roma, Italy p.pagliani@agora.stm.it 2 Department of Pure Mathematics, University of Calcutta 35, Ballygunge Circular Road, Calcutta-700019, India mihirc99@vsnl.com

Abstract. Rough Set Theory may be considered as a formal interpretation of observation of phenomena. On one side we have objects and on the other side we have properties. This is what we call a Property System. Observing is then the act of perceiving and then interpreting the binary relation (of satisfaction) between the two sides. Of course, the set of properties can be given a particular structure. However, from a pure ”phenomenological” point of view, a structure is given by the satisfaction relation we observe. So it is a result and not a precondition. Phenomena, in general, do not give rise to topological systems but to pre-topological systems. In particular, ”interior” and ”closure” operators are not continuous with respect to joins, so that they can ”miss” information. To obtain continuous operators we have to lift the abstraction level of Property Systems by synthesizing relations between objects and properties into systems of relations between objects and objects. Such relations are based on the notion of a minimal amount of information that is carried by an item. This way we can also account for Attribute Systems, that is, systems in which we have attributes instead of properties and items are evaluated by means of attribute values. But in order to apply our mathematical machinery to Attribute Systems we have to transform them into Property Systems in an appropriate manner. Keywords: approximation spaces, formal topology, Galois adjunctions, rough sets, information quanta, information systems, pointless topology, pretopology.

1

Introduction

Rough Sets arise from information systems in which items are evaluated against a set of attributes or properties. In Computer Science properties are often interpreted as ”open subsets” of some topological space. M. Smyth pioneered this interpretation in 1983 when he observed that semi-decidable properties are analogous to open sets in a topological space (cf. [28]). This intuition was developed by distinguished scholars such as D. Scott who introduced Domain Theory and the so-called Scott Topology to study continuous approximating maps between structures of information called domains. J.F. Peters et al. (Eds.): Transactions on Rough Sets VI, LNCS 4374, pp. 253–297, 2007. c Springer-Verlag Berlin Heidelberg 2007

254

P. Pagliani and M.K. Chakraborty

This research is paralleled, in a sense, by logical studies such as Cohen’s forcing and Kripke models, where the notion of approximable sets of properties (or approximable information) is the core of the construction. W. Lawvere showed that these constructions can be synthesized into the notion of a topos as the abstract form of continuously variable sets. S. Vickers combined the logical and the Computer Science approach. In [33], a prominent role is played by topological systems where just the formal properties of open sets are considered, without mentioning points (pointless topology). Indeed, this approach originates in Stone’s and Birkhoﬀ’s representation theorems where the notion of an abstract point is de facto, introduced. And an abstract point is nothing else but a bunch of properties (once we interpret the elements of a lattice as properties). Inﬂuenced by P. Martin-L¨ of’s Intuitionistic Type Theory, G. Sambin undertook his own way to deal with pointless topology, and speciﬁcally pointless pretopology, as related to Logic (namely Linear Logic), which led to the notion of a Formal Topology (see [25]) which later on has been presented as a result of a construction arising from binary relations between the concrete side (points) and the abstract side (properties) of an observational system called a Basic Pair (cf. [26]). As a matter of fact, the interrelations between concrete points and abstract properties is considered by Vickers, too. However in Formal Topology one does not impose any pre-established structure on the set of properties, not even that suggested by ”Observation Logic” in Vicker’s approach, which makes a system of observations into a frame1 . In [18] it was noted that the properties of the operators of Formal Topology may be deduced from the fact that they are based on constructors which enjoy adjointness relations. The pretopological approach was applied to account for approximation operators arising from families of Property Systems in [19] and, ﬁnally, to generalize approximation operators arising from single Property Systems ([21] and [22]). Moreover, this machinery was applied to Attribute Systems too, by transforming them into Property Systems (cf. [20]). The latter researches were, in turn, inﬂuenced by A. Skowron, J. Stepaniuk and T. Y. Lin’s pioneering investigations which have shown that neighborhood systems may account for natural organizations of information systems (for this topic and its applications the reader is referred to [11] and [29]). Moreover, it must be noticed that neighborhood systems give rise to pretopological operators which turn into topological operators just under particular conditions. Therefore, we claim that pre-topology is a most natural setting for approximation operators as induced by information systems. Also, this claim ﬁts with recent suggestions on the role of ”true borders” (hence non topological) to account for a more dynamic approach to data analysis and granular computing (see, for instance [19] and for ”rough topologies” see [24]). Speciﬁcally, the present investigation is induced by the observation that, from a very general point of view, Rough Set Theory arises from a sort of ”phenomenological” approach to data analysis with two peculiar characteristics: 1

A frame is a lattice with ﬁnite meets distributing over arbitrary joins.

Formal Topology and Information Systems

255

– data is analyzed statically at a given point in time of a possibly evolving observation activity; – as a consequence, the analyzed data provides us only with an approximated picture of the domain of interest. We shall see that the status of an observation system at a certain point in time is essentially a triple P = G, M, , that we call a Property system were G is a set of objects (also called ”points”), M a set of properties (also called ”formal neighborhoods”) and ⊆ G × M is intended as a fulﬁllment relation2 . From the concept of an ”observation” we shall deﬁne a family of basic ”perception constructors” mapping sets of objects into sets of properties, called intensional constructors, and sets of properties into sets of objects, called extensional constructors. We show that some pairs of constructors from opposite sides, fulﬁll adjunction properties. That is, one behaves in a particular way with respect to properties if and only if the other behaves in a mirror way with respect to objects. Hence, adjunction properties state a sort of ”dialectic” relationship, or mutual relationship, between perception constructors, which is exactly what we require in view of a ”phenomenological” approach. Adjunction properties make some combinations of these basic constructors into generalized closure and generalized interior operators. Particularly, some combinations happen to be pre-topological operators in the sense of Sambin’s Formal Topology. Actually, we shall see that they are generalizations of the approximation operators provided by Rough Set Theory. However, for they are pretopological and not topological, these approximation operators are not continuous, that is, they exhibit ”jumps” in the presence of set-theoretical operations. Therefore we synthesize the structuring properties of a Property system, P, into a second level informational structure G, G, RP , called an Information Quantum Relational System - IQRS, where RP is a relation between objects - hence no longer between objects and properties - embedding the relevant informational patterns of P. In IQRSs adjointness makes second level approximation operators fulﬁll nice properties. Also, this way we shall be able to account for Attribute systems after appropriately transforming them into Property Systems3 . This study aims at presenting the state-of-the-art of a conception of Rough Set Theory as a convergence of diﬀerent approaches and diﬀerent techniques, such as Formal Topology, duality and representation theory, Quantum Logic, adjoint functors and so on, as sketched in the following ﬁgure: 2

3

Property Systems may be also regarded as ”Chu Spaces” ([37]). However Chu Spaces have additional features, namely a pair of covariant and contravariant functors, which links spaces together. For Chu Spaces, see the WWW site edited by V. Pratt, http://boole.stanford.edu/chuguide.html. The term ”Quantum” refers to the fact that a basic information grains is given by the minimal amount of information which is organised by RP around an item what, technically, is linked to the notion of a quantum of information at a location once one substitute ”item” for ”location” - cf. [4].

256

P. Pagliani and M.K. Chakraborty

Fig. 1. A Rough Set connection

More precisely, in the present paper we shall deal with the boxed topics along the ﬂow displayed by bold lines. Dotted arrows display the relationships between logical systems and some topics connected to Rough Sets Theory. Dotted lines show some interesting links between some techniques used in the present paper and topics connected with data and information analysis. Bold arrows display some well-established links between logico-algebraic systems and Rough Set systems, while the dotted arrow marked with ”?” suggests links between the modal-style approach applied in the paper and logico-algebraic interpretations to be explored4 .

2

Formal Relationships Between ”Objects” and ”Observables”

Observation is a dynamic process aimed at getting more and more information about a domain. The larger the information, the ﬁner the picture that we have about the elements of the domain. Using topological terms, an observation process makes it possible to move from a trivial topology on the domain, in 4

We have proved that Rough Set Systems are semi-simple Nelson algebras (or equivalently, three-valued L ukasiewicz algebras) (cf. [14] and [15]). What algebraic systems arise from this generalisation has to be studied yet. Brouwer-Zadeh Lattices, Bilattices and Semi-Post algebras may provide some hints.

Formal Topology and Information Systems

257

which everything is indistinguishable to a topology in which any single element is sharply separable from all the other elements (say a discrete topology or a Hausdorﬀ space). In this case we can ”name” each single element of the domain by means of its characteristic properties. However, in this respect, observation is an asymptotic process. What usually happens is that at a certain point in time we stop our observation process, at least temporarily, and analyse the stock of pieces of information we have collected so far. In a sense we consider a ”ﬂat slice” of the observation process.

Fig. 2. A process of diﬀerentiation via observations

Therefore, our slice is basically composed by: (a) a set G of ’objects’; (b) a set M of ’observable properties’; (c) a relation between G and M , denoted with the symbol . Given g ∈ G and m ∈ M we shall say that if g m, then g enjoys property m, or that g induces the observable property m. We have also to assume that is deﬁned for all the elements of G and M because we consider immaterial any property which cannot help making any distinction among objects and, symmetrically, if an object g does not manifest any property, then it is a ”non-object” from a phenomenological point of view5 . We encode everything in the following deﬁnition: Deﬁnition 1. A triple G, M, where G and M are ﬁnite sets, ⊆ G × M is a relation such that for all g ∈ G there is m ∈ M such that g m, and for all m ∈ M there is g ∈ G such that g m, is called a property system or a P-system. Among P-systems we distinguish: a) Functional systems, or FP-systems, where is functional in the sense that for any element g ∈ G, g m and g m implies m = m . 5

The symbols ”G” and ”M” are after the German terms ”Gegenst¨ ande” (”objects”) and, respectively, ”Merkmale (”properties”). A ”Gegenstand” is what stays in front of a subject, while the German term ”Object” means an interpreted ”Gegenstand”. These are the terms used in Formal Concept Analysis and we adopt them for their philosophical meaning.

258

P. Pagliani and M.K. Chakraborty

b) Dichotomic systems or DP-systems, if for all p ∈ M there is p ∈ M such that for all g ∈ G, g p if and only if g p. Functional and dichotomic systems enjoy particular classiﬁcation properties6 . Moreover we shall also consider Deterministic attribute systems: Deﬁnition 2. A structure of the form G, At, {Va }a∈At , , where G, At and Va are sets (of objects, attributes and, respectively, attribute-values) and for each a ∈ At, a : G −→ Ata , is called a deterministic Attribute System or an Asystem7 . From now on we assume that P always denotes a Property System G, M, and that A denotes an Attribute System G, At, {Va }a∈At , . Moreover, we shall use the following notation: If f : A −→ B and g : B −→ C are functions, then: (a) with (f ◦g)(x) or, equivalently, g(f (x)) we denote the composition of g after f ; (b) f → : ℘(A) −→ ℘(B); f → (X) = {f (a) : a ∈ X} - denotes the image of X via f ; (c) f ← : ℘(B) −→ ℘(A); f ← (Y ) = {a : f (a) ∈ Y } - denotes the pre-image of Y via f ; (d) the set f → (A) is denoted with Imf ; 1A , denotes the identity function on A; (e) the map f o : A −→ Imf ; f o (a) = f (a) denotes the corestriction of f to Imf and the map fo : Imf −→ B; fo (b) = b denotes the inclusion of Imf into B; (f) the equivalence relation kf = {a, a : f (a) = f (a )} is called the kernel of f or the ﬁbred product A ×B A obtained by pulling back f along itself. 2.1

Ideal Observation Situations

If G, M, is an FP-system, we are in a privileged position for classifying objects, for the reasons we are going to explain. The ”best” case is when is an injective function. Indeed in this case the converse relation (or, also, −1 ) is a function, too, and we are able to distinguish, sharply, each object. In mathematical words we can compute the retraction of . Deﬁnition 3. Let f : A −→ B be a function. Then a morphism r : B −→ A is called a retraction of f , if f ◦ r = 1A . But, ﬁrst of all, this is an unusual situation, from a practical point of view. Further, ”observation” and ”interpretation” is a modeling activity. Thus, from an epistemological point of view we may wonder if the best model of a horse is really a horse. Actually, ”modeling” means ”abstracting” and ”abstracting” means ”losing something”, some quality or characteristic. Thus the situation depicted in Figures 3 and 4 cannot be but the result of some reduction process. 6 7

Indeed FP-systems and DP-systems are closely linked together, as we shall see. The traditional term in Rough Set Theory is ”Information System”.

Formal Topology and Information Systems

259

Fig. 3. An ideal situation: the existence of retractions

Fig. 4. Is a horse the best model of a horse?

A second optimal situation is when is a surjective function (what always happens of FP-systems). Indeed, by reading back we obtain an equivalence relation E , so that any element of G will belong to one and just one equivalence class modulo E , without ambiguity and borderline situations, what is a perfect case of a classiﬁcation. Indeed, E is the kernel of and it induces a classiﬁcation of the elements of G through properties in M . This is tantamount to the construction of stalk spaces, or espace etal´e, through ﬁbers (or stalks, sorts). This means that has a section or a co-retraction. Deﬁnition 4. Let f : A −→ B be a function. Then a morphism s : B −→ A is called a section or co-retraction of f , if s ◦ f = 1B . We can interpret f as a way to list or parametrise (some of) the elements of B, through the elements of A. In turn the notions of a section and a retraction are special cases of a more fundamental concept: a divisor.

260

P. Pagliani and M.K. Chakraborty

Deﬁnition 5. Let f : A −→ B, g : B −→ C and h : A −→ C be three functions. Then, g is called a right divisor of h by f and f is called a left divisor of h by g if h = f ◦ g, that is, if the following diagram commutes: B f

A

@ @ g @ @ @ R - C

h From the above deﬁnitions, we immediately deduce that if s is a section of f , then f is a retraction of s; and vice-versa. Moreover, it is not diﬃcult to verify that f does not have any section if it is not surjective on B (otherwise, how would it be possible to obtain 1B ?). Intuitively if there is a b ∈ B that is not f -image of any a ∈ A, then b would be associated with a void sort. Vice-versa, a function f does not have any retraction if f is not injective in B. In fact, if f (a) = f (a ) = b, for a = a , then any morphism from B to A either maps b onto a and forgets a , or it maps b onto a and forgets a, because of unicity of the image, and we could not obtain 1A (thus, for any function f : A −→ A, fo is a section with retraction f o ).

Fig. 5. An almost ideal situation: the existence of co-retractions

If r : A −→ B is a retraction of a function h : B −→ A, then r◦h is an idempotent endomorphism of A: (r ◦ h) ◦ (r ◦ h) = r ◦ (h ◦ r) ◦ h = r ◦ 1B ◦ h = r ◦ h. It follows that if s : B −→ A is a section of f : A −→ B, then f ◦ s is an idempotent endomorphism in A, provided s is onto, because f is a retraction of s (see above). Clearly, if a = s(b), then s(f (a)) = s(f (s(b))) = s(1B (b)) = s(b) = a. Hence, any image of the section s is a ﬁxed point of the endomorphism f ◦ s.

Formal Topology and Information Systems

261

Thus sections, retractions and kernels of a function f make it possible to organise the domain of f into sharp classiﬁcation categories (groups) and to single out representatives of such categories.

3

Categorizing Through Relational P-systems

On the contrary, if we deal with generic relational P-systems, not necessarily functional, it hardly happens to obtain sharp classiﬁcations, at least without a proper manipulation of the given P-system that, in turn, may be or may not be an appropriate maneuver. It follows that the identity relation in the deﬁnition of left and right divisors must be weakened to an inequality relation ”≥” or ”≤”. Therefore, to deal with generic cases we need a more subtle mathematical machinery. Such a machinery is based on the notion of an ”approximation”. However, this notion depends on another one. Indeed, we cannot speak of ”approximation” without comparing a result with a goal and this comparison depends on the granularity of the target and of the instruments to get it. For instance, in a document search system, in general we face a situation in which queries refer to a set of possible answers and not to single objects. Otherwise we would not have ”queries” but ”selections” (the realm of sections and retractions). In other words, objects constitute, in principle, a ﬁner domain than those obtained by any modeling or interpretation activity. So we can distinguish an extensional granulation, related to objects, and an intensional granulation, related to properties, and assume that the extensional granulation is ﬁner than the intensional one. Thus, when we have to determine a point on the extensional scale by means of the intensional ruler, we hardly will be able to get a precise determination. We can approximate it. But in order to be able to have ”best approximations” the intensional granulation and its relationships with the extensional granulation must fulﬁll non trivial properties. First of all we need an order. Suppose X is a set of candidate results of a query. Then we generally do not have a selection criterion to single out elements of X. But if the elements of X are ordered in some way, we can use this order to choose, for instance, the least or the largest element in X, if any.

Fig. 6. A usual P-system needs a scale

262

P. Pagliani and M.K. Chakraborty

But this is not enough. In fact dealing with functions (either surjective or injective) is, in a sense, a lucky event in this business which happens only if we are able to completely reduce a structure to a simpler one. This means that generally we cannot have sections and retractions, so that we cannot directly manipulate pre-images of single elements of the codomain or images of single elements of the domain of some ”connecting” function. On the contrary we have to manipulate some kinds of subsets of the domain and co-domain which, we hope, embed enough ordering features to compute approximations. Having this picture in mind, we underline what follows. From an observational point of view the only relationships between objects are induced by the fulﬁllment relation and they are grouping relationships so that we can compare subsets of objects (or properties) but not, directly, objects (or properties). In other words in this paper we assume that there is no relation (hence any order) either between objects or between properties. Hence the result of an approximation activity is, generally, a ”type” not a ”token”8. It follows that we shall move from the level of pure P-systems G, M, to that of Perception systems ℘(G), ℘(M ), {φi }i∈I where φi is a map from ℘(G) to ℘(M ) or from ℘(M ) to ℘(G).

Fig. 7. Approximation deals with types, not with tokens

4

Concrete and Formal Observation Spaces

Given a P-system, the ﬁrst, and obliged, step we have to do is ”observing”, in the elementary sense of ”perceiving” the manifested properties. Thus if P is a P-system let us deﬁne an ’observation function’ obs : G → ℘(M ), by setting m ∈ obs(g) ⇔ g m.

(1)

Technically, obs is what is called a constructor for it builds-up a set from a point. Indeed, for each point g, obs(g) = {m ∈ M : g m}. We shall call obs(g) the ’intension of g’. In fact, any element g appears through the series of its observable properties, so that obs(g) is actually the intensional description of g. The intension of a point g is, therefore, its description through the observable properties listed in M . We shall also say that if g m (i. e. if m ∈ obs(g)), then m is an observable property connected with g and that g belongs to the ﬁeld of m. 8

By the way, note that in [3], classiﬁcation is achieved at ”type” level.

Formal Topology and Information Systems

263

Symmetrically we can introduce a ”substance function” sub : M → ℘(G) deﬁned by setting g ∈ sub(m) ⇔ g m. (2) This symmetry reﬂects the intuition that a point can be intensionally conceived as the set of properties it is connected with, just as a property may be extensionally conceived as the set of points belonging to its ﬁeld. Dually to obs, given a property m ∈ M , sub(m) = {g ∈ G : g m}, so that sub(m) is the ’extension’, or the ﬁeld, of m. The link between these two functions is the relation : g ∈ sub(m) ⇔ m ∈ obs(g) ⇔ g m, ∀g ∈ G, ∀m ∈ M

(3)

Fig. 8. A ﬁrst level perception process

We now notice that since the set M is given and ﬁxed, any P − system will provide only partial observations of the members of G so that a single point x possibly fails to be uniquely described by its intension obs(x). We shall also say that obs(x) is an intensional approximation of a ’partially describable’ member x of G and claim that if obs(x) = obs(y), then x and y cannot be discerned by means of the observable approximating properties (or ”partial descriptions”) at hand, so that x and y will be said to be indiscernible in the given P-system. If x and y are indiscernible they will collapse into the same intentional description. Indeed, if obs fails to be injective then we know that it cannot have a retraction and this means that the identity 1℘(G) cannot be determined by means of the properties at our disposal (that is, the subsets of M mapped by obs), so that a ”loss of identity” literally happens. However, we can deﬁne, by means of obs and sub some approximation operators. 4.1

The Basic Perception Constructors

The second step after observing, is an initial interpretation of what we have observed. Thus we shall introduce the ”perception constructors” that are induced by a P-system.

264

P. Pagliani and M.K. Chakraborty

These constructors will make it possible to deﬁne diﬀerent kinds of structures over ℘(G) and ℘(M ). Since such structures are deﬁned as extensions of the two functions obs and sub and since, in turn, these two functions are linked by the relation (3), it is clear that any structurization on points will have a dual structurization on observables, and vice-versa. Deﬁnition 6 (Basic contructors). Let P = G, M, be a P-system. Then: – – – – – –

e : ℘(M ) −→ ℘(G); e(Y ) = {g ∈ G : ∃m(m ∈ Y & g ∈ sub(m))}; [e] : ℘(M ) −→ ℘(G); [e](Y ) = {g ∈ G : ∀m(g ∈ sub(m) =⇒ m ∈ Y )}; [[e]] : ℘(M ) −→ ℘(G); [[e]](Y ) = {g ∈ G : ∀m(m ∈ Y =⇒ g ∈ sub(m))}; i : ℘(G) −→ ℘(M ); i(X) = {m ∈ M : ∃g(g ∈ X & m ∈ obs(g))} [i] : ℘(G) −→ ℘(M ); [i](X) = {m ∈ M : ∀g(m ∈ obs(g) =⇒ g ∈ X)}; [[i]] : ℘(G) −→ ℘(M ); [[i]](X) = {m ∈ M : ∀g(g ∈ X =⇒ m ∈ obs(g))}.

Fig. 9. Basic constructors derived from a basic pair

An intuitive interpretation of the above functions is in order. As for the constructors decorated with ’e’ (because the result of the function is an extent), we notice that if we want to extend function sub from elements to subsets of M , we have essentially two choices: a ”disjunctive” or ”existential” extension and a ”conjunctive” or ”universal” extension. The former is e while the latter is [[e]]. Obviously, e = sub→ = obs← . It is not diﬃcult to see that [e] is the dual of e, hence it is the ”co-existential extension” of sub (the dual of [[e]] is not discussed in this paper). Given Y ⊆ B, the set [[e]](Y ) collects the points that fulﬁll at least all the properties from Y (and, possibly, others), while e(Y ) gives the set of points which fulﬁll at least one property from Y . Finally, [e](Y ) collects the points which fulﬁll at most all the properties from Y (but possibly not all the properties in Y ). The same considerations apply symmetrically to the operators decorated with ’i’ (because the result of these functions is an intent). Indeed i and [[i]] are the disjunctive and, respectively, conjunctive extensions to subsets of G of the function obs and i = obs→ = sub← . More precisely, i(X) collects the set of properties that are fulﬁlled at least by one point of X, while [[i]](X) collects the set of properties that are fulﬁlled at least by all the points of X, that is, the properties com-

Formal Topology and Information Systems

265

mon at least to all the points of X. Finally, [i](X), the ”co-existential extension” of obs, gives the set of properties that are fulﬁlled at most by all of the points in X. In particular, [i]({x}) is the set of properties which uniquely characterize x. We summarize these remarks in the following table: ... property/ies in Y ... point/s in X at least one ... e(Y ) i(X) at least all ... [[e]](Y ) [[i]](X) at most all ... [e](Y ) [i](X) If one of the above or following operators on sets, say Op, is applied to a singleton {x} we shall also write Op(x) instead of the correct Op({x}), if there is no risk of confusion. 4.2

A Modal Reading of the Basic Constructors

At this point a modal reading of the basic constructors is in order. This will be formalized in Section 6. Indeed, we can read these constructors by means of operators taken from extended forms of modal logic, namely, possibility, necessity and suﬃciency. x∈X b∈B if x ∈ X then b is enjoyed by i(X) = B it is possible that some element x enjoys elements in B collected in X Operator

[i](X) = B

to enjoy elements in B b is enjoyed by it is necessary at most all the to be in X element of X

[[i]](X) = B

to enjoy elements in B b is enjoyed by it is suﬃcient at least all the to be in X elements of X

Example reading there are examples of elements in X that enjoy b there are not examples of elements enjoying b that are not in X there are not examples of elements of X that do not enjoy b

Remarks: Suﬃciency was introduced in modal logic by [10]. Recently it was discussed in [8] and in [16] from an informational point of view. Suﬃciency happens to be the fundamental operator to deﬁne Formal Concepts, which are pairs of the form [[e]][[i]](X), [[i]](X), for X ⊆ G (see [34]). From the point of view of pointless topology, the operators [i], i, [e] and e have been studied by [26], (where the notation ””, ”♦”, ”rest” and, respectively, ”ext” is used) and in [19]. From an informational point of view they have been investigated in [8], [7], [36], [21] and [22]. It is worth noticing that variations of concept lattices have been introduced by means of these operators, namely ”object oriented concepts” of the form (e[i](X), [i](X) (by Y. Y. Yao) and ”property oriented concepts” of the form [e]i(X), i(X) (by D¨ untsch and Gegida). From Proposition 2 and Corollary 3 below one can easily deduce some of the properties discussed in the quoted papers. A pictorial description of the above modal reading follows:

266

P. Pagliani and M.K. Chakraborty

Fig. 10. Possibility, Necessity, Suﬃciency

The above functions are linked by some structural relationships. First, recall that our operators are deﬁned on Boolean algebras of type ℘(A), ∩, ∪, A, −, ∅, where A is either G or M , so that negation coincides with settheoretical complementation. We say that if an operator Opo is obtained by negating all of the deﬁning subformulas in an operator Op, and by further applying the contraposition law according to negations (or, equivalently, by ﬁrst putting the deﬁnition in disjunctive normal form), then Opo and Op are called opposite or orthogonal (to each other), or ”o” in symbols9 . If Opd (X) =∼ Op(∼ X) then Opd is called the dual of Op and we denote the relation of duality with ”d”. Furthermore we can easily observe that functions decorated with e and functions decorated with i are symmetric with respect to the relation , and we denote this fact with ”s”. The following table summarizes these relationships between basic operators (some of these connections are well known in literature: cf. [26], [7] and [8] - but see also the literature about Galois connections): e i [e] [i] [[e]] [[i]] e = s d sd od ods i s = sd d ods od [e] d sd = s o os [i] sd d s = os o [[e]] od ods o os = s [[i]] ods od os o s = Obviously, symmetric functions fulﬁll the same formal properties, opposite functions fulﬁll opposite properties, while dual and symmetric-dual operators fulﬁll dual properties.

5

Fundamental Properties of the Basic Constructors

Let us investigate the fundamental properties of basic constructors. We carry on this job in a more general dimension concerning binary relations at large. 9

So, for instance, if α =⇒ β appears in a deﬁning formula, of Op, then in Opo we have ∼ β =⇒∼ α.

Formal Topology and Information Systems

267

Deﬁnition 7. Let R ⊆ A × B and Q ⊆ C × D be binary relations, X ⊆ A, Y ⊆ B, x ∈ A, y ∈ B. Then we deﬁne10 : 1. R = {y, x : x, y ∈ R} - the inverse relation of R. 2. R (X) = {y ∈ B : ∃x ∈ X(x, y ∈ R)} - the left Peirce product of R and X. We shall also call R (X) the R- neighborhood of X. In particular, if X is a singleton {x}, then we shall usually write R (x) instead of R ({x}). 3. R(Y ) = {x ∈ A : ∃y ∈ Y (x, y ∈ R)} - the left Pierce product of R and Y . Clearly, R(Y ) is the R -neighborhood of Y . 4. [R ](X) = {y ∈ B : ∀x(x, y ∈ R =⇒ x ∈ X)} - the right residual of R and X. 5. [R](Y ) = {x ∈ A : ∀y(x, y ∈ R =⇒ y ∈ Y )} - the right residual of R and X. 6. [[R ]](X) = {y ∈ B : ∀x(x ∈ X =⇒ x, y ∈ R)} - the left residual of X and R . 7. [[R]](Y ) = {x ∈ A : ∀y(y ∈ Y =⇒ x, y ∈ R)} - the left residual of X and R. 8. R⊗Q = {a, d : ∃z ∈ B∩C(a, z ∈ R & z, d ∈ Q)} - the right composition of R with Q or the left composition of Q with R. If deﬁned, R ⊗ Q ⊆ A × D. Lemma 1. Let R ⊆ A × B. Then for any X ⊆ A, Y ⊆ B, a ∈ A, b ∈ B: 1. (a) b ∈ [R ] (X) iﬀ R(b) ⊆ X; (b) a ∈ [R] (Y ) iﬀ R (a) ⊆ Y ; 2. (a) a ∈ [[R]] (Y ) iﬀ Y ⊆ R (a); (b) b ∈ [[R ]] (X) iﬀ X ⊆ R(b); 3. (a) [[R]](∅) = A, (b) [[R ]](∅) = B, (c) R (∅) = R(∅) = ∅, (d) [R ](A) = B. (e) If R is onto then R (A) = B and [R ](∅) = ∅; (f) [R](B) = A. (g) If R is onto then R(B) = A and [R](∅) = ∅. 4. If X and Y are singletons, then (a) R (X) = [[R ]] (X); (b) R(Y ) = [[R]] (Y ); 5. (a) If R is onto, [R ](X) ⊆ R (X); (b) If R is onto [R](Y ) ⊆ R(Y ); 6. If R is a functional relation then [R](Y ) = R(Y ); 7. If R is a functional relation then [R ](X) = R (X). Proof. (1) (a) By deﬁnition b ∈ [R ] (X) iﬀ ∀a(a, b ∈ R =⇒ a ∈ X) iﬀ R(b) ⊆ X.(b) By symmetry. (2) from (1) by swapping the position of the 10

As the reader will probably note, the operations we denote with R (X) and R(Y ) are often denoted with R(X) and, respectively, R (Y ). Moreover, the left composition of R with Q is usually denoted with R; Q in mathematical literature. However, there are several reasons which suggest to adopt the following symbols, mostly depending on both logic and relational algebra. In particular logical reasons are related to Kripke models for Modal Logics and, as to the left composition, to Linear Logic. Apart from symbols, R(Y ) coincides with what in [26] is called the ”extension” of Y along R and [R](Y ) the ”restriction” of Y along R . In Formal Concept Analysis [[R]](Y ) is the ”(derived) extent” of Y , while [[R ]](X) is called the ”(derived) intent” of X. The terminology used here is that of Relation Algebra and connected topics, (strictly speaking, residuals are deﬁned provided X and Y are right ideal elements - see for instance [16]).

268

P. Pagliani and M.K. Chakraborty

relations ∈ and R. (3) (a), (b) and (c) are obvious. (d) For any b ∈ B, either a, b ∈ R for some a ∈ A or the premise of the implication deﬁning the operator [R ] is false. (e) If R is surjective then for all a ∈ A there is a b ∈ B such that a, b ∈ R. Moreover, in [R ](∅) the consequence is always false. Similar proofs for (f) and (g). (4) Applied on singletons the deﬁnitions of [[α]] and α operators trivially coincide, for α = R or α = R . (5) For all b ∈ B, b ∈ [R ](X) iﬀ R(b) ⊆ X iﬀ (for isotonicity of R ) R (R(b)) ⊆ R (X). But b ∈ R (R(b)). Hence b ∈ R(X). Symmetrically for [R] and R. (6) If R is a functional relation, by deﬁnition R is onto, thus from point (5) [R](Y ) ⊆ R(Y ) for any Y ⊆ B. Suppose x ∈ R(Y ) and x ∈ / [R](Y ). Then there is y ∈ Y such that x, y ∈ R and there is a y ∈ / Y such that x, y ∈ R and x, y ∈ R. Hence R is not functional. (7) It is an instance of (6). 5.1

Solving the Divisor Inequalities

Now we come back for a while to the divisor diagram of Deﬁnition 5. Our instance of this diagram reads as in Figure 3. Thus we have to understand under what conditions we can have ”best approximating” maps. Therefore, suppose in general φ is a function which maps subsets of a set A into subsets of a set B (possibly in dependence on how the elements of A are related via a binary relation R ⊆ A × B with the members of B). If φ(X) ⊇ Y we can say that X approximates Y from above via φ. The smallest of these X can therefore be thought of as a ”best approximation from above” via φ, for its image is the closest to Y . In order to get such a best approximation, if any, we should take φ← (↑ Y ), where ↑ Y = {Y ⊆ B: Y ⊇ Y }. In fact φ← (↑ Y ) = {X : φ(X) ⊇ Y }. Dually, if we take φ← (↓ Y ), where ↓ Y = {Y ⊆ B : Y ⊆ Y }, we should obtaina ”best approximation from below” of Y , if any, via φ, because φ←1 (↓ Y ) = {X : φ(X) ⊆ Y }. To be sure, this approach is successful if φ( φ← (↑ Y )) ⊇ Y and, dually, φ( φ← (↓ Y )) ⊆ Y . So we now shall examine, in an abstract setting, the conditions under which the above operations are admissible and behave as expected. Indeed, we have a mathematical result which states rigorously these informal intuitions11 . 11

We remind that in a preordered set O: (a) ↑ X = {y : ∃x(x ∈ X & x y)} = (X) is called the order ﬁlter generated by X. In particular ∀p ∈ A, ↑ p =↑ {p} is called the principal order ﬁlter generated by p. If O is partially ordered p = min(↑ p), where, given a set X, min(X) is the minimum element of X. (b) ↓ X = {y : ∃x(x ∈ X & y x)} = (X) is called the order ideal generated by X. In particular, ∀p ∈ A, ↓ p =↓ {p} is called the principal order ideal generated by p. If O is partially ordered p = max(↓ p), where, given a set X, max(X) is the maximum element of X. From now on O = A, ≤ and O = A , ≤ will denote preordered or partial ordered sets. Furthermore, with L and L we shall denote two arbitrary complete and bounded lattices L = L, ∨, ∧, 0, 1 and, respectively, L = L , ∨ , ∧ , 0 , 1 .

Formal Topology and Information Systems

269

Proposition 1. Let A and B be partially ordered sets, φ a functor (i. e. an isotone map) between A and B. Then, the following conditions are equivalent: 1. (a) there exists a functor ψ : B −→ A such that φ◦ψ ≥A 1A and ψ◦φ ≤B 1B ; (a’) for all b ∈ B, φ← (↓ b) is a principal order ideal of A. 2. (b) there exists a functor ϑ : A −→ B such that ϑ◦φ ≥B 1B and φ◦ϑ ≤A 1A ; (b’) for all a ∈ A, ϑ← (↑ a) is a principal ﬁlter of B. The proof can be found in [5]. 5.2

Galois Adjunctions and Galois Connections

The conditions stated in Proposition 1 deﬁne a basic mathematical notion which is at the core of our construction (NOTE: the following materials are known and we have included them to render completeness to this paper). Deﬁnition 8. Let σ : O −→ O and ι : O −→ O be two maps between partial ordered sets. Then we say that ι and σ fulﬁll an adjointness relation if the following holds: ∀p ∈ O, ∀p ∈ O , ι(p ) ≤ p if and only if p ≤ σ(p)

(4)

If the above conditions hold, then σ is called the upper adjoint of ι and ι is called the lower adjoint of σ. This fact is denoted by O ι,σ O

(5)

and we shall say that the two maps form an adjunction between O and O . If the two preorders are understood we shall denote it with ι σ, too12 . When an adjointness relationship holds between two preordered structures we say that the pair σ, ι forms a Galois adjunction or an axiality. This name is after the notion of a Galois connection which is deﬁned by means of a similar but covariant condition where, indeed, ι and σ are antitone: ∀p ∈ O, ∀p ∈ O , ι(p) ≥ p if and only if p ≤ σ(p )

(6)

We read this fact by saying that the pair σ, ι forms a Galois connection or a polarity. Clearly, a Galois connection is a Galois adjoint with the right category O turned into its opposite Oop . In other words, σ, ι is a polarity if and only if O ι,σ Oop . 12

Sometimes in mathematical literature, the lower adjoint is called ”left adjoint” and the upper adjoint is called ”right adjoint”. However, the reader must take care of the fact that we have two levels of duality. The ﬁrst swaps the partial order (≤ into ≥ and vice-versa). The second swaps the order of application of the functors (ι ◦ σ into σ ◦ ι, and the other way around) and the position of the two structures (by the way, we notice that in usual literature given a map φ, the upper residual is denoted with φ∗ and the lower residual is denoted with φ∗ ).

270

P. Pagliani and M.K. Chakraborty

We now state without proof a number of properties fulﬁlled by adjoint maps. Proposition 2. Let σ : O −→ O and ι : O −→ O be mappings, p ∈ O and p ∈ O . Then, (a) the following statements are equivalent: (a.1) O ι,σ O; (a.2) σι(p ) ≥ p and ισ(p) ≤ p, and both ι and σ are isotone; (a.3) σ is isotone and ι(p ) = min(σ ← (↑ p )); (a.4) ι is isotone and σ(p) = max(ι← (↓ p)); If O ι,σ O, then: (b) σ preserves all the existing infs and ι preserves all the existing sups; (c) ι = ισι, σ = σισ; (d) σι and ισ are idempotent. (e) σ is surjective iﬀ ι(p ) = min(σ ← ({p })) iﬀ σι(p ) = p iﬀ ι is injective; (f) σ is injective iﬀ σ(p) = max(ι← ({p})) iﬀ ισ(p) = p iﬀ ι is surjective. Notice that (e) and (f) are the reader’s digest of the story about retraction and coretraction we told in Section 2.1. Now we have a good stock of results in order to ”implement” a suﬃciently large body of useful operators, actually those operators which will constitute the backbone of all the present story. Deﬁnition 9. Let φ : O → O be an operator on a partially ordered set and ϑ : L → L be an operator between two lattices. Then, (1) φ is a projection operator on O iﬀ it is isotone and idempotent; (2) a projection operator is a closure operator iﬀ it is increasing; (3) a projection operator is an interior operator iﬀ it is decreasing; (4) ϑ is a modal operator iﬀ it is normal (i. e. ϑ(0) = 0 ) and additive; (5) a closure operator ϑ on a lattice is topological iﬀ it is modal; (6) ϑ is a co-modal operator iﬀ it is co-normal (i. e. ϑ(1) = 1 ) and multiplicative; (7) an interior operator ϑ on a lattice is topological iﬀ it is co-modal; (8) ϑ is an anti-modal operator iﬀ it is anti-normal (i. e. ϑ(0) = 1 ) and antiadditive (i. e. ϑ(x ∨ y) = ϑ(x) ∧ ϑ(y)). Notice that in our deﬁnition of modal operators we do not require L = L . Then from Proposition 2 we immediately obtain: Corollary 1. Let O ι,σ O and O ε,ς Oop hold (hence the latter is a Galois connection between O and O ). Then, 1. (a) σι is a closure operator on O ; (b) ισ is an interior operator on O. 2. (a) ςε is a closure operator on O ; (b) ες is a closure operator on O. It is worth underlining that none of these operators needs to be topological. Moreover given the above adjointness situations we can underline what follows:

Formal Topology and Information Systems

271

(m1) σ is half of a co-modal operator: it lacks co-normality; (m2) ι is half of a modal operator: it lacks normality; (m3) and ς are half of an anti-modal operator: they lack anti-normality. The lack of properties concerning preservation of operations may be partially amended when we restrict domains to the families of ﬁxed points of the operators σι, ισ, ες and ςε. To this end the following two results are fundamental: Lemma 2. Let φ : O −→ O be a map. Then, o o (a) if φ is closure then O φ ,φo Imφ ; (b) if φ is interior then Imφ φo ,φ O. Corollary 2. Let φ : L −→ L be a map. Then, 1. if φ is closure then φo is additive, φo (Imφ ) is closed under infs, and Satφ (L) = Imφ , ∧, , 1, where for all x, y ∈ Imφ , x y = φ(x ∨ y), is a lattice; 2. if φ is interior then φo is multiplicative, φo (Imφ ) is closed under sups, and Satφ (L) = Imφ , , ∨, 0, where for all x, y ∈ Imφ , x y = φ(x ∧ y), is a lattice. We want to point out that if φ is closure then sups in L and sups in Imφ may diﬀer. Hence, although for all x ∈ L, φ(x) = φo (x) and φo is additive, nonetheless φ in general is not sup-preserving so that φo (Imφ ) is not closed under sups (dually if φ is interior). These results give the following proposition (where we have just to notice that turning Lop upside-down interiors turns into closures, sups into infs and viceversa): Proposition 3. Let L ι,σ L and L ε,ς Lop hold. Then: 1. Satισ (L) = Imισ , , ∨, 0, where for all x, y ∈ Imισ , x y = ισ(x ∧ y), is a lattice; 2. Satσι (L ) = Imσι , ∧ , , 1 , where for all x, y ∈ Imσι , x y = σι(x ∨ y), is a lattice; 3. Satςε (L ) = Imςε , ∧ , , 1 , where for all x, y ∈ Imςε , x y = ςε(x ∨ y), is a lattice; 4. Satες (L) = Imες , ∧, , 1, where for all x, y ∈ Imες , x y = ες(x ∨ y), is a lattice.

6

Formal Operators on Points and on Observables

Now let us come back to our basic constructors. Proposition 4 (Fundamental relationships). Let A and B be two sets, X ⊆ A, Y ⊆ B, R ⊆ A × B a relation and fˆ ⊆ A × B a functional relation. Then the following holds: 1. (a) R(Y ) ⊆ X iﬀ Y ⊆ [R ](X); (b) R (X) ⊆ Y iﬀ X ⊆ [R](Y ) . 2. Y ⊆ [[R ]](X) iﬀ X ⊆ [[R]](Y );

272

P. Pagliani and M.K. Chakraborty

3. fˆ (X) ⊆ Y iﬀ X ⊆ fˆ(Y ); 4. The operators R, R , [R] and [R ] are isotone; [[R]] and [[R ]] are antitone. Proof. (1) (a) R(Y ) ⊆ X iﬀ R (y) ⊆ X, ∀y ∈ Y iﬀ (from Lemma 1.(1)) Y ⊆ [R ](X). (b) By symmetry. (2) X ⊆ [[R]](Y ) iﬀ ∀x ∈ X(Y ⊆ R(x)) (from Lemma 1.(2)), iﬀ ∀x ∈ X, ∀y ∈ Y (y ∈ R(x)) iﬀ ∀y ∈ Y (X ⊆ R (y)) iﬀ ∀y ∈ Y (y ∈ [[R ]](X)) iﬀ Y ⊆ [[R ]](X) (in view of Lemma 1.(1).(b)). (3) Directly from Proposition 1. (4) Easily from the position of the subformula ”y ∈ Y ” and ”x ∈ X” in the deﬁnitions. From the above discussion we trivially have: = i = e [ ] = [i] [] = [e] [[ ]] = [[i]] [[]] = [[e]] Therefore it is clear that if given a P-system, P, we set M = ℘(M ), ⊆ and G = ℘(G), ⊆, in view of Proposition 4 the following adjointness properties hold: (a) M e,[i] G; (b) G i,[e] M; (c) M [[e]],[[i]] Gop ; (d) G [[i]],[[e]] Mop . The lack of properties involving top and bottom elements, such as ”normality” and ”co-normality”, for generic adjoint functions, is quite obvious since they depend on the adjoint structures. But in the case of the basic constructors Lemma 1.(3) and Proposition 2 immediately prove that e and i are modal operators, [e] and [i] are co-modal operators and, ﬁnally, [[e]] and [[i]] are anti-modal operators. Moreover, in view of these adjunction properties, some sequences of constructors with alternate decorations provide a number of useful operators on ℘(G) and ℘(M ). Indeed axiality says that if one operator lowers an element then its conjugate operator lifts it, and vice-versa, so that by combining them either we obtain the maximum of the lowering elements or the minimum of the lifting elements of a given argument. Deﬁnition 10. Let G, M, be a P -system. Then: – – – – – –

int : ℘(G) −→ ℘(G); int(X) = e([i] (X)). cl : ℘(G) −→ ℘(G); cl(X) = [e] (i(X)). est : ℘(G) −→ ℘(G); est(X) = [[e]] ([[i]] (X)). A : ℘(M ) −→ ℘(M ); A(Y ) = [i] (e(Y )). C : ℘(M ) −→ ℘(M ); C(Y ) = i([e] (Y )). IT S : ℘(M ) −→ ℘(M ); IT S(Y ) = [[i]] ([[e]] (Y )).

The above operators inherit ’d’, ’o’, ’s’, ’sd’, ’os’, ’od’ and ’ods’ reciprocal relationships from the outermost constructors which deﬁne them. Proposition 5. In any P-system G, M, , for any X ⊆ G, Y ⊆ M , g ∈ G, m ∈ M: 1. (a) m ∈ A(Y ) iﬀ e(m) ⊆ e(Y ), (b) g ∈ cl(X) iﬀ i(g) ⊆ i(X); 2. (a) g ∈ int(X) iﬀ i(g) ∩ [i] (X) = ∅, (b) m ∈ C(Y ) iﬀ e(m) ∩ [e] (Y ) = ∅; 3. (a) g ∈ est(X) iﬀ [[i]](X) ⊆ i(g), (b) m ∈ IT S(Y ) iﬀ [[e]](Y ) ⊆ e(m).

Formal Topology and Information Systems

273

Proof. (1) (a) By deﬁnition m ∈ A(Y ) iﬀ m ∈ [i] (e(Y )). Hence from Lemma 1.(1), m ∈ A(Y ) iﬀ e(m) ⊆ e(Y ). (b) By symmetry. (2) (a) g ∈ int(X) iﬀ g ∈ e([i] (X)) iﬀ g ∈ e({m : e(m) ⊆ X}), iﬀ i(g) ∩ {m : e(m) ⊆ X} = ∅, iﬀ i(g) ∩ [i] (X) = ∅. (b) By symmetry. (3) (a) Directly from Lemma 1.(2) and the deﬁnition of ”est”. (b) By symmetry13 . Therefore, g ∈ est(X) if and only if g fulﬁlls at least all the properties that are shared by all the elements of X. In this sense est(X) is the extent of the set of properties that characterises X as a whole. Symmetrically, m ∈ IT S(Y ) if and only if m is fulﬁlled by at least all the objects that enjoy all the properties from Y . In this sense IT S(Y ) is the intent of the set of objects that are characterised by Y as a whole. In order to understand the meaning of the other operators, let us notice that the elements of M can be interpreted as ”formal neighborhoods”14 . In fact, in topological terms a neighborhood of a point x is a collection of points that are linked with x by means of some nearness relation. For a member m of M is associated, via e with a subset X of G, m may be intended as a ’proxy’ of X itself. Thus if X is a concrete neighborhood of a point x, then m may be intended as a formal neighborhood of x, on the basis of the observation that the nearness relation represented by X states that two points are close to each other if they both fulﬁll property m15 . It follows that obs(g) is the family of formal neighborhoods of g (symmetrically for sub(m) we have the concrete neighborhoods of m). This is the intuitive content of the following gift of the adjointness relationships between basic constructors: Interior operators int, C Closure operators cl, A, est, IT S In view of the observation after Corollary 1 one easily notices that none of the above operators needs to be topological. 6.1

Fundamental Properties of the Formal Perception Operators

Deﬁnition 11. Let P be any P-system. Then we deﬁne the following families of ﬁxpoints of the operators induced by P: 1. Ωint (P) = {X ⊆ G : int(X) = X}; Γcl (P) = {X ⊆ G : cl(X) = X}; 2. Γest (P) = {X ⊆ G : est(X) = X}; ΩA (P) = {Y ⊆ M : A(Y ) = Y }; 3. ΓC (P) = {Y ⊆ M : C(Y ) = Y }; ΓIT S (P) = {Y ⊆ M : IT S(Y ) = Y }. 13

14

15

Moreover, [[i]](X) = {m : X ⊆ e(m)} = {m : ∀x ∈ X(x m)}. Henceforth [[i]](X) ⊆ i(g) iﬀ g m for all m such that x m, for any member x of X, that is, iﬀ ∀m ∈ M ((∀x ∈ X(x m)) =⇒ g m). Indeed, this is the framework in which the operators cl, int, A and C are introduced, although not by means of adjointness properties, by the Padua School of Formal Topology (see [26]). This interpretation is close to the approach of [11].

274

P. Pagliani and M.K. Chakraborty

It is understood that the partial order between saturated subsets is inherited from the category they are derived from. Thus, for instance, we shall have ΓC (M ), ⊆. Proposition 6. Let P be a P-system. Then the following are complete lattices: 1. Satint (P) = Ωint (P), ∪, ∧, ∅, G, where i∈I Xi = int( i∈I Xi ); 2. SatA (P) = ΩA (P), ∨, ∩, ∅, M , where i∈I Yi = A( i∈I Yi ); 3. Satcl (P) = Γcl (P), ∨, ∩, ∅, G, where i∈I Xi = cl( i∈I Xi ); 4. SatC (P) = ΓC (P), ∪, ∧, ∅, M , where i∈I Yi= C( i∈I Yi ); 5. Satest (P) = Γest (P), ∩, ∨, est(∅), G, where i∈I X i ); i = est( i∈I X 6. SatIT S (P) = ΓIT S (P), ∩, ∨, IT S(∅), M , where i∈I Yi = IT S( i∈I Yi ). Proof. Much work has already been done in Proposition 3. We just need to justify the choice of top and bottom elements. To this end, remember that in any P-system both and are onto. Hence in view of Lemma 1.(3). int(G) = e[i](G) = [ ](G) = (M ) = G, and analogously for the other operators. The only diﬀerence is for IT S and est because [[]](∅) = G but [[ ]](G) = {m : (m) = G} ⊇ ∅, dually for [[]][[ ]](∅). Lemma 3. Let P be a P-system. Then for all X ⊆ G, Y ⊆ M , X ∈ Ωint (P) iﬀ X = e(Y ) X ∈ Γcl (P) iﬀ X = [e](Y ) X ∈ Γest (P) iﬀ X = [[e]](Y ) Y ∈ ΩA (P) iﬀ Y = [i](X ) Y ∈ ΓC (P) iﬀ Y = i(X ) Y ∈ ΓIT S (P) iﬀ Y = [[i]](X )

for some Y ⊆ M, X ⊆ G. Proof. If X = e(Y ) then X = e[i]e(Y ), from Proposition 2.(c). Therefore, by deﬁnition of int, X = int(e(Y )) = int(X). Vice-versa, if X = int(X), then X = e[i](X). Hence, X = e(Y ) for Y = [i](X). The other cases are proved in the same way, by exploiting the appropriate equations of Proposition 2.(c). Corollary 3. Let P be a P-system. Then the following are isomorphisms: 1. (a) e : SatA (P) −→ Satint (P); (b) [i] : Satint (P) −→ SatA (P); 2. (a) [e] : SatC (P) −→ Satcl (P); (b) i : Satcl (P) −→ SatC (P); 3. The following are anti-isomorphisms (where − is the set-theoretical complementation): (a) [[i]] : Satest (P) −→ SatIT S (P); [[e]] : SatIT S (P) −→ Satest (P); (b) − : Satcl (P) −→ Satint (P); − : SatC (P) −→ SatA (P). Proof. Let us notice, at once, that the proof for an operator requires the proof for its adjoint operator. Then, let us prove (1).(a) and (b) together: First, let us prove bijection for e and [i]. From Lemma 3 the codomain of e is Ωint (P) and the codomain of [i] is ΩA (P). Moreover, for all X ∈ Ωint (P), X = e[i](X) and for all Y ∈ ΩA (P), Y = [i]e(Y ). From the adjointness properties we have:

Formal Topology and Information Systems

275

(i) e is surjective on Ωint (P) and (ii) [i] is injective from Ωint (P). (iii) e is injective from ΩA (P) and (iv) [i] is surjective onto ΩA (P). Moreover, if [i] is restricted to Ωint (P), then its codomain is the set H = {Y : Y = [i](X) & X ∈ Ωint (P)}. Clearly, H ⊆ ΩA (P). In turn, if e is restricted to ΩA (P), then its codomain is the set K = {X : X = e(Y ) & Y ∈ ΩA (P)}. Clearly K ⊆ Ωint (G). Therefore, (i) and (iii) give that e is bijective if restricted to ΩA (P), while (ii) and (iv) give that [i] is a bijection whenever restricted to Ωint (P)16 . Now we have to showthat e and [i] preserve joins and meets. For e we proceed as follows: (v) e( i∈I (A(Yi ))) =def e(A( i∈I (A(Yi ))). But eA = e, from Proposition2.(c). Moreover, e distributes over unions. Hence the right side of (v) equals to i∈I e(A(Yi )). But in view of Proposition 6, the union of extensional open subsets is concrete open and from Lemma 3 e(A(Y i )) belongs indeed to ∈ Ω (P), so that the right side of (v) turns into int( int i∈I e(A(Yi ))) =def e(A(Y )). i i∈I (vi) e( i∈I A(Yi )) = e( i∈I [i]e(Yi )). Since [i] distributes over intersections, the right side of (vi) turns into e[i]( i∈I e(Y )) = int( i i∈I e(Yi )). But e = eA, so that the last term is exactly i∈I e(A(Yi )). Since [i] is the inverse of e, qua isomorphism, we have that [i] preserves meets and joins, too. As to (2) the results come by symmetry. (3) (a) As in the above proof by noticing that in polarities the right structure is reversed upside-down (we can optimize a passage by noticing that [[e]] and [[i]] are both upper and lower adjoints). (b) By duality between the operators. 6.2

Pre-topological Approximation Spaces

Now we are in position to show how the above mathematical machinery may be used to generalise the upper and lower approximation operators provided by Rough Set Theory. Given X ⊆ G we know that [e]i(X) ⊇ X and e[i](X) ⊆ X. We can interpret these relationships by saying that – cl is an upper approximation of the identity map on ℘(G); – int is a lower approximation of the identity map on ℘(G). More precisely, i(X) = min([e]← (↑ X)) = min{X ⊆ G : [e](X ) ⊇ X}, it follows that [e]i(X) (i. e. cl) is the best approximation from above to X via function [e]. Dually, [i](X) = max(e← (↓ X)) = max{X ⊆ G : e(X ) ⊆ X}. Hence, e[i](X) (i. e. int) is the best approximation from below to X, via function e. Of course, if i is injective (or, equivalently, [e] is surjective), then we can exactly 16

As side results, we have: (i) ΩA (P) = H and (ii) Ωint (P) = K. This is not surprising, because if Y ∈ ΩA (P) then Y = [i]e(Z) for some Z ⊆ M and e(Z) ∈ Ωint (P), any Z ⊆ M . Vice-versa, if X ∈ Ωint (P), then X = e(Z). Hence [i](X) = [i]e(Z) belongs to ΩA (P). Symmetrically for (ii).

276

P. Pagliani and M.K. Chakraborty

reach X from above by means of [e]. The element that must be mapped is, indeed, i(X). Dually, if [i] is injective (or e is surjective), then we can exactly reach X from below by means of e applied to [i](X).

7

Information, Concepts and Formal Operators

So far we have discussed a number of instruments that act on either abstraction sides we are dealing with, that is, points and properties. Indeed, the introduced operators are based on well-deﬁned mathematical properties, such as adjointness, and feature proper informational and conceptual interpretations. Also, the use of the terms ”open” and ”closed” is not an abuse because, on the contrary, these operators translate the usual topological deﬁnitions of an interior and, respectively, a closure of a set X ⊆ G, into the language of observation systems, provided the elements of M are interpreted as formal neighborhoods. For instance, the usual deﬁnition tells us that for any subset X of G, a point a belongs to the interior of X if and only if there is a neighbourhood of a included in X. If the elements of the set M are intended as formal neighbourhoods, then the relation a b (hence, a ∈ e(b)) says that b is a formal neighborhood of a and e(b) ⊆ X says that the extension of this neighbourhood b is included in X. But this is precisely a reading of a ∈ e[i](X), because in view of the adjunction properties, e(b) ⊆ X if and only if b ∈ [i] (X). Thus we have made a further step beyond M. Smyth’s seminal observation that semi-decidable properties are analogous to open sets in a topological space, with the aid of the interpretation of basic constructors elaborated by the Padua School on Formal Topology17 . Moreover, we have seen that int and cl provide us with lower and upper approximations of any set X ⊆ G. But are we really happy with this machinery? The answer is ”yes and no”. Yes, for we have found a mathematically sound way to deal with approximations which enjoy a reliable intuitive interpretation. No, for both int and cl are discontinuous (non topological) operators because int is not multiplicative and cl is not additive, so that we have to face ”jumps” which can be too wide and make us miss information. EXAMPLE 1 Here we give an example of a P-system and its induced operators: a a1 a2 a3

17

See for instance [25] and [27].

b b1 1 1 0 1 0 1 0 0

b2 0 0 1 0

b3 0 1 1 1

Formal Topology and Information Systems

277

Let us try and compute some instances of basic formal operators on system P: 1) Extensional operators: int({a, a1 }) = e[i]({a, a1 }) = e({b}) = {a}; int({a2 }) = e({b2 }) = {a2 }. cl({a, a1 }) = [e]i({a, a1 }) = [e]({b, b1 , b3 }) = {a, a1 , a3 }; cl({a, a2 }) = [e](M ) = G. est({b1 , b2 }) = [[i]][[e]]({b1 , b2 }) = [[i]]({a2 }) = {b1 , b2 , b3 }. int(int({a, a1 })) = int({a}) = e[i]({a}) = e({b}) = {a}. cl(cl({a, a1 })) = cl({a, a1 , a3 }) = [e]i({a, a1 , a3 }) = [e]({b, b1 , b3 }) = {a, a1 , a3 }. est(est({b1 , b2 }))= est({b1 , b2 , b3 })= [[i]][[e]]({b1 , b2 , b3 })= [[i]]({a2 })= {b1 , b2 , b3 }. Thus, this is also an example of the fact that int is decreasing while cl and est are increasing and all of them are idempotent. Moreover, one can see that int({a, a1 }) ∪ int({a2 }) = {a, a2 } ⊆ {a, a1 , a2 } = int({a, a1 } ∪ {a2 }) and cl({a, a1 }) ∩ cl({a, a2 }) = {a, a1 , a2 } ⊇ {a} = cl({a, a1 } ∩ {a, a2 }). 2) Intensional operators: A({b, b1 }) = [i]e({b, b2 }) = [i]({a, a1 , a2 }) = {b, b1 , b2 }. C({b2 , b3 }) = i[e]({b2 , b3 }) = i({a3 }) = {b3 }. IT S({b1 , b2 }) = [[i]][[e]]({b1 , b2 }) = [[i]]({a2 }) = {b1 , b2 , b3 }. A(A({b, b1 })) = A({b, b1 , b2 }) = [i]e({b, b1 , b2 }) = [i]({a, a1 , a2 }) = {b, b1 , b2 }. C(C({b2 , b3 })) = C({b3 }) = i[e]({b3 }) = i({a3 }) = {b3 }. IT S(IT S({b1 , b2 }))= IT S({b3 })= [[i]][[e]]({b1 , b2 , b3 }) = [[i]]({a2 }) = {b1 , b2 , b3 }. Thus, this is also an example of the fact that C is decreasing while A and IT S are increasing and all of them are idempotent. Let us now visualise the lattices of saturated sets: Satint (P)

Satcl (P)

G G bb bb "" "" {a1 , a2 , a3 } {a, a1 , a2 } {a1 , a2 , a3 } {a, a1 , a3 } bb bb bb bb "" bb bb {a1 , a3 } {a, a2 } bb bb bb bb "" {a2 } {a} {a3 } {a} bb bb "" "" ∅ ∅ SatA (P) SatC (P) M M bb bb "" "" {b2 , b3 } {b, b1 , b2 } {b1 , b2 , b3 } {b, b1 , b3 } bb bb bb bb "" bb bb {b1 , b3 } {b, b2 } bb bb bb bb "" {b2 } {b} {b3 } {b, b1 } bb bb "" "" ∅ ∅

278

P. Pagliani and M.K. Chakraborty

SatIT S (P)

Satest (P)

G G bb bb "" "" {b, b1 } {b1 , b2 , b3 } {a1 , a2 , a3 } {a, a1 , a2 } bb bb bb bb "" bb bb {a1 , a2 } {b1 , b3 } bb bb bb bb "" {b1 } {b3 } {a2 } {a} bb bb "" "" ∅ ∅

Pay attention that SatIT S (P) and SatA (P) have the same shape just by chance (idem for Satest (P) and Satcl (P)). The non topological nature of these operators is openly visible in the above pictures. For instance in Satint (P) we do not have the intersection of {a1 , a2 , a3 } and {a, a1 , a2 }. Hence int does not distribute over intersections. in Satcl (P) we lack the union of {a3 } and {a}, so that cl does not distribute over unions. However we have a few results which will be useful. Proposition 7. Let θ and φ be two dual basic operators. Then, (a) θ is a closure operator if and only if φ is an interior operator; (b) θ is topological if and only if φ is topological. Proof. (a) Trivially, since complementation reverses the order. (b) Suppose θ is additive, then φ(X ∩Y ) = −θ−(X ∩Y ) = −θ(−X ∪−Y ) = −(θ(−X)∪θ(−Y )) = −θ(−X) ∩ −θ(−Y ) = φ(X) ∩ φ(Y ). Dually for the opposite implication. In order to try and solve the above issue, we must notice that any answer and solution depends on the nature of the P-system at hand. Generally, the nature of points is not really important. More important is the nature of properties. And, even more important is the nature of the operator supposed to better represent the basic perception act. 7.1

Choosing the Initial Perception Act

We have assumed that our ﬁrst act of knowledge is a grouping act, a sort of ”data abstraction”. However, we can basically perform this act in two opposite ways: either collect around an object g the elements which fulﬁlls at least all the properties (or attribute-values) of g, or the elements fulﬁlling at most all the properties (or attribute-values) of g. Otherwise stated, in the ﬁrst case we collect the objects which are characterised at least as g by the properties (attributes) at hand, while in the second case we collect the objects which are characterised at most as g. However if we consider attribute-values the two conditions collapse (see later on). Moreover notice that the grouping rule just asserted does not imply any form of symmetry. Indeed, g could manifest all the properties of g but also additional properties that are not manifested by g. To put it another way, we are claiming

Formal Topology and Information Systems

279

that our basic grouping act is not based on the notion ”to manifest exactly the same properties”, but on the notion ”to manifest at least (or at most) the same properties”. Indeed, the set of properties which are manifested by g, is the attracting phenomenon around which we form our perception. Thus from an analytical point of view we have just to focus on these properties and not to take into account additional properties. In the present paragraph we shall see that the former notion is subsumed by the latter. That is, if we deﬁne this way the basic ”cells” of our categorisation process, we shall be able to cover a wider range of cases. Let G, At, {Va }a∈At be an Attribute System. Then for all g ∈ G deﬁne: Qg = {g : ∀a ∈ At, ∀x ∈ Va ((a(g) = x) =⇒ (a(g ) = x))} Let G, M, be a Property System. Then for all g ∈ G deﬁne: Qg = {g : ∀p ∈ M (p ∈ i(g) =⇒ p ∈ i(g ))} Qg will be called the quantum of information at g. In view of the previous discussion we adopt quanta of information because they reﬂect the idea ”g is perceived together g whenever it manifests at least the same properties as g”. Therefore, given (the properties manifested by) an object g the perception cell organised around g should be Qg which should be referred to as the ”minimum perceptibilium at location g”, for it is not possible to perceive g without perceiving (the manifestations of) the other members of that ”perception parcel”. Therefore, we call such perception parcel a quantum of perception or a quantum of information at g. This terminology drew its inspiration from [4] and this term expresses a sort of programme that may be epitomized by the slogan ”any information is a quantum of information”. As to quanta of information from a Property System, we can elaborate a little further (we shall resume Attribute Systems later on). Proposition 8. Let G, M, be a P-system and g, g ∈ G. Then, 1. 2. 3. 4.

Qg = est(g); g ∈ Qg iﬀ i(g) ⊆ i(g ) g ∈ Qg iﬀ for all X ∈ Γest , g ∈ X =⇒ g ∈ X; g ∈ Qg iﬀ for all p ∈ M, g ∈ e(p) =⇒ g ∈ e(p); g ∈ Qg iﬀ g ∈ cl(g ).

Proof. (1) Indeed, g ∈ est(g) iﬀ i(g ) ⊇ [[i]](g). But [[i]](g) = i(g), whence g ∈ est(g) if and only if p ∈ i(g) =⇒ p ∈ i(g ) if and only if g ∈ Qg if and only if i(g) ⊆ i(g ). (2) () Suppose X ∈ Γest and g ∈ X =⇒ g ∈ X. Then i(g) ⊇ [[i]](X) implies i(g ) ⊇ [[i]](X). Since this happens for all est-saturated X, it happens for [[i]](g) too and we trivially obtain i(g ) ⊇ i(g), so that g ∈ Qg . () If g ∈ Qg then i(g ) ⊇ i(g). If, moreover, g ∈ X, for X = est(X), then i(g) ⊇ [[i]](X). By transitivity, i(g ) ⊇ [[i]](X), whence g ∈ X too. (3)

280

P. Pagliani and M.K. Chakraborty

Indeed, g ∈ e(x) if and only if g x if and only if x ∈ i(g). Hence for all p ∈ M, g ∈ e(p) =⇒ g ∈ e(p) if and only if for all p ∈ M, p ∈ i(g) =⇒ p ∈ i(g ). (4) So, g ∈ Qg if and only if i(g) ⊆ i(g ) if and only if g ∈ cl(g ). Indeed, these results, trivial consequences of Lemma 1 and Proposition 4, formally state that g is perceived together with g if and only if it fulﬁlls at least all the properties fulﬁlled by g. Moreover, a quantum of perception at location g is the universal extension of function sub to the set of properties (g) fulﬁlled by g. In turn, since we start from a singleton {g}, (g) is both a universal and an existential extension of function obs. When we have to move from grouping maneuvers around a single object to grouping maneuvers around two or more objects we have essentially two kinds of choice: universal extensions from singletons to a generic set X and existential extensions. The existential extension is deﬁned as Qx (7) Q∪ X = x∈X

This is not the sole choice, but for the very reasons discussed so far, we shall adopt it. Moreover, it makes a uniform treatment of both P-systems and Asystems possible. As for universal extensions we brieﬂy discuss only the following alternative: Q⊗ X = [[e]][[i]](X) = est(X).

(8)

The superscript ⊗ underlines the fact that in est(X) we consider the properties which glue the elements of X together. Otherwise stated, we extract from i(X) those properties P which are shared by all the elements of X. Then we make exactly the same thing with respect to P . Thus, according to this universal extension, an object g belongs to Q⊗ X if g fulﬁlls all the properties fulﬁlled by all the elements of X. Whenever we need to distinguish the system inducing a quantum of information, we shall use the name of the system as an exponent. The same for P any operator (for instance we shall write QP g and int if needed). Moreover, if O = X, R, we set −O = X, −R and O = X, R . 7.2

Information Quantum Relational Systems

Now that we have chosen the basic mechanisms (basis and step) leading from atomic perception (or ”elementary perception cells”) to complex perception, let us analyse what kinds of a relation arise between elements of G from these grouping maneuvers. Let us then set the following deﬁnition: Deﬁnition 12 (Information Quantum Relational System). Let S be an A-system or a P-system over a set of points G. Let R be a binary relation on G. We say that R is induced by S whenever the following holds, for all g, g ∈ G: g, g ∈ R iﬀ g ∈ Qg .

Formal Topology and Information Systems

281

We call R the information quantum relation - or i-quantum relation or quantum relation in short - induced by S and it will be denoted as RS . Moreover, Q(S) will denote the relational system G, RS , called the Informational Quantum Relational System - IQRS, or Quantum Relational System in short, induced by S. Finally we set ΩQ (S) = {RS (X) : X ⊆ G}18 . Since g ∈ Qg says that g fulﬁlls at least all the properties fulﬁlled by g , then g , g ∈ RS has the same meaning. Clearly the properties of i-quantum relations depend on the patterns of objects induced by the given systems. However, they uniformly fulﬁll some basic properties. In view of the additivity property of generalised quanta, we can conﬁne our attention to i-quanta at a location. Lemma 4. In any A-system S over a set G, for all g, g , g ∈ G: 1. (a) g ∈ Qg (q-reﬂexivity); (b) g ∈ Qg & g ∈ Qg =⇒ g ∈ Qg (qtransitivity). 2. If S is an A-system, a functional or a dichotomic P-system then g ∈ Qg implies g ∈ Qg (AFD-q-symmetry). Proof. The ﬁrst two statements are obvious consequences of transitivity and reﬂexivity of the relation ⊆. Notice that antisymmetry does not hold because of the obvious fact that g ∈ Qg and g ∈ Qg does not imply g = g. Now, let S be an A-system. Suppose g ∈ Qg and a(g) = x, then a(g) = x for some x = x so that a(g ) = x , because g ∈ Qg , whence a(g ) = x too. Therefore, g ∈ Qg implies g ∈ Qg , so that the induced relation is also symmetric. If S is a functional P-system then we trivially obtain the proof from deﬁnitions and the fact that i(g) is a singleton. Finally, if S is dichotomic and g ∈ Qg , then g fulﬁlls at least the same properties as g. Now, if g p while g p, then g p, where p is a complementary copy of p. But g p, since it fulﬁlls p. Hence we cannot have i(g) ⊆ i(g ), whence g ∈ Qg . Contradiction. So notice that in A-systems the universal quantiﬁcation over attribute-values hides a bi-implication because the set of attribute-values of g and that of g must coincide in order to have g ∈ Qg . As an immediate consequence of the above result we have: Proposition 9. Let S be an A-system or a P-system. Then: 1. The i-quantum relation RS induced by S is a preorder; 2. If S is an A-system or an FP or DP system, then RS is an equivalence relation; 3. If S is an FP-system then RS = ⊗ and g ∈ Qg iﬀ g ∈ [g]k . 18

I-quanta and i-quantum relations from A-systems were introduced in [13], with different names. If the entire set M is considered as a multi-valued property, then i-quantum relations coincide with the so-called ”forward inclusion relations” introduced in [12].

282

P. Pagliani and M.K. Chakraborty

Proof. We just have to prove statement (3). In view of Proposition 8.(4) we just need to show that if is a map then for all x ∈ G, cl({x}) = [x]κ . From Proposition 7.1, a ∈ cl({a }) if and only if i({a}) ⊆ i({a }). Therefore, if happens to be a map, we have that a ∈ cl({a }) if and only if i({a}) = i({a }), since exactly one value is admitted. We can conclude that for all x ∈ G, cl({x}) = {x : i(x ) = i(x)} = ei({x}) = [x]k (that is, the kernel of ). EXAMPLE 2 Consider the P-system P of EXAMPLE 1. Let us compute some quanta of information of P: Qa = {a}, Qa = {a , a }, Qa = {a }, Qa = {a , a , a }, Q{a,a } = {a, a , a } and so on addictively. Thus a , a ∈ RS because a ∈ Qa but the opposite does not hold. Here below the i-quantum relations RP and RQ(P) are displayed: RP a a a a

a a 1 0 0 1 0 0 0 1

a 0 1 1 1

a RQ(P) 0 a 0 a 0 a 1 a

a a 1 0 0 1 0 1 0 0

a 0 0 1 0

a 0 1 1 1

It is easy to verify that both of the above relations are reﬂexive and transitive, Q(P) = RQ(P) . Moreover one can see that, for instance, Qa = {a , a } and RP or Qa = {a , a , a }. Indeed we have that a ∈ Qa and a ∈ Qa , or Q(P) a ∈ Qa whereas a ∈ Qa , and so on. Now consider the A-system A = G = {a, a1 , a2 , a3 }, At = {A, B, C}, {VA = {0, 1, 3}, VB = {b, c, f }, VC = {α, δ}} such that A(a) = A(a2 ) = 1, A(a1 ) = 0, A(a3 ) = 3, B(a) = B(a2 ) = b, B(a1 ) = c, B(a3 ) = f, C(a) = C(a1 ) = C(a2 ) = A A A α, C(a3 ) = δ. We have: QA a = Qa2 = {a, a2 }, Qa1 = {a1 }, Qa3 = {a3 }. The resulting i-quantum relation RA = {a, a2 , a2 , a, a, a, a2 , a2 , a1 , a1 , a3 , a3 } is an equivalence relation. Q(P)

Q(P)

From the above results we obtain immediately some interesting consequences about functional P-systems: Corollary 4. Let P be an FP-system. Then, (a) cl is a topological closure operator; (b) int is a topological interior operator. Proof. From Proposition 8.(4) and Proposition 9.(3) we have that cl(x) = [x]k . But k is the kernel of and the kernel of a function is a congruence. It follows by induction that cl(X) ∪ cl(Y ) = [X]k ∪ [Y ]k = [X ∪ Y ]k = cl(X ∪ Y ). Hence cl is additive. Since int is dual of cl we immediately obtain that int is multiplicative. We now list some results in terms of IQRSs. Since i-quantum relations are preorders, it is useful to prove some general facts about this kind of relations:

Formal Topology and Information Systems

283

Proposition 10. Let O = X, R be any preordered set. Then for any x, y ∈ X the following are equivalent: (1) y ∈ R (x); (2) R (y) ⊆ R (x); (3) x ∈ QO y ; (4) x ∈ R(y); (5) O y ∈ Qx . Proof. (1 2) y ∈ R(x) iﬀ x, y ∈ R. Suppose y, y ∈ R. Since R is transitive, for all y ∈ X, y, y ∈ R =⇒ x, y ∈ R so that R (x) ⊇ R (y). Conversely, since R is reﬂexive, y ∈ R(y) holds. Thus if R (x) ⊇ R (y) then y ∈ R (x). All the other equivalences are obvious consequences or even just deﬁnitions. Corollary 5. Let S be an A-system or a P-system over a set G. Then, g ∈ QS iﬀ g ∈ Qg g iﬀ g ∈ RS (g) iﬀ g ∈ Qg g ∈ Q−S g . Q(S)

Q(Q(S))

iﬀ g ∈ RS (g ) iﬀ

Proof. The ﬁrst equivalence is just a deﬁnition. Now, g ∈ RS (g) iﬀ RS (g ) ⊆ Q(S) RS (g) iﬀ g ∈ Qg iﬀ g ∈ RQ(S) (g ) iﬀ RQ(S) (g) ⊆ RQ(S) (g ) iﬀ g ∈ Qg . From this we have that Qg = RS (g) so that in view of trivial set-theoretic considerations, (X ⊆ Y iﬀ −Y ⊆ −X) we obtain the last two equivalences. Q(Q(S))

Q(S)

These equivalences show that IQRSs of level higher than 1 do not provide any further information. Corollary 6. If S is an A-system, an FP-system or a DP-system over a set Q(S) G, then for all g, g ∈ G, X ⊆ G, (a) g ∈ QS ; (b) RS (X) = g iﬀ g ∈ Qg RQ(S) (X). Moreover, since a P-system is a generic relational system we have that all facts valid for P-systems are valid for any relational system. The notion of a quantum of information is asymmetric for P-systems, because if g fulﬁlls strictly more properties than g, we have g ∈ Qg but g ∈ Qg . On the contrary it is symmetric in the case of A-systems and dichotomic or functional P-systems.

8

Higher Level Operators

Let S be an A-system and let Q(S) = G, G, RS be its induced IQRS. What kinds of patterns of data can we collect by applying our operators to these derivative systems? First of all, since in IQRSs there is no longer the distinction between objects and properties and intension or extension, it is better we change once more our symbols and notation:

284

P. Pagliani and M.K. Chakraborty

The operator deﬁned as turns into i i(X) = {g : ∃g (g ∈ X & g , g ∈ RS )} RS e e(X) = {g : RS (g) ∩ X = ∅} RS [i] [i](X) = {g : RS (g) ⊆ X} [RS ] [e] [e](X) = {g : RS (g) ⊆ X} [RS ] Let us call the above operators decorated with RS ”quantum operators” (notice that in this context [[RS ]] and [[RS ]] are not quantum operators). Quantum operators behave in a very particular way, because, we remind, they fulﬁll adjoint properties. Namely RS [RS ] and RS [RS ]. Actually, the following results apply to any preorder. Proposition 11. Let Q(S) = G, G, RS be a IQRS. Let Oi and Oj be any two adjoint quantum operators from the set {RS , RS , [RS ], [RS ]}. Then (a) Oi Oj = Oj ; (b) Oj Oj = Oj ; (c) the ﬁxpoints of Oi and Oj coincide. Proof. (a) (i) In view of Proposition 5, for all g ∈ G and X ⊆ G, g ∈ [RS ]RS (X) iﬀ RS (g) ⊆ RS (X), iﬀ (from Proposition 10) g ∈ RS (X). One can prove g ∈ [RS ]RS (X) iﬀ g ∈ RS (X) similarly. (ii) In view again of Proposition 5, a ∈ RS [RS ](X) iﬀ RS (a) ∩ [RS ](X) = ∅, iﬀ there is a such that a ∈ RS (a ) and a ∈ [RS ](X). But a ∈ [RS ](X) iﬀ RS (a ) ⊆ X iﬀ RS (a) ⊆ X. Hence, a ∈ RS [RS ](X) iﬀ a ∈ [RS ](X). (b) From point (a) and Proposition 6.3, Oj Oj = Oi Oj Oi Oj = Oi Oj = Oj . (c) Let X = Oj (X). Then from point (a) Oi (X) = Oi Oj (X) = Oj (X) = X. Deﬁnition 13. Let S be an A-system or a P-system. With ΩQ (S) we shall denote the family {QS X : X ⊆ G}. With Qn (S) we denote the n-nested application of the functor Q to S. In view of these results we can prove a number of properties. Lemma 5. For any P-system P, (a) Ωint (P) ⊆ ΩQ (P); (b) Γcl (P) ⊆ Ωint (Q(P)). Proof. (a) Assume X QX . Thus we must have some x such that x ∈ / X but x ∈ QX . Thus there is g ∈ X such that (x) ⊇ (g), so that for all / intP (X) m ∈ M such that g m surely (m) X. It follows that g ∈ P P and, hence, int (X) = X. (b) We remind that x ∈ cl (X) iﬀ i(x) ⊆ i(X) iﬀ (x) ⊆ (X). Moreover, if x ∈ X, i(x) ⊆ i(X). Now suppose X = intQ(P) (X). Then there is x ∈ X such that RP (x) X. Hence {y : x ∈ QP y} X. Thus {y : i(y) ⊆ i(x)} X. This means that there is g ∈ / X such that i(g) ⊆ i(x) ⊆ i(X) so that i(g) ⊆ i(x) ⊆ i(X). It follows that X clP (X). Corollary 7. Let Q(S) be an IQRS. Then, Q(S)

(a) QS (...) = RS = cl; (b) Q(...) = RS = A; (c) [RS ] = int; (d) [RS ] = C.

where the operators cl, int, C, A are intended over Q(S).

Formal Topology and Information Systems

285

Corollary 8. Let S be an A-system or a P-system and A, B ⊆ G. Then, Q(S)

1. QS (...) , Q(...) , RS and RS are topological closure operators and their images are closed under intersections. 2. [RS ] and [RS ] are topological interior operators and their images are closed under unions. Q(S)

Proof. (1) From Corollary 7, we have that QS (...) , Q(...) , RS and RS are closure operators. Moreover, since they are lower adjoints in the category ℘(G), ⊆ they preserve unions. Finally, from Proposition 7.2 they are normal and their images are closed under intersections. (2) From Corollary 7, [RS ] and [RS ] are interior operators. Moreover as they are lower adjoints in the category ℘(G), ⊆ they preserve intersections. Finally from Proposition 7.2 they are conormal and their images are closed under unions.

Corollary 9 (I-quantum systems). Let S be an A-system or a P-system. Then, 1. SatQ (S) = ΩQ (S), ∪, ∩, ∅, G is a distributive lattice, called the I-quantum system - IQS induced by S. 2. SatQ (Q(S)) = ΩQ (Q(S)), ∪, ∩, ∅, G is a distributive lattice, called the coI-quantum system - co-IQS induced by S. 3. The set theoretical complement is an antisomorphism between SatQ (S) and SatQ (Q(S)). 4. Satint (Q(S)), Satcl (Q(S)), SatA (Q(S)) and SatC (Q(S)), equipped with the set-theoretical operations, are distributive lattices. 5. G, ΩQ (S) and G, ΩQ (Q(S)) are topological spaces, where the interior operators are intQ2 (S) and, respectively, intQ1 (S) . Proof. (1) We know that the operator Q(...) is additive. Thus ΩQ (S) is closed under unions. From Corollary 8 it is closed under intersections too. Moreover, for ΩQ (S) is a (ﬁnite) lattice of sets Sat(S) inherits distributivity from the corresponding property of unions and intersections. (2) Since Q(S) is a P-system the above considerations apply to this structure. (3) From Corollary 5 we know that −RS = RS , so that we obtain immediately the thesis. (4) From Proposition 7.2 and Corollary 8. (5) Any family of open sets of a topological space enjoys distributivity of arbitrary unions over ﬁnite intersections and of intersection over arbitrary unions. Moreover, from Proposition 11 and Corollary 7 we obtain that, ΩQ (S) = Γcl (Q(S)) = Ωint (Q2 (S)) and ΩQ (Q(S)) = ΩA (Q(S)) = Ωint (Q(S)). Now by means of the above mathematical machinery we prove a key statement in the theory of Approximation Space and Rough Sets, namely the well-known fact that the family of deﬁnable sets can be made into a Boolean algebra. Proposition 12 (Quantum relations and Boolean algebras). Let S be an Information system. If RS is an equivalence relation, then SatQ (S) is a Boolean algebra.

286

P. Pagliani and M.K. Chakraborty

Proof. We show that if RS is an equivalence relation then any element QX of ΩQ (S) is complemented by QX = z∈QX Qz . First, let us prove that QX ∪QX = G. In fact for all g ∈ G if g ∈ QX then g ∈ QX because g ∈ Qg (q-reﬂexivity). Now we prove that QX ∩ QX = ∅. Assume z ∈ QX . We have just to prove that if z ∈ Qz then z ∈ QX . So let z ∈ Qz . For q-symmetry z ∈ Qz . So, if there is an x ∈ X such that z ∈ Qx we have z ∈ Qx too (for q-transitivity), hence z ∈ QX . Contradiction.19 Corollary 10. Let S be an Information system. Then, if S is an A-system, a dichotomic or a functional P-system, then SatQ (S) is a Boolean algebra. About the family of co-prime elements of SatQ (S) we have: Lemma 6. Let S be an A-system or a P-system. Then for any X ∈ J (SatQ (S)), X = Qg for some g ∈ X. Proof. Trivial from the very additive deﬁnition of the operator Q and its increasing property. Lemma 7. Let P be a P-system and g ∈ G. Then Qg = {e(m) : m ∈ i(g)} Proof. Indeed, x ∈ Qg iﬀ i(x) ⊇ i(g) iﬀ x ∈ e(m), ∀m ∈ i(g). Proposition 13. Let P be a P-system such that cl (int) is topological. Then SatQ (P) = Satint (P). Proof. We have seen in Lemma 5 that Ωint (P) ⊆ ΩQ (P). Now we need just to show that if X ∈ J (SatQ (P)) then X = int(X). The proof is immediate. Indeed, the family {e(m) : e(m) ⊆ X} is a base of Ωint (S). Moreover, if int is topological then it is multiplicative and since for all m ∈ M , int(e(m)) = e(m) (from Lemma 3), in view of the above Lemma 7 we have the result. Corollary 11. Let F be an FP-system. Then ΩQ (Qn (F)) = Ωint (Qn (F)), n ≥ 0. Proof. From Corollary 4 and an inductive extension of Proposition 13. Hence we can note that P-systems such that int and cl are topological behave like functional systems. Corollary 12. If S is a preordered set (that is, G = M and R ⊆ G × G is a preorder), then ΩQ (S) = Ωint (S). 19

There is another way to obtain this result. In fact, J. L. Bell proved that if T = A, T is a relational structure with T a tolerance relation (that is, reﬂexive and symmetric) then the family ΩQL (T) of all unions of principal order ﬁlters ↑T x (i. e. T (x) i. e., for symmetry of T , T (x)) can be made into an ortholattice. But if RS is an equivalence relation then it is a tolerance relation too and for any x ∈ A, ↑RS x =↓RS x = cl(x) = Qx so that SatQ (S) can be made into a distributive ortholattice, that is, a Boolean algebra.

Formal Topology and Information Systems

287

Proof. From Corollary 11 and Corollary 9 (3). At this point we can end this subsection with an analogue of the duality between distributive lattices and preorders in the context of i-quantum relations and Psystems. Proposition 14 (Duality between preorders and P-systems) 1. Let O = G, R be a preorder, then there is a P-system I(O) over G such that RI(O) = R (hence, Q(I(O)) ∼ =I O). 2. Let S be an A-system or a P-system. Then I(Q(S)) ∼ =I S. Proof. (1) Let F (O) be the set of order ﬁlters of O. Thus F (O) = ΩQ (Q(O)) (i. e. ΩQ (O ))), so that we know that F(O) can be made into the distributive lattice SatQ (Q(O)). Then let J (SatQ (Q(O))) be the set of co-prime elements of SatQ (Q(O)). Notice that co-prime elements have the form ↑R x, i. e. R (x), for some element x ∈ G and that they may be understood as properties fulﬁlled by the elements of G such that g x only if x, g ∈ R. Let us then deﬁne I(O) I(O) as G, J (SatQ (Q(O))), . Thus, g, g ∈ RI(O) iﬀ g ∈ Qg , iﬀ i(g) ⊆ i(g ), iﬀ g ∈ R (x) =⇒ g ∈ R (x) for all R (x) ∈ J (SatQ (Q(O))). In particular, since R is reﬂexive, g ∈ R (g) so that g ∈ R (g) holds, i. e. g, g ∈ R. Conversely, if g, g ∈ R and x, g ∈ R, for transitivity x, g ∈ R too. It follows that g ∈ R (x) g ∈ R (x), all x ∈ G. (2) For Q(S) is a preorder, from the previous point we have Q(I(Q(S))) ∼ =I Q(S) so that trivially I(Q(S)) ∼ =I S. EXAMPLE 3 Consider the P-system P and the A-system A of Example 2. Here below we display the lattices SatQ (P) and SatQ (A): SatQ (P)

SatQ (A)

G , ll G ,, l " ll " {a1 , a2 , a3 } {a, a1 , a2 } " l " ll , ll {a, a1 , a2 } {a, a2 , a3 } {a1 , a2 } , l , l " l ll "" " {a1 , a2 } {a, a2 } " " ll l " " ll , ll {a, a2 } {a1 } {a3 } l ,, l " l " {a2 } {a} ll " " ll , ∅ l ,, ∅

It is easy to verify that both of them are distributive lattices and that, moreover, SatQ (A) is a Boolean algebra. Indeed, for instance, the element QA {a,a1 } = A A Q = Q = {a {a, a1 , a2 } is complemented by the element z∈Q 3 }. z / A {a3 } {a,a1 }

288

P. Pagliani and M.K. Chakraborty

We can straightforwardly verify that Satcl (Q(P)) is Satint (P) plus some missed elements which are the diﬀerence between SatQ (P) and Satint (P). Indeed, the missed element is {a1 , a2 } which equals clQ(P) ({a1 , a2 }). On the contrary, intP ({a1 , a2 }) = ∅.

9

Generalised Topological Approximation Operators

In view of the duality between Information Systems and preorders, we can develop the rest of the theory from a more abstract point of view. Thus, from now on we shall deal with preordered structures and assume, intuitively, that they represent some information quantum relation system. Corollary 13. Let O = G, G, R be a preordered set. Let X ⊆ G. Then: The application is the least ﬁxpoint of including X R [R ] A int R(X) R [R] cl C R (X) The application is the largest ﬁxpoint of included in X [R] R C cl [R](X) [R ] R int A [R ](X) Proof. First notice that from Proposition 11 the listed ﬁxpoints collapse. Nonetheless it is worthwhile proving the ﬁrst two cases by means of two diﬀerent approaches. (a) Obviously R(X) ⊇ X and for idempotence R(X) is a ﬁxed point of R. Suppose Z is a ﬁxed point of R such that X ⊆ Z. From monotonicity R(X) ⊆ R(Z) = Z. Hence R(X) is the least ﬁxpoint of R including X. (b) From Proposition 6.3 [R]R (X) is the smallest image of [R] greater than or equal to X. Since [R] is idempotent it is the least ﬁxpoint of [R] which includes X and from Proposition 11, it is also the least ﬁxpoint of R (X) including X. The remaining cases are proved analogously. Corollary 14. Let O = G, G, R be a preordered set. Then for all X ⊆ G, (i) R(X) = {Z : Z ∈ ΩA (O) & Z ⊇ X}; (ii) [R](X) = {Z : Z ∈ ΓC (O) & Z ⊆ X}; (iii) R (X) = {Z : Z ∈ Γcl (O) & Z ⊇ X}; (iv) [R ](X) = {Z : Z ∈ Ωint (O) & Z ⊆ X}. Henceforth, for obvious reasons we shall adopt the following terminology: R(X) direct upper R-approximation of X, also denoted with (uR)(X) R (X) inverse upper R-approximation of X, also denoted with (uR )(X) [R](X) direct lower R-approximation of X, also denoted with (lR)(X) [R ](X) inverse lower R-approximation of X, also denoted with (lR )(X)

Formal Topology and Information Systems

289

The information-oriented reading of the above operators is: R(X)

Set of the elements specialised by some member of X (or, which approximate some member of X) R (X) Set of the elements approximated by some member of X (or, which specialise some member of X) [R](X) Set of the elements specialised only by members of X (or, which approximate only members of X) [R ](X) Set of elements approximated only by members of X (or, which specialise only elements of X) Particularly we can give an information-oriented interpretation to some combinations of operators: Set of the elements which are specialised just by elements specialised by some member of X (x ∈ [R]R(X) only if each element which specialises x is specialised by some member of X) Set of the elements which are approximated just by elements [R ]R (X) approximated by some member of X (x ∈ [R ]R (X) only if each element which approximates x is approximated by some member of X )

[R]R(X)

Besides these operators we add also the interpretation of [[R]] and [[R ]]: [[R]](X)

Set of the elements specialised by all the members of X (or, which approximate all the members of X) [[R ]](X) Set of the elements approximated by all the members of X (or, which specialise all the members of X) 9.1

Topological Approximation Spaces

Eventually we deﬁne some interesting examples of topological Approximation Spaces. Deﬁnition 14. Let Q(S) = G, G, RS be an IQRS. Then, 1. 2. 3. 4.

G, [RS ], RS - will be called a Direct Intuitionistic Approximation Space. G, [RS ], RS - will be called an Inverse Intuitionistic Approximation Space. G, [RS ], RS - will be called a Galois Intuitionistic Approximation Space. G, [RS ], RS - will be called a co-Galois Intuitionistic Approximation Space.

Deﬁnition 15. Let G, G, E be a relational structure such that E is an equivalence relation. Let I and C be the interior and, respectively, topological operators of the topological space induced by taking {[x]E : x ∈ G} as a subbasis. Then G, I, C is called a Pawlak Approximation Space From the above discussion the following statement is obvious: Proposition 15. Let E = G, G, E be a relational structure such that E is an equivalence relation. Then G, intE , clE is a Pawlak Approximation Space.

290

P. Pagliani and M.K. Chakraborty

But we can prove a further fact. To this end we introduce the notion of an Approximation Equivalence, or a-equivalence between (pre) topological Approximation Spaces: Deﬁnition 16. Let A = G, α, β and A = G, γ, δ two topological or pretopological Approximation Spaces. Then we say that A and A are a-equivalent, in symbols, A ∼ =a A if and only if Ωα (G) = Ωγ (G) and Γβ (G) = Γδ (G). Clearly, by duality one equality implies the other. We use this deﬁnition in the following statement: Proposition 16. Let S be an A-system or FP-system or DP-system. Let us set ♦ = RS and = [RS ]. Then ♦ = RS , = [RS ] and G, , ♦ is a Pawlak Approximation Space. Moreover, if S is an FP-system then G, , ♦ ∼ =a G, intS , clS . Proof. Immediate, from the fact, proved in Proposition 9.(3), that RS in this case is an equivalence relation. For the last part it is suﬃcient to use in addition Proposition 4.(3) together with Proposition 7, or the latter Proposition and Proposition 9.(3) which together with Proposition 8.(4) states that clS (g) = [g]kf , any g ∈ G. Note that the former system of the previous Proposition is, actually, G, G, , ♦ while the latter is G, M, intS , clS . Therefore, we cannot put G, , ♦ = G, intS , clS . However, if S is an FP-system, then Single-agent (pre)topological Approximation Spaces, and Pawlak Approximation Spaces induce the same family of ﬁxed points.

10

Comparing Information Systems

The notion of an i-quantum makes it possible to compare Information Systems. First of all we should ask whether it is possible to compare two quanta of information Qg and Qg . At ﬁrst sight we would say that Qg is ﬁner than Qg if Qg ⊆ Qg . However, this intuition works for P-systems, but not for A-systems because from Proposition 9.(2) if Qg ⊆ Qg then Qg ⊆ Qg . Thus non trivial comparisons of quanta of information in A-systems require a specialised notion of an i-quantum, which, in any case, is useful for P-systems too. Deﬁnition 17 (Relativised quanta of information) – Let A be an A-system. The quantum of information of g relative to a subset A ⊆ At is deﬁned as: Qg A = {g ∈ G : ∀a ∈ A, ∀x ∈ Va ((a(g) = x) =⇒ (a(g ) = x))}. – Let P be a P-system. The quantum of information of g relative to a subset A ⊆ M is deﬁned as: Qg A = {g ∈ G : ∀a ∈ A(g a =⇒ g a)}. Deﬁnition 18 (I-quantum dependence) Let S be an A-system or a P-system. Let A, A ⊆ At (or A, A ⊆ M ), g ∈ G.

Formal Topology and Information Systems

291

1. We say that A functionally depends on A at g, in symbols A →g A , if for all g ∈ G, g ∈ Qg A =⇒ g ∈ Qg A (that is, if Qg A ⊆ Qg A ). 2. We say that A functionally depends on A, in symbols A → A , if for all g ∈ G, A →g A . 3. If A → A and A → A, we say that A and A are informationally equivalent, A∼ =I A (thus, A ∼ =I A if for all g ∈ G, Qg A = Qg A ). So, a set of attributes (properties) A functionally depends on a set of attributes (properties) A if A has a higher discriminatory capability than A . Clearly, if S is an A-system then the notion of an i-quantum dependence relation turns into the usual notion of a functional dependence. From now on, if X denotes the relation with co-domain restricted to X then with S X we shall denote the subsystem G, X, X. If S is an A-system and X ⊆ At, with S X we shall denote the subsystem G, X, {Va }a∈X . The following statement formalises the above intuitions with respect to iquantum relations: Proposition 17. Let S be an A-system or a P-system. Let A, A ⊆ At (A, A ⊆ M ) such that A → A . Then R(AA) ⊆ R(AA ) . Proof. The proof is immediate. Suppose A → A . Then for all g ∈ G, Qg A ⊆ Qg A , so that g, g ∈ R(AA) implies g, g ∈ R(AA ) . It follows that we can naturally extend the notion of a functional dependence in order to compare two sets X and X of properties or attributes from two distinct (property or attribute) systems S and S over the same set of points G. Thus, we can extend the notion of ”informational equivalence” to entire systems: Deﬁnition 19. Let S and S be A-systems or P-systems over the same set of points G. Let S and S be the sets of attributes (properties) of S and, respectively, S . We say that S and S are informationally equivalent, in symbols S ∼ =I S , if S and only if for any g ∈ G, Qg S = Qg S , that is, if and only if Qg = QS g . Informational equivalence tells something about the behaviour of cl and int: Proposition 18. Let P and P be P-systems and P ∼ G, =I P . Then for all x ∈ P P P P cl (x) = cl (x). If both cl and cl are topological, then clP (X) = clP (X) and intP (X) = intP (X), for any X ⊆ G.

Proof. Suppose clP (x) = clP (x). Then there is g ∈ G such that, say, g ∈ clP (x) and g ∈ / clP (x). It follows that (g) ⊆ (x) but (g) (x). Thus x ∈ QP (g) and x ∈ / QP (g), so that P ∼ operators =I P . If both closure are additive, then by easy induction we obtain that clP (X) = clP (X) for any X ⊆ G. Moreover, suppose intP (X) = intP (X). Then −intP (X) = −intP (X), so that clP (−X) = clP (−X) - contradiction.

Notice that if either clP or clP is not topological then the equality between clP and clP is guaranteed just for singletons so that clP (−X) = clP (−X) is not

292

P. Pagliani and M.K. Chakraborty

a contradiction. Moreover, we can have P and P such that intP (x) = intP (x) but still P ∼ =I P . Therefore, the relation ∼ =I is far to be considered the ”best” way to compare P-systems, though very useful for our purposes. Now we want to stress the fact that we can compare not only the informational behaviour of the same point g with respect two diﬀerent sets of properties (attributes) X and X , but we can also compare the behaviours of two diﬀerent points g and g with respect to the same set of properties (attributes) P . Deﬁnition 20. Let S be an A-system or a P-system, X ⊆ M (or X ⊆ At) and g, g ∈ G. 1. We say that g is an X-specialisation of g (or that g is an X-approximation of g), in symbols g X g, if and only if the following condition holds: ∀x ∈ G(g ∈ Qx X =⇒ g ∈ Qx X). 2. We say that g is a specialisation of g , g g, if and only if g M g. Since for q-reﬂexivity x ∈ Qx , any x ∈ G, if g X g then g ∈ Qg X, so that g X g says that g fulﬁlls at least all the properties from X that are fulﬁlled by g . Therefore, g g implies g , g ∈ RS . Conversely, if g , g ∈ RS then g ∈ Qg . Hence g ∈ Qx implies g ∈ Qx , any x ∈ G, from transitivity of RS . It follows that the two relations and RS coincide. In fact they are the same instance of the usual topological notion of a specialisation preorder. In view of Proposition 9.(1) we can construct a topological space G, ImQ on G whose specialisation preorder is indeed (that is, RS ).

11

Transforming Perception Systems

Now we are equipped with a suﬃcient machinery in order to compare transformed systems. Let A be an A-system. To get a P-system out of A, the basic step derives from the observation that any attribute a is actually a set of properties, namely the possible attribute values for a. Thus we start associating each attribute a with the family N (a) = {av }v∈Va . We set N (At) = a∈At N (a). For each value v, av is the property ”taking value v for attribute a”. This transform is usually called a ”scale nominalisation”. Now let us set a relation N as: g N av if and only if a(g) = v, all g ∈ G, a ∈ At, v ∈ Va . We call the resulting system, N (A) = G, N (At), N , the ”nominalisation of A”. N (A) will be called a nominal A-system or NA-system. Proposition 19. Let A be an A-system. Then: (a) N (A) is a P-system; (b) N (A) ∼ =I A. Proof. (a) is obvious. (b) Let us prove that for any g ∈ G, Qg At = Qg N (At). Indeed, if g ∈ Qg At, then a(g) = x if and only if a(g ) = x, all a ∈ At.

Formal Topology and Information Systems

293

Therefore for any x ∈ N (a), g ax if and only if g ax , whence g ∈ Qg N (a). Finally, g ax for any other x = x, so we have the reverse implication. Moreover, if we formally consider P-systems as binary A-systems, we can also nominalise P-systems. But in this case we have a further property: Proposition 20. Let P be a P-system. Then N (P) is a dichotomic system. Proof. This is obvious, because for any property p, the nominalisation N (p) = {p1 , p0 } forms a pair of complementary properties, since for all g ∈ G, g N p1 if and only if g p and g N p0 if and only if g p. Nominalisation of dichotomic or functional systems does not give rise to any further result. Proposition 21. If P is a DP system or an FP system, then N (P) ∼ =I P. Proof. If P is dichotomic let p, p be a pair of complementary properties. After nominalisation we shall obtain two pairs N (p) = {p1 , p0 } and N (p) = {p1 , p0 }. Clearly, for any g ∈ G, g p in P if and only if g N p1 in N (P). But g N p1 if and only if g N p0 if and only if g N p0 . Conversely, g p if and only if g N p0 if and only if g N p1 if and only if g N p1 . If P is functional and g ∈ Qg M then g m if and only if g m, since (g) = (g ) = m. Thus the proof runs as in Proposition 19.(b). For N (A) is not only a P-system but it is still an A-system with At = {0, 1}, we obtain the following corollary: Corollary 15. Let S be an A-system or a P-system. Then N (S) ∼ =I N (N (S)). Proof. If S is a P-system then N (S) is a dichotomic systems so that from Proposition 21 N (N (S)) ∼ =I N (S). If S is an A-system then N (S) is a binary A-system and from Proposition 19.(b) N (S) ∼ =I N (N (S)). Corollary 16. If A is an A-system then there is a dichotomic system D such that D ∼ =I A. Proof. Since N (A) is a P-system, from Proposition 20 N (N (A)) is dichotomic. But from Proposition 19.(b) and Corollary 15 A ∼ =I N (A) ∼ =I N (N (A)). As a side result we again obtain Proposition 9.(2). Notice that this Proposition, as well as Corollary 15, relies on the fact that we are dealing with deterministic A-systems so that either two objects converge on the same attribute-value, or they diverge, but not both. EXAMPLE 4 Here are some examples: a P-system P = G, M, , an FP-system F = G, M , fˆ and an A-system A = G, At, V over the same set G:

294

P. Pagliani and M.K. Chakraborty

a a a a

b b 11 01 01 00

b 0 0 1 0

b 0 1 1 1

fˆ a a a a

m m 1 0 0 1 1 0 0 0

m 0 0 0 1

a a a a

A A 1 b 0 c 1 b 3 f

A α α α δ

Considering the system P let A = {b, b } and B = {b , b }. Then Qa A = {a, a , a } while Qa B = {a }. It follows that B →a A. On the contrary, Qa A = {a, a , a } and Qa B = {a , a , a } are not comparable. Hence B → A does not hold. If we compare the above systems we notice what follows: P a) A ∼ I P because QA = a = {a, a } while Qa = {a}. Neither P → A because P A F ∼ Qa = {a , a } while Qa = {a }. b) F =I A, because for all g ∈ G, QA g = Qg . Let us now nominalise the systems A and P: N A a a a a

A0 0 1 0 0

A1 1 0 1 0

A3 0 0 0 1

Ab 1 0 1 0

Ad 0 1 0 0

Af 0 0 0 1

Aα 1 1 1 0

Aδ 0 0 0 1

N P a a a a

b1 1 0 0 0

b0 0 1 1 1

b1 1 1 1 0

b0 0 0 0 1

b1 0 0 1 0

b0 1 1 0 1

b 1 0 1 1 1

b 0 1 0 0 0

Thus N (A) = {A0 , A1 , A3 }, N (b) = {b1 , b0 } and so on. It is evident that, for N (A) N (A) instance, a ∈ Qa and a ∈ Qa . But the same happens already in A. N (P) A A = {a }. Indeed, Qa = {a, a } = Qa . On the contrary, QP a = {a , a } but Qa P In fact a ∈ Qa because it fulﬁlls all the properties fulﬁlled by a (i. e. b and b ) plus the additional property b . But in N (P) this latter fact prevents N (P) a from belonging to Qa , because property b splits into the pair b0 , b1 N and a N P b0 while a P b1 , what are mutually exclusive possibilities. If we further nominalise N (P) and split, for instance, b0 , b1 into b01 , b00 , b11 , b10 , it is obvious that the pairs b01 , b10 and b00 , b11 give the same information as b0 and, respectively, b1 . It is not diﬃcult to verify that RN (A) = RA so that N (A) ∼ =I Q(A). 11.1

Dichotomic, Functional and Nominal Systems

First notice that the reverse of Proposition 9.(2) does not hold. For instance, if P is such that G = {1, 2, 3, 4}, M = {A, B, C} and (1) = {A, B}, (2) = {A, B}, (3) = {B, C} and (4) = {B, C}, Qg is an equivalence class, any g ∈ G though P is neither dichotomic nor functional. Also, if A is an A-system, then N (A) is not necessarily dichotomic. However N (A) ∼ =I N (N (A)) which is dichotomic (see Corollary 15). Indeed, notice that N (N (A)) is informationally equivalent to the system deﬁned as follows: 1) For each av in N (A), if Va is not a singleton set ¬av = {av }v =v,v ∈Va , while if Va = {v} then set ¬av = {av }. We set P = {av }v∈Va ∪ {¬av }v∈Va . 2) For each g ∈ G set g ∗ ¬av if and only if g av and g ∗ av if and only if g av . Clearly ¬av is the complementary copy of av . Thus, 3) set S = G, P, ∗ . We can easily verify that S is a dichotomic system and that S ∼ =I N (A).

Formal Topology and Information Systems

295

In reversal, since for any P-system P, N (P) induces an equivalence relation, we can ask whether N (P) itself ”is”, in some form, an A-system. Indeed it is trivially an A-system with set of attributes values V = {0, 1} and such that m1 (g) = 1 iﬀ g m1 iﬀ g m0 iﬀ m0 (g) = 0 and m1 (g) = 0 iﬀ g m0 iﬀ g m1 iﬀ m0 (g) = 1, all m ∈ M and by trivial inspection one can verify that G, N (M ), {0, 1} ∼ =I N (P). Finally we discuss another natural equivalence. We know that if S is an Asystem, or a DP or a FP system then RS is an equivalence relation (see Proposition 9). Thus a question arises as how to deﬁne a functional system F (S) informationally equivalent to a given A or DP system S. The answer is simple. If S is a P-system consider it as an A-system. Any tuple t ∈ a∈At Va is a combination of attribute-values and has the form a1m , . . . , ajn . We setg ∗ t only if a1 (g) = aim for any ai ∈ At and aim ∈ t. The resulting system G, a∈At Va , ∗ is the required F (S). Indeed ∗ is a map because no g ∈ G can satisfy diﬀerent tuples. Thus RF (S) is an equivalence relation such that g, g ∈ RF (S) only if a(g) = a(g ) for all a ∈ At (or in M ). It follows that N (S) ∼ =I F (S) so that if S is dichotomic or it is an A-system then RS = RF (S) and S ∼ =I F (S).

12

Conclusions

We have seen how modal operators naturally arise from the satisfaction relation which links points and properties in a Property System. Combinations of two modal operators which fulﬁll a adjunction relations deﬁne pre-topological interior and closure operators, as studied in Formal Topology. Thus we have shown that approaching approximation problems by means of the mathematical machinery provided by Formal Topology and Galois Adjunction theory makes it possible to deﬁne well-founded generalization of the classical upper and lower approximation operators. Moreover Galois Adjunction theory provides a set of results that can be immediately applied to these operators, so that we have a good understanding of the structure of the system of their ﬁxed points (i. e. exact sets). We have also seen how to deﬁne higher order information systems, namely Information Quantum Relation Systems, from property systems in order to deﬁne topological (that is, continuous) approximation operators, through the notion of a ”quantum of information”. And we have shown when these operators coincide with the lower and upper approximations deﬁned in classical Rough Set Theory. Eventually, we have seen how we can make diﬀerent kinds of information systems, property systems and attribute systems, into a uniform theoretical framework, and control these manipulations by means of a particular notion of an ”informational equivalence” induced by the concept of quanta of information. This has practical consequences too. Indeed, the relational modal or/and topological operators that we have deﬁned over P-systems may be directly translated into extremely simple constructs of functional languages such as LISP or APL (see [16]), thus providing a sound implementation. Therefore, this approach directly links the logical interpretation of approximation operators to the manipulation of concrete data structures for it coherently embeds the con-

296

P. Pagliani and M.K. Chakraborty

crete operations on Boolean matrices into a very general logical framework (the same relational interpretation of a modal operator applies to any sort of binary Kripke frame).

References 1. M. Banerjee & M. Chakraborty, ”Rough Sets Through Algebraic Logic”. Fundamenta Informaticae, XXVIII, 1996, pp. 211-221. 2. Banerjee, M., Chakraborty, M. K.: Foundations of Vagueness: a Category-theoretic Approach. In Electronic Notes in Theoretical Comp. Sc., 82 (4), 2003. 3. Barwise; J., Seligman, J.: Information Flow: the Logic of Distributed Systems. Cambridge University Press, Cambridge, 1997. 4. Bell, J. L.: Orthologic, Forcing, and the Manifestation of Attributes” In C. TR. Chong & M. J. Wiks (Eds.): Proc. Southeast Asian Conf. on Logic. North-Holland, 1983, pp. 13-36. 5. Blyth, T. S., Janowitz, M. F.: Residuation Theory. Pergamon Press, 1972. 6. Chakraborty, M. K., Banerjee, M.: Dialogue in Rough Context. Fourth International Conference on Rough Sets and Current Trends in Computing 2004 (RSCTC’2004), June 1-June 5, 2004, Uppsala, Sweden. 7. D¨ untsch, I., Gegida, G.: Modal-style operators in qualitative data analysis. Proc. of the 2002 IEEE Int. Conf. on Data Mining, 2002, pp. 155-162. 8. D¨ untsch, I., Orlowska, E.: Mixing modal and suﬃciency operators. In Bulletin of the Section of Logic, Polish Academy of Sciences, 28, 1999, pp. 99-106. 9. Gierz, G., Hofmann, K. H., Keimel, K., Lawson, J. D., Mislove, M. and Scott, D. S.: A compendium of Continuous Lattices. Springer-Verlag, 1980. 10. Humberstone, I. L.: Inaccessible worlds. In Notre Dame Journal of Formal Logic, 24 (3), 1983, pp. 346-352. 11. Lin, T. Y.: Granular Computing on Binary Relations. I: Data Mining and Neighborhood Systems. II: Rough Set Representation and Belief Functions. In Polkowski L. & Skowron A. (Eds.): Rough Sets in Knowledge Discovery. 1: Methodology and Applications, Physica-Verlag, 1998, pp.107-121 and 122-140. 12. Orlowska, E.: Logic for nondeterministic information. In Studia Logica, XLIV, 1985, pp. 93-102. 13. Pagliani, P.: From Concept Lattices to Approximation spaces: Algebraic Structures of some Spaces of Partial Objects. In Fund. Informaticae, 18 (1), 1993, pp. 1-25. 14. P. Pagliani: A pure logic-algebraic analysis on rough top and rough bottom equalities. In W. P. Ziarko (Ed.): Rough Sets, Fuzzy Sets and Knowledge Discovery, Proc. of the Int. Workshop on Rough Sets and Knowledge Discovery, Banﬀ, October 1993. Springer-Verlag, 1994, pp. 227-236. 15. Pagliani, P.: Rough Set Systems and Logic-algebraic Structures. In E. Orlowska (Ed.): Incomplete Information: Rough Set Analysis, Physica Verlag, 1997, pp. 109-190. 16. Pagliani, P.: Modalizing Relations by means of Relations: a general framework for two basic approaches to Knowledge Discovery in Database. In Proc. of the International Conference on Information Processing and Management of Uncertainty in Knowledge-Based Systems. IPMU 98, July, 6-10, 1998. ”La Sorbonne”, Paris, France, pp. 1175-1182. 17. Pagliani, P.: A practical introduction to the modal relational approach to Approximation Spaces. In A. Skowron (Ed.): Rough Sets in Knowledge Discovery. Physica-Verlag, 1998, pp. 209-232.

Formal Topology and Information Systems

297

18. Pagliani, P.: Concrete neighbourhood systems and formal pretopological spaces (draft). Conference held at the Calcutta Logical Circle Conference on Logic and Artiﬁcial Intelligence. Calcutta October 13-16, 2003. 19. Pagliani, P.: Pretopology and Dynamic Spaces. In Proc. of RSFSGRC’03, Chongqing, R. P. China 2003. Extended version in Fundamenta Informaticae, 59(2-3), 2004, pp. 221-239. 20. Pagliani, P.: Transforming Information Systems. In Proc. of RSFDGrC 2005, Vol. I, pp. 660-670. 21. Pagliani, P., Chakraborty, M. K.: Information Quanta and Approximation Spaces. I: Non-classical approximation operators. In Proc. of the IEEE Int. Conf. on Granular Computing. Beijing, R. P. China 2005, pp. 605-610. 22. Pagliani, P., Chakraborty, M. K.: Information Quanta and Approximation Spaces. II: Generalised approximation operators. In Proc. of the IEEE Int. Conf. on Granular Computing. Beijing, R. P. China 2005, pp. 611-616. 23. Pawlak, Z.: Rough Sets: A Theoretical Approach to Reasoning about Data. Kluwer, 1991. 24. Polkowski, L.: Rough Sets: Mathematical Foundations. Advances in Soft Computing, Physica-Verlag, 2002. 25. Sambin, G.: Intuitionistic formal spaces and their neighbourhood. In Ferro, Bonotto, Valentini and Zanardo (Eds.) Logic Colloquium ’88, Elsevier (NorthHolland), 1989, pp. 261-285. 26. Sambin, G., Gebellato, S.: A Preview of the Basic Picture: A New Perspective on Formal Topology. In TYPES 1998, pp. 194-207. 27. Sambin, G.: Formal topology and domains. In Proc. of the Workshop on Domains, IV. Informatik-Bericht, Nr. 99-01, Universit¨ at GH Siegen, 1999. 28. Smyth, M.: Powerdomains and predicate transformers: a topological view. In J. Diaz (Ed.) Automata, languages and Programming. Springer LNCS, 154, 1983, pp. 662-675. 29. Skowron, A. & Stepaniuk, J.: Tolerance Approximation Spaces. Fundamenta Informaticae, 27 (2-3), IOS Press, 1996, pp. 245-253. 30. Skowron, A., Swiniarski, R. & Synak, P.: Approximation Spaces and Information Granulation. Transactions on Rough Sets III, LNCS 3400, Springer, 2005, pp. 175-189. 31. Skowron, A., Stepaniuk, J., Peters, J. F. & Swiniarski, R.: Calculi of approximation spaces. Fundamenta Informaticae, 72 (1-3), 2006, pp. 363–378. 32. Vakarelov, D.: Information systems, similarity relations and modal logics. In E. Orlowska (Ed.) Incomplete Information - Rough Set Analysis Physica-Verlag, 1997, pp. 492-550. 33. Vickers, S.: Topology via Logic. Cambridge University Press, 1989. 34. Wille, R.: Restructuring Lattice Theory. In I. Rival (Ed.): Ordered Sets, NATO ASI Series 83, Reidel, 1982, pp. 445-470. 35. Yao, Y. Y.: Granular computing using neighborhood systems. In R. Roy, T. Fumhashi, and P.K. Chawdhry (Eds.): Advances in Soft Computing: Engineering Design and Manufacturing, Springer-Verlag, London, U.K., 1999. 36. Yao, Y. Y.: A comparative study of formal concept analysis and rough set theory in data analysis. Manuscript 2004. 37. Zhang, G. -Q.: Chu spaces, concept lattices and domains. In Proc. of the 19th Conf. on the Mathematical Found. of Programming Semantics. March 2003, Montreal, Canada. Electronic Notes in Theor. Comp. Sc., Vol. 83, 2004.

On Conjugate Information Systems: A Proposition on How to Learn Concepts in Humane Sciences by Means of Rough Set Theory Maria Semeniuk–Polkowska Chair of Formal Linguistics, Warsaw University Browarna 8/12, 00991 Warsaw, Poland m polkowski@hotmail.com

To the memory of Professor Zdzislaw Pawlak Abstract. Rough sets, the notion introduced by Zdzislaw Pawlak in early 80’s and developed subsequently by many researchers, have proved their usefulness in many problems of Approximate Reasoning, Data Mining, Decision Making. Inducing knowledge from data tables with data in either symbolic or numeric form, rests on computations of dependencies among groups of attributes, and it is a well–developed part of the rough set theory. Recently, some works have been devoted to problems of concept learning in humane sciences via rough sets. This problem is distinct as to its nature from learning from data, as it does involve a dialogue between the teacher and the pupil in order to explain the meaning of a concept whose meaning is subjective, vague and often initially obscure, through a series of interchanges, corrections of inappropriate choices, explanations of reasons for corrections, ﬁnally reaching a point, where the pupil has mastered enough knowledge of the subject to be able in future to solve related problems fairly satisfactorily. We propose here an approach to the problem of learning concepts in humane sciences based on the notion of a conjugate system; it is a family of information systems, organized by means of certain requirements in order to allow a group of students and a teacher to analyze a common universe ofobjects and to correct faulty choices of attribute value in order to reach a more correct understanding of the concept. Keywords: learning of cognitive concepts, rough sets, information systems, conjugate information systems.

1

Introduction

In addition to a constant ﬂux of research papers on inducing knowledge from data expressed in either symbolic or numerical form, there are recently papers on learning cognitive concepts by means of the rough set theory, see, e.g., [2], [12], [13], [14], [15]. We propose in this work an approach to the problem of learning/teaching of concepts in humane sciences that stems from an analysis of the process of J.F. Peters et al. (Eds.): Transactions on Rough Sets VI, LNCS 4374, pp. 298–307, 2007. c Springer-Verlag Berlin Heidelberg 2007

On Conjugate Information Systems

299

learning in humane sciences, and of learning approach in library sciences, in particular, that has been worked out during our seminars at Warsaw University, [12], [13], [14], [15]. In the process of learning of humane concepts, in general, a teacher, a tutor, is directing the pupil, the student, toward understanding of the problem, and toward its correct solutions, by means of a dialogue that involves expositions, responses, corrections, explanations etc., etc., aimed at developing a satisfactory understanding of the concept meaning by the student. In order to formally render this mechanism and to study this problem, we recall here a notion of a conjugate information system introduced in [14] (under the name of SP–systems), and discussed shortly in [12]. The main idea underlying this approach can be introduced as follows: 1. Both the tutor and the student are equipped with information/decision systems (see Sect. 2, for all relevant notions of the rough set theory) that possess identical sets of attributes and the same universe of objects, and they differ from each other only in value assignment to attributes; the assumption is, that the values of attributes are correctly assigned in the tutor system whereas the student can initially assign those values incorrectly, which results in a faulty classiﬁcation of objects to decision classes, and those values are gradually corrected during the interchange of messages with the tutor; 2. In order to learn a correct assignment of values to attributes, the student has also in his disposal a family of auxiliary decision systems, one for each attribute. Attributes in those decision systems are for simplicity (and, actually, in conformity with the practice of learning in many parts of humane sciences) assumed to be Boolean; this means that the value of an attribute on an object is selected on the basis of whether the object has/has not given Boolean features (for instance, when deciding whether romance books should be acquired for a library, one may look at the feature: majority/minority in a poll opted for romance books in the library). In what follows, we present a formal framework for conjugate information systems. We refrain from discussing here the interface between the tutor and the pupil, being satisﬁed with presenting the formal apparatus of conjugate information systems. Our methodology presented in what follows was tested in our courses given to the students in the Department of Library and Information Sciences at the University of Warsaw. The author does express gratitude to her students to whom she is indebted for many works on applying the ideas presented in this paper.

2

Auxiliary Notions of Rough Set Theory

All basic notions relevant to rough sets may be found in [5], [9], or in [3]. We recall here, for the reader’s convenience, some basic notions that are used in the sequel.

300

M. Semeniuk–Polkowska

2.1

Information Systems

An information system A is deﬁned as a triple (U, A, h) where: 1. U is a ﬁnite set of objects; 2. A is a ﬁnite set of conditional attributes. In the sequel, we will use the term attribute instead of the term conditional attribute. Attributes act on objects, to each pair of the form (a, u), where a is an attribute and u is an object, a value a(u) is assigned. In our setting, we wish to work with systems where U and A are ﬁxed, but value assignments are distinct, hence the need for the component h, representing the value assignment in a given information system; 3. A mapping h : U × A → V , with h(u, a) = a(u), is an A–value assignment, where V = {Va : a ∈ A} is the attribute value set. 2.2

Decision Systems

Decision systems are a variant of information systems in which a new attribute d∈ / A, called the decision, is introduced; formally, a decision system is a quadruple (U, A, d, h) where U, A are as in sect. 2.1, d : U → Vd is the decision with values in the decision value set Vd , and the value assignment h does encompass d, i.e., h : U × (A ∪ {d}) → V ∪ Vd with the obvious proviso that values of h on pairs of the form (d, u) belong in Vd . 2.3

Indiscernibility and Its Extensions

The crucial notion on which the idea of rough sets does hinge is that of the indiscernibility relation [5], [6]. For an information system I = (U, A, h), the indiscernibility relation IN DI (B), induced by a set B ⊆ A of attributes, is deﬁned as follows, IN DI (B) = {(u, v) : h(a, u) = h(a, v) for a ∈ B},

(1)

and equivalence classes of IN DI (B) generate by means of the set–theoretical operation of the union of sets the family of B–exact sets (or, concepts); concepts that are not B–exact are called B–rough. Rough set theory deals with constructs that are invariant with respect to indiscernibility relations hence they can be expressed in terms of indiscernibility classes. Indiscernibility classes may be generalized to μ–granules, where μ is a rough inclusion [11]. Rough inclusions are relations of the form μ(u, v, r) read as: ”u is a part of v to degree at least r”. Examples of rough inclusions and deeper applications can be found, e.g., in [10]; let us quote from there an example of the rough ukasiewicz t–norm tL (x, y) = max{0, x+y−1} by inclusion μL induced from the L ) ≥ r in which DIS(u, v) = {a ∈ means of the formula μL (u, v, r) ⇔ g( |DIS(u,v)| |A| A : h(a, u) = h(a, v)}, |A| stands for the cardinality of A, and g is a function from

On Conjugate Information Systems

301

the reals into [0, 1] that ﬁgures in the representation : tL (x, y) = g(f (x) + f (y) (see,e.g., [7]). As g(x) = 1 − x, the formula for μL is ﬁnally: μL (u, v, r) iﬀ

|IN D(u, v)| ≥ r. |A|

(2)

In case r = 1, one obtains from (2) indiscernibility classes as sets (granules) of the form g(u)1 = {v : μ(u, v, 1)}; for r < 1, one obtains a collection of granules being unions of indiscernibility classes with respect to various sets of attributes. 2.4

Approximations to Rough Concepts

One more crucial notion due to Zdzislaw Pawlak is the notion of an approximation. In the classical case, any set (concept) X ⊆ U is approximated with indiscernibility classes [u]B of the relation IN DI (B), where B ⊆ A, from below (the lower approximation) and from above (the upper approximation): (3) BX = {[u]B : [u]B ⊆ X}, BX =

{[u]B : [u]B ∩ X = ∅}.

(4)

More generally, one can replace in above deﬁnitions classes [u]B with the μ–granules g(u)r of a ﬁxed or subject to some conditions radius r.

3

Conjugate Information Systems

The notion of a conjugate information system reﬂects the mechanism of learning in the interaction between the tutor and the student (or, students). In this process, the correct evaluation scheme is transferred from the tutor to students, who initially may have incorrect evaluation schemes and gradually learn better evaluations in order to ﬁnally come up with schemes satisfactorily close to the correct one. 3.1

On Conjugate Information Systems

By a conjugate information system, we understand a triple, d : a ∈ A, i ∈ I}, i0 ), C = ({Ai = (Ui , Ai , hi ) : i ∈ I}, {Fa,i

(5)

where I is a set of participants in the learning process, with i0 denoting the tutor and i ∈ I \ {i0 } denoting students, consisting of: 1. a family of information systems {Ai = (Ui , Ai , hi ) : i ∈ I} such that for some ﬁnite sets U, A we have Ui = U, Ai = A for i ∈ I; d 2. a family of decision systems {Fa,i : a ∈ A and i ∈ I}. Thus the diﬀerence between information systems Ai , Aj with i = j, i, j ∈ I is in functional assignments hi , hj . The information system corresponding to i0 is said to be the tutor system;

302

M. Semeniuk–Polkowska

d 3. for each pair (a, i), a ∈ A, i ∈ I, the decision system Fa,i is a decision system (U, F eata , a, ha,i ), where U is the C–universe, F eata is the set of a–features, each f ∈ F eata being a binary attribute, a is the decision attribute of the d , and ha,i is the value assignment. system Fa,i d will be regarded as the system belonging to the tutor, The system Fa,i 0 d while its realization by an agent i ∈ I, i = i0 will be the system Fa,i of the student i for the evaluation of the attribute a ∈ A. d An Assumption. We assume that each system Fa,i is reduced in the sense that for each value va of the attribute a, there exists at most one object u with the property that ha,i (u, a) = va . In the case when such an object u exists (and then, by our assumption, it is unique) we will denote by h← a,i (va ) the information vector (f (u) : f ∈ F eata ), i.e.,

h← a,i (va ) = (f (u) : f ∈ F eata ).

(6)

← The symbol (h← a,i (va ))f will denote the f −th coordinate of the vector ha,i (va ). We deliberately omit the communication aspect of the process; formally, its presence could be marked with some mappings between the corresponding systems. However, we deem it unnecessarily complicating the picture for the purpose of this paper.

3.2

A Metric on Conjugate Systems

It is now important to introduce some means for organizing the unordered as of now set of participants in the learning process; to this end, we exploit the idea in [12] of some distance function on a conjugate system. We introduce a distance function dist on the conjugate system C. To this end, we ﬁrst introduce for an object u ∈ U the set, DISi,j (u) = {a ∈ A : hi (a, u) = hj (a, u)},

(7)

of attributes discerning on u between systems Ai and Aj . Thus, DISi,j (u) collects attributes which are assigned distinct values on the object u by students i = j. Now, we let, (8) dist(Ai , Aj ) = maxu |DISi,j (u)|. The function dist(Ai , Aj ) is a pseudo–metric, i.e., it has all properties of a metric (see, e.g., [1]) except for the fact that its value may be 0 whereas the arguments may be formally distinct as discerned by distinct indices i and j; we oﬀer a simple argument showing that dist(., .) is a pseudo–metric. Only the triangle inequality may need a proof. Thus, assume that information systems Ai , Aj , and Ap are given. If hi (a, u) = hj (a, u) and hj (a, u) = hp (a, u) then hi (a, u) = hp (a, u); thus, hi (a, u) = hp (a, u) implies that either hi (a, u) = hj (a, u) or hj (a, u) = hp (a, u).

On Conjugate Information Systems

303

In consequence, DISi,p (u) ⊆ DISi,j (u) ∪ DISj,p (u),

(9)

and from (9) one gets that, maxu |DISi,p (u)| ≤ maxu |DISi,j (u)| + maxu |DISj,p (u)|.

(10)

The formula (10) is the required triangle equality. Learning starts with the pupil(s) closest to the tutor, and continues in the decreasing order of the distance. 3.3

Basic Parameters for Learning

At the learning stage, each agent Aj (represented by the corresponding information system) learns to assign values of attributes in its set A from features in d decision systems {Fa,j : a ∈ A}. Parameters for Learning Feature Values and Attribute Values. First, at the training stage, each agent student learns to assign correct values to features in sets F eata = {fk,a : k = 1, 2, ..., na } for each attribute a ∈ A. We assume that values at the tutor system are already established as correct. The measure of learning quality is the assurance–level–function mj (k, a); for each triple (j, k, a), where j ∈ I \ {i0 }, k ≤ na , and a ∈ A, it is deﬁned as follows: mj (k, a) =

pos(j, k, a) , ex(j, k, a)

(11)

where ex(j, k, a) is the number of examples for learning fk,a and pos(j, k, a) is the number of positively classiﬁed examples in the set U by the agent Aj. The process of learning, as mentioned above, proceeds in a dialogue between the tutor and the student, aimed at explaining the meaning of the attribute a, and its dependence on features in the set F eata ; after that discussion, the j − th student proceeds with assigning values to features for objects from the universe U , in the training sample. The assignment is evaluated by the tutor and on the basis of that evaluation, assurance levels are calculated, to judge the understanding of the pupil. d according to decreasing value of mj (k, a); the reWe order features in Fa,j d sulting linear order is denoted ρaj , and the system Fa,j with values assigned by d the agent Aj is denoted with the symbol Fa,j (ρ). Metrics on Value Sets of Attributes. We set a distance function φja (v, w) on values of the attribute a for each a ∈ A, v, w ∈ Va , and j ∈ I, estimated in d the system Fa,j by letting,

304

M. Semeniuk–Polkowska

where,

φja (v, w) = |DISa,j (v, w)|,

(12)

← DISa,j (v, w) = {f ∈ F eata : (h← a,j (v))f = (ha,j (w))f }.

(13)

d in This deﬁnitions are possible, due to our assumption about systems Fa,j Sect.3.1. Thus, φja (v, w) does express the distance at the pair v, w of values of the attribute a measured as the number of diﬀerently classiﬁed features in the row deﬁned by v, w, respectively.

4

Learning of Attribute Values

We now address the problem of learning from the tutor of the correct evaluation of attribute values. Objects u ∈ U are sent to each agent Ai for i ∈ I one–by–one. Step 1. Assignment of attribute values based on training examples. At that stage the values dist(i, i0 ) of distances from agents Ai to the tutor Ai0 are calculated. Step 2. The feedback information passed from the tutor to the agent Ai is the following: Infi = (r, Error− set− i = {ai1 , ...aipr }, Error− vector− i = [si1 , ..., sipr ]), (14) where: 1. r is the value of the distance dist (Ai0 , Ai ) from the student i to the tutor i0 ; 2. pr is the number of misclassiﬁed attributes in A between agents i, i0 . Clearly, p(r) ≤ |U | · r, depends on r; 3. aij is the j th misclassiﬁed attribute; 4. for j ∈ {1, .., pr }, the value sij is the distance φaij (vj , wj ) where vj is the correct (tutor) value of the attribute aij and wj is the value assigned to aij by the agent Ai . Step 3. The given agent Ai begins with the attribute a = aiError−seti for which the value of the assurance-level-function is maximal (eventually selected at random from attributes with this property). For the attribute a, the value s = sa is given, meaning that s × 100 percent of features has been assigned incorrect values by Ai in the process of determining the value of a. d (ρ) according to deStep 4. Features in F eata are now ordered into a set Fa,i creasing values of the assurance–level–function mi (k, a) i.e. by ρai : starting with the feature f = fFi,aeata giving the minimal value of the function mi (k, a), the agent i goes along the ordered set changing the value at subsequent nodes. If the value of φ remains unchanged after the change at the node, the error counter remains unchanged, otherwise its value is decremented/incremented by one.

On Conjugate Information Systems

305

Step 5. When the error counter reaches the value 0, stop and go to the next feature. Step 6. Go to the next attribute in the set A.

5

An Example

Our example is a simple case that concerns grading essays written by students in French, taken from [13], [12]. Grading is done on the basis of three attributes: a1 : grammar, a2 : structure, and a3 : creativity. We present below tables showing the tutor decision systems Fa for a = a1 , a2 , a3 . Example 1. Decision systems Fa1 , Fa2 , Fa3 Table 1. Decision systems Fa1 , Fa2 , Fa3 fa11 + -

fa21 + -

fa31 + + -

a1 3 2 1

fa12 +

fa22 + -

fa32 + -

a2 3 2 1

fa13 + +

fa23 + +

fa33 +

a3 3 2 1

where fa11 takes value +/- when the percent of declination errors is ≥ /< 20 ; fa21 is +/- when the percent of conjugation errors is ≥ / < 20, and fa31 is +/- when the percent of syntax errors is ≥ / < 20; fa12 takes value +/- when the structure is judged rich/not rich, fa22 is +/- when the structure is judged medium/not medium, and fa32 is +/- when the structure is judged to be weak/ not weak. fa13 takes value +/- when the lexicon is judged rich/not rich, fa23 is +/- when the source usage is judged extensive/not extensive, and fa33 is +/- when the analysis is judged to be deep/ not deep. Consider a pupil A1 and a testing information system with U = {t1 , t2 , t3 }, A = {a1 , a2 , a3 } which is completed with the following value assignments. Example 2. Information systems A0 of the tutor and A1 of the pupil. Table 2. Decision systems Fa1 , Fa2 , Fa3 t t1 t2 t3

a1 1 1 3

a2 2 1 2

a3 1 1 3

t t1 t2 t3

a1 1 1 3

a2 2 1 2

a3 2 2 3

The distance dist(A0 , A1 ) is equal to 1 as DIS0,1 (t1 ) = {a3 } = DIS0,1 (t2 ); DIS0,1 (t3 ) = ∅.

306

M. Semeniuk–Polkowska

Thus, the pupil misclassiﬁed the attribute a3 due to a faulty selection of feature values: in case of t1 , the selection by the tutor is +,+,+ and by the pupil: +,+,-. The distance φa3 ,1 is equal to 1 and the information sent to the pupil in case of t1 is Inf1 = (1, {a3 }, (1)). Assuming the values of assurance–level–function m(1, k, a3 ) are such that f 3,a3 = fa33 , the pupil starts with fa33 and error–counter =1 and changing the value at that node reduces the error to 0. This procedure is repeated with t2 etc.

6

Conclusion

We have presented a skeleton on which the mechanism of learning cognitive concepts can be developed. It has been the principal aim in this paper to show that the notion of a conjugate information system may be helpful in fulﬁlling this task as a model of dependence between the tutor and the student.

Acknowledgement The topic of the paper has been discussed at seminars conducted by the author at the Institute of Library and Information Sciences at Warsaw University. The author wishes to thank the participants in those seminars. Thanks go also to Professor Lech Polkowski for useful discussions and valuable help with the preparation of this note.

References ´ ements de Math´ematique. Topologie G´en´erale. Hermann, Paris, 1. Bourbaki, N.: El´ France (1960). 2. Dubois, V., Quafafou, M.: Concept learning with approximations: rough version spaces. In: Lecture Notes in Artiﬁcial Intelligence vol. 2475, Springer–Verlag, Berlin, Germany. (2002) 239–246. 3. Komorowski, J., Pawlak, Z., Polkowski, L., Skowron, A.: Rough sets: A tutorial. In: Pal, S. K., Skowron, A. (Eds.): Rough–Fuzzy Hybridization: A New Trend in Decision Making. Springer–Verlag, Singapore Pte. Ltd. (1999) 3–98. 4. Pal, S. K., Polkowski, L., Skowron, A. (Eds.): Rough-Neural Computing. Techniques for Computing with Words. Springer–Verlag, Berlin, Germany (2004). 5. Pawlak, Z.: Rough Sets: Theoretical Aspects of Reasoning about Data. Kluwer, Dordrecht, the Netherlands (1991). 6. Pawlak, Z.: Rough sets. International Journal of Computer and Information Science 11 (1982) 341–356. 7. Polkowski, L.: Rough Sets. Mathematical Foundations. Physica–Verlag, Heidelberg, Germany (2002). 8. Polkowski, L., Tsumoto, S., Lin, T. Y. (Eds.): Rough Set Methods and Applications. Physica–Verlag, Heidelberg, Germany (2000). 9. Polkowski, L., Skowron, A. (Eds.): Rough Sets in Knowledge Discovery 1,2. Physica–Verlag, Heidelberg, Germany (1998).

On Conjugate Information Systems

307

10. Polkowski, L., Semeniuk–Polkowska, M.: On rough set logics based on similarity relations. Fundamenta Informaticae 64 (2005) 379–390. 11. Polkowski, L., Skowron, A.: Rough mereology: A new paradigm for approximate reasoning. International Journal of Approximate Reasoning 15 (1997) 333-365. 12. Semeniuk-Polkowska, M., Polkowski, L.: Conjugate information systems: Learning cognitive concepts in rough set theory. In: Lecture Notes in Artiﬁcial Intelligence 2639, Springer–Verlag, Berlin, Germany. (2003) 255–259. 13. Semeniuk–Polkowska, M.: Applications of Rough Set Theory. Seminar Notes (in Polish), Fasc. II, III, IV. Warsaw University Press, Warsaw, Poland (2000–2002). 14. Semeniuk–Polkowska, M.: On Some Applications of Rough Sets in Library Sciences (in Polish). Warsaw University Press, Warsaw, Poland (1997). 15. St¸epie´ n, E.: A study of functional aspects of a public library by rough set techniques.PhD Thesis, Warsaw University, Department of Library and Information Sciences, M. Semeniuk-Polkowska, supervisor (2002).

Discovering Association Rules in Incomplete Transactional Databases Grzegorz Protaziuk and Henryk Rybinski Institute of Computer Science, Warsaw University of Technology gprotazi@ii.pw.edu.pl, hrb@ii.pw.edu.pl

Abstract. The problem of incomplete data in the data mining is well known. In the literature many solutions to deal with missing values in various knowledge discovery tasks were presented and discussed. In the area of association rules the problem was presented mainly in the context of relational data. However, the methods proposed for incomplete relational database can not be easily adapted to incomplete transactional data. In this paper we introduce postulates of a statistically justiﬁed approach to discovering rules from incomplete transactional data and present the new approach to this problem, satisfying the postulates. Keywords: association rules, frequent itemsets, incompleteness, transactional data.

1

Introduction

Very often one of the main restrictions in using data mining methodology is imperfection of data, which is a common fact in real-life databases, especially those exploited for a long period. Imperfection can be divided into several different categories: inconsistency, vagueness, uncertainty, imprecision and incompleteness [19]. In the paper we consider the problem of discovering knowledge from incomplete database. Within the knowledge discovery process the incompleteness of data can be taken into consideration at two stages, namely (1) at the preprocessing step, and (2) at the data mining step. The objective of (1) is to ﬁll missing values in order to pass to the next steps of the process and process data as they were complete. Here, one can use simple approaches, such as replacing unknown values by special ones (e.g. average or dominant value), as well as more advanced methods, such as e.g. completing data methods based on classiﬁers or sets of rules [8]. In the case of (2), missing or unknown values are subject of processing by the data mining algorithms. In the literature many such approaches for diﬀerent data mining tasks were introduced. In particular, the problem of classifying incomplete objects has been addressed in the context the rough set theory [20,21]. The main idea of the approach is based on the indiscernibility relation and lower and upper approximation of a given set X. Originally proposed for complete information system,

Research has been supported by the grant No 3 T11C 002 29 received from Polish Ministry of Education and Science.

J.F. Peters et al. (Eds.): Transactions on Rough Sets VI, LNCS 4374, pp. 308–328, 2007. c Springer-Verlag Berlin Heidelberg 2007

Discovering Association Rules in Incomplete Transactional Databases

309

the approach was successfully extended to deal with incomplete data. Various modiﬁcations have been proposed and discussed in the papers [10,16,28,29,30]. Yet another group of data mining algorithms dealing with incomplete data can be distinguished, the algorithms from this group are based on the methods for building decision tree. The modiﬁcation of the C4.5 algorithm was introduced in [25]. In the CART algorithm [6] the surrogate tests are used for dealing with missing values. Diﬀerent aspects of using decision trees for working with incomplete data have been presented in [9,17,32]. In the paper we concentrate on the algorithms discovering association rules from incomplete data sets. In [12] a notion of legitimate approach has been deﬁned. It consists in satisfying a set of postulates resulting from statistical properties of the support and conﬁdence parameters and being necessary conditions. We claim that any method dealing with incompleteness should satisfy the postulates of the approach in order to properly asses expected support and conﬁdence. The original deﬁnition of the postulates referred to the relational database. Here we generalize it, so that it also covers discovering association rules from transactional data. In addition we deﬁne a novel data mining algorithm (DARIT), very well suited for discovering rules from transactional databases and satisfying the postulates of the statistically justiﬁed approach. The rest of the paper is organized in the following manner. Section 2 formally deﬁnes association rules and their properties. Section 3 reviews the methods of discovering association rules from incomplete data. Section 4 presents the concept of support and conﬁdence under incompleteness. In Section 5 we discuss details of the presented DARIT algorithm. The results of experiments are presented in Section 6, whereas Section 7 concludes this paper.

2

Association Rules and Their Properties

Below we introduce basic notions necessary for analyzing the process of discovering rules from incomplete data. We consider two types of databases, namely transactional and relational ones. A transactional database, denoted as DT , consists of ﬁnite set of transactions, DT = {t1 , t2 , t3 , . . . , tn }. Each transaction has a unique identiﬁer and a non empty set of elements (items). Each element belongs to the ﬁnite set of items I = {elem1, elem2 , . . . , elemm}. A relational database, denoted by DR, is a ﬁnite set of records DR = {r1 , r2 , . . . , rk }. Each record consists of n scalar values, belonging to the domains of n attributes respectively. The set of attributes is further denoted by A. By k-itemset we denote a set of k items from the database. In the sequel, if we do not distinguish between relational and transactional database, we denote it by D. Association rules are one of the simplest and the most comprehensive forms for representing discovered knowledge. The problem of discovering association rules was ﬁrst deﬁned in [1], in the context of market basket data with the goal to identify customers’ buying habits. An exemplary association rule would state that 70% customers who buy bread also buy milk.

310

G. Protaziuk and H. Rybinski

The basic property of an itemset is its support. It is deﬁned as percentage of those transactions in the database D, which contain given itemset. It is referred to as a relative support, and is formally deﬁned as: support(X) =| {t ∈ D | X ⊆ t}|/ | D | .

(1)

where X – itemset, t – transaction or record. Sometimes the notion of absolute support of an itemset is used, which is deﬁned as a number of the transactions supporting given itemset. If a given transaction (record) includes an itemset X we say that the transaction (record) supports the itemset. Frequent itemset is an itemset with support not less than a given minimal level called minimal support and denoted by minSup. Association rule is an expression in the form: X → Y , where X, Y are itemsets over I and X = ∅, Y = ∅ and X ∩ Y = ∅. X is called an antecedent of the rule, Y is called a consequent of the rule. The support of the rule X → Y is equal to support(X ∪ Y ). Conﬁdence of the rule X → Y , denoted by conﬁdence(X → Y ), is deﬁned as: conf idence(X → Y ) = support(X → Y )/support(X).

(2)

The parameter minConf is deﬁned by the user and indicates minimal conﬁdence that the discovered rules should satisfy. The basic task concerning association rules is to ﬁnd all such rules which satisfy the minimal support and minimal conﬁdence requirements.

3

Related Works

The most known algorithm for discovering all association rules is the Apriori algorithm, proposed in [2]. In the algorithm candidate sets (potentially frequent itemsets) of the length k are generated from frequent itemsets of the length k −1. Another approach proposed in [11] is based on the special data structure called FP-tree (frequent pattern tree). Various aspects of discovering association rules are widely presented and discussed in the literature. In [3,15,24,27,31] the problem of generating only interesting rules with respect to additional measure is considered. In [4,5,23] the methods for discovering rules from dense data were introduced. The lossless concise representations of frequent patterns were proposed in [7,14,22]. However, the problem of discovering association rules from incomplete data is discussed relatively rarely, especially with respect to transactional data sets. In the case of missing values the fundamental problem is evaluating the real support of a given itemset, so that one can determine if the given itemset is frequent or not. In [26] the new deﬁnitions of support and conﬁdence of the rule were introduced for the relational databases with missing values. In the deﬁnitions, the notions of disabled data and valid database are used. A record r is disabled for a given set X if, it includes at least one unknown value for the attributes, for which there are elements in X. A set of disabled records for X is denoted by

Discovering Association Rules in Incomplete Transactional Databases

311

Dis(X). A valid database vdb for the set X consists only of those records from DR which are not disabled for X: vdb = DR\Dis(X). Given these notions the authors deﬁne support as: support(X) = |DX|/|vdb(X)|, where DX stands for a set of records containing the set X. The conﬁdence was deﬁned as follows: conﬁdence(X → Y ) = |DXY |/|DX | − |Dis(Y ) ∩ DX |). It is worth mentioning that the proposed deﬁnition of support may give rise to the situations in which a multielement set has greater support than its subsets. This drawback has been eliminated in [13], where the probabilistic approach was presented. The approach is dedicated to relational databases. It is based on the assumption that the missing values of an attribute do not depend on the values of other attributes. Given an attribute a, unknown values in all records are assigned a probability distribution over all known values from the domain of the attribute. Each value v has assigned a probability, denoted by prob(v, a), which is equal to its frequency of occurring in all records having known values for the considered attribute. The main idea of the approach is based of the notion of probable support. Probable support, denoted by probSupr , is calculated for the element elem(v, a) of the value v from the domain of the attribute a for single record in the following manner: 1 if r.a = v probSupr (elem(v, a)) = prob(v, a) if r.a is unknown 0 otherwise

(3)

where r.a stands for the value for the attribute a in the record r. The support probSupr of a set X = {elem(v1 , a1 ), elem(v2 , a2 ), . . . , elem(vk , ak )} is computed by the following formula: probSupr(X) = probSupr (elem(v1 , a1 )) ∗ probSupr (elem(v2 , a2 )) ∗ . . . ∗ (4) probSupr (elem(vk , ak )) A similar approach has been applied for transactional databases in [18] where the algorithm ∼AR has been proposed. The algorithm is a modiﬁcation of the well known Apriori algorithm. Also here the main idea is based on partial support of itemsets, in this case by transactions. The following way of calculating support of itemsets for single transaction was introduced: given k-itemset Z k , each element included in a transaction t and the set Z k contributes in 1/k to total value of support of the Z k set calculated for the transaction t. The total value of support of the Z k set is computed by summing up all values contributed by elements included in the transaction. If the set Z k is contained in the transaction t then the value of support of the set calculated for the transaction is equal to k ∗ (1/k) = 1. The ∼AR algorithm starts from replacing each unknown element u elem in the transactions by all known items k elem corresponding to the unknown ones in other transactions. Each element k elem replaces the u elem with the probability evaluated based on its frequency of occurring in the transaction, of which

312

G. Protaziuk and H. Rybinski

the k elem elements come form. The authors assumed that such replacing is possible because of presence of names of classes of elements or uniﬁed ordering of items in the transactions. A value contributed by a single k elem to the total value of support of the Zk set calculated for the transaction t is additionally multiplied by the probability of its occurring.

4

Discovering of Association Rules Under Incompleteness

Incompleteness of data creates problem in interpreting minimal support and minimal conﬁdence thresholds given by the user. In the case of missing or unknown values it is not possible to calculate exact values of the support and conﬁdence measures. Instead, we can provide the estimation of these values and the range, limited by optimistic and pessimistic support (conﬁdence), in which the true value is placed. 4.1

Support and Conﬁdence Under Incompleteness

In the order to express the properties of data incompleteness we will apply the following notions: • by minSet(X) we denote the maximal set of records (transactions) which certainly support the itemset X. • by maxSet(X) we denote the maximal set of records (transactions) which possibly support the itemset X. • by nkD we denote the maximal set of records (transactions) which are incomplete (i.e. include at least one item with unknown or missing value). • by kmD we denote the maximal set of records (transactions) which include only items with known values. Deﬁnition 1. Minimal (pessimistic) support of an itemset X, denoted further as pSup(X), is deﬁned as the number of records (transactions) which certainly support the itemset X, i.e. pSup(X) = |minSet(X)|. Deﬁnition 2. Maximal (optimistic) support of an itemset X, denoted further as oSup(X), is deﬁned as the number of records (transactions) which possibly support the itemset X : oSup(X) = |maxSet(X)|. The estimated (probable) support of an itemset X is denoted further as probSup(X). Deﬁnition 3. Minimal (pessimistic) conﬁdence of a rule X → Y , denoted further as pConf (X → Y ), is deﬁned as: pConf (X → Y ) = |minSet(X ∪ Y )|/|maxSet(X)|. Deﬁnition 4. Maximal (optimistic) conﬁdence of a rule X → Y , denoted further as oConf (X → Y ), is deﬁned as: oConf (X→Y )=|maxSet(X∪Y )|/(|maxSet(X∪Y )|+|minSet(X)\minSet(Y )|).

Discovering Association Rules in Incomplete Transactional Databases

313

Deﬁnition 5. Estimated (probable) conﬁdence of a ruleX → Y , denoted further as probConf (X → Y ), is deﬁned as: probConf (X → Y ) = probSup(X ∪ Y )/probSup(X). 4.2

Support Calculation for Single Record or Transaction

The standard method of calculating support for a complete record can be expressed by the formula: Supr (X) =

1 if X is present in a record r 0 otherwise

(5)

where Supr (X) denotes support of the itemset X calculated for the record r, r ∈ DR. Instead, in the case of incomplete records we estimate probable support, for which we can use a generalization of (5) in the form: probSupr (X) = prob(sat(X, r)) if X may be present in a record r

(6)

where prob(sat(X, r)) denotes probability that the itemset X is present in the record r. Obviously if X is certainly present in a record r (all needed values in r are present) then probSupr (X) = 1. If X cannot be present in r then probSupr (X) = 0. For transactional data the method of calculating support in a single transaction is analogous. 4.3

Postulates for Relational Data

The simplest approach to the calculation of support and conﬁdence based on optimistic or pessimistic estimation does not promise good results, especially if the incompleteness of data is signiﬁcant and cannot be neglected. To obtain result of a higher quality more advanced techniques for the support and conﬁdence estimation should be used. In the literature some proposals in this direction have been published. As mentioned above, we claim that any such method should satisfy the postulates deﬁned in [12], in order to properly asses statistically justiﬁed expected support and conﬁdence. The original postulates refer to the relational database. Let us recall them: (P1) (P2) (P3) (P4) (P5)

probSup(X) ∈ [pSup(X), oSup(X)] probSup(X) ≥ probSup(Y ) for X ⊂ Y probConf (X → Y ) = probSup(X ∪ Y )/probSup(X) probConf (X) ∈ [pConf (X → Y ), oConf (X → Y )] XInstances(A) probSup(X) = 1 for any A ⊆ AT

where X, Y are itemsets, and Instances(A) is the set of all possible tuples over the set of attributes A. Postulate P1 assures the natural limitation of estimated support – it can not be greater than optimistic support and less than pessimistic support. The second

314

G. Protaziuk and H. Rybinski

postulate says that an itemset can not be present in database more often than its proper subset. The postulate 3 introduces the standard way for conﬁdence calculation. The fourth one is analogous to P1 but refers to conﬁdence. The last one states that for freely chosen set of attributes the sum of support of all itemsets consisting of items which belong to a domain of selected attributes is equal to the number of records included in a database. In the [16] it is shown that P4 is redundant. With the satisﬁed condition of P2 the estimated support is consistent with the fact that any superset of an itemset X does not occur more often than X. However, satisfying these criteria does not necessary means that the sum of support of supersets of an itemset X, speciﬁed on the same set of attributes, is not greater than support of X. Example: For the record r8 from the base DR n1 (Table 1) the probabilities P r o of occurrences are deﬁned for the following sets: P r w(z1 = {atr3 = a}) = 2/3, P r w(z2 = {atr3 = a1}) = 1/3, P rw (z3 = {atr4 = b}) = 1/2, P rw (z4 = {atr4 = c}) = 1/2, P rw (z5 = {atr3 = a, atr4 = b}) = 1/2, P rw (z6 = {atr3 = a, atr4 = c}) = 1/2. The support of an itemset Y calculated for the record r8 is as follows: P r w(zN ) if Y = zN ∪ r8.knowni if Y ⊂ r8.known probSup(Y ) = 1 0 Otherwise where: zN – is one of the sets zi deﬁned above, r8.known – is a set of all sets which can be generated from known values included in the record r8, r8.knowni – one of sets included in the r8.known set. The support deﬁned in this way satisﬁes all the postulates, however after counting supports of the following set [Y 1 = {atr2 = x, atr3 = a}, probSup(Y 1) = 42/3], [Y 2 = {atr2 = x, atr3 = a, atr4 = b}, probSup(Y 2) = 2/1], [Y 3 = {atr2 = x, atr3 = a, atr4 = c}, probSup(Y 3) = 2/1] one can conclude that the sets Y 2 and Y 3 occur in the database more often than the set Y 1, which of course is not possible. Table 1. Relational database DR n1 id atr2 atr3 atr4 atr5 atr6 r1 x A b Y v1 r2 x A b Y v2 r3 x A c Y v3 r4 x A c Y v4 r5 x a1 b1 Z v1 r6 x a1 c1 P v3 r7 k a2 b2 P v3 r8 x ∗ ∗ Y v2

* – the missing value

Discovering Association Rules in Incomplete Transactional Databases

315

To exclude such situations the postulate P2 has to be modiﬁed. Below the new version of the postulate is presented. (P2n) : probSup(X) = probSup(X ∪ Z) for any A ⊆ AT. (7) Z∈Instances(A)

where X = ∅, Z = ∅ and X ∩ Z = ∅ for any set Z. Rationale: Support of a set X can be counted by testing values in records only for those attributes over which the set X is deﬁned. If we consider larger set of attributes then of course the support of the set X does not change. In this case we can say that we examine whether in a given record the set X is present along with an additional set Z deﬁned over attributes out of those in X. If in place of Z consecutively all the sets deﬁned over the additional attributes will be considered then ﬁnally we will obtain valid value of the support of X. 4.4

Postulates for Transactional Data

For transactional data the postulates from [12] can be used directly, except for P2n and P5. In the deﬁnitions below we use the following notations: • superset(X, k) is a set of all supersets of X which include k more elements than X. • tmax is a maximal transaction i.e. transaction belonging to the given DT which has the most known elements. • |t| is the length of transaction t – number of elements included in this transaction • DTk – set of transaction which include at least k elements • sets(Ik ) denotes sets of k-itemsets • ti denotes the ith transaction. The postulate 2 for transactional data is deﬁned as follows: tmax |−|X|

probSup(X) ≥

k=1

(−1)(k+1)

probSup(Y ).

(8)

Y ∈superset(X,k)

The inequality in the formula results from taking into consideration the diﬀerent lengths of transactions. The postulate 5 for transactional data is deﬁned as follows: X∈sets(Ik ) probSup(X) =1 (9) |Dk | k (C|ti | ti ∈ DTk k where C|t| is a number of all combinations of k items from t. In the deﬁnition of the postulate the absolute support is used.

Rationale for Postulate 2: Calculation of a support of an itemset X can be done by summing number of transactions which contain only all elements

316

G. Protaziuk and H. Rybinski

included in X with the number of transactions that include the set X and at least one more element i.e. which contain a set z = {X ∪{y}|{y}∩X = ∅, y −an item}. There are up to n such sets z, where n is a number of items y, which occur in the database but not in X. A number of transaction in which at least one of the sets z occurs as the percentage of the number of all transaction can be calculated by adopting the following formula (the probability of the sum of events): P(A1 ∪ A2 ∪ . . . ∪ An ) = P(A1 ) + P(A2 ) + . . . + P(An ) − P(A1 ∩ A2 ) − P(A1 ∩ A3 ) − . . . − P(An−1 ∩ An ) + P(A1 ∩ A2 ∩ A3 ) + P(A1 ∩ A2 ∩ A4 ) + . . . + P(An−2 ∩ An−1 ∩ An ) + . . . + (−1)n−1 P(A1 ∩ . . . ∩ An ).

(10)

Assuming that an event Ai represents occurrence of sets zi1 in a transaction (where zi1 is an ith set composed from item yi and elements included in X) we can write the following equation: P (Z 1 ) =

n

(−1)k+1

P (zik )

(11)

zik ∈Z k

k=1

where Z k = superset(X, k), and P(Z 1 ) is the probability that the transaction t contains X and at least one additional item. If we replace the probability by frequencies, and express the frequencies by relative support of the sets we obtain the following formula: ptobSup(Z 1) =

n

(−1)(k+1)

probSub(zik )

(12)

zik ∈Z k

k=1

As we are interested in calculating support of itemsets, we should consider only such z k sets for which k + |X| ≤ |tmax |. All the more numerous sets have support equal to 0. Hence, we can rewrite our formula in the following manner: |tmax |−|X| 1

probSup(Z ) =

(−1)(k+1)

probSup(zik ).

(13)

z1k ∈Z k

To calculate support of the set X we have to sum the values: (i) resulting from the formula (13) and (ii) support of X for the transaction with exactly the same items as in X. This leads us ﬁnally to the formula (8) above. Rationale for postulate 5: Each transaction supports n k-itemsets, where k ≤ |t| and n is the number of all diﬀerent k-elements sets which can be created the transaction. Hence for each transaction from the items belonging to k t: X∈sets(Ik ) probSupt (X) = C|t| , where probSupt (X) denotes the support of the itemset X counted for the transaction t. The fulﬁllment of the postulate 5 requires estimation of support for each possible itemset, even if the given itemset is certainly infrequent. In the tasks of discovering association rules only frequent sets are interested, so there is no need to

Discovering Association Rules in Incomplete Transactional Databases

317

take into consideration infrequent itemsets. In our opinion the postulate 5 can be weaken by replacing the equality relation (=) by the relation ≤ in the formula (9). This modiﬁcation preserves natural deﬁnition of the postulate and makes it more practical. The modiﬁed postulates will be denoted as SJATD (Statistically Justiﬁed Approach for Transactional Data). 4.5

Postulate for Single Record or a Transaction

The postulates SJATD presented above provide conditions for the support calculated for the entire database. On the other hand they say nothing about required properties of support counted for a single record or transaction. For the methods that estimate support of an itemset X by summing its support calculated for single record it seems that more practical is to deﬁne conditions that should be fulﬁlled by support calculated for single record or transaction rather than calculated for entire database. Below we present such requirements. Postulates for single record (P1r) probSupR (X) ∈ [0, 1] (P2r) probSupR (X) ≥ probSupR (Y ) for X ⊂ Y (P3r) probSupR (X) ≥ Z∈sets(At) probSupR (X∪Y )X = ∅, Z = ∅ and X∩Z = ∅ for any set Z (P4r) X∈set(At) , probSupR (X) = 1 for each set of attributes At ⊆ A. For transactional data the postulates: (P3r) and (P4r) have to be redeﬁned. The appropriate formulas are given below. Postulate P3r for transactional data: |T |−|X|

probSupT (X) ≥

(−1)(k+1)

probSupT (Y )

(14)

Y ∈superset(X,k)

k=1

Postulate P4r for transactional data X∈sets(Ik )

probSupT (X)

k C|T |

=1

(15)

where probSupT (X) is a support of an itemsets X counted for transaction T .

5

Algorithm DARIT

In this section we present the new algorithm for discovering association rules from incomplete transactional data, called in the sequel DARIT (Discovering Association Rules in Incomplete Transactions). In our approach we allow that incomplete transaction may have any number of missing element. We start from the description of the data structure, called mT-tree, which is used in the algorithm.

318

5.1

G. Protaziuk and H. Rybinski

The mT-Tree

The tree-like mT-tree structure is used for storing information concerning sets tested by DARIT. Each node of this tree has assigned level, (the root is at level 1). Data associated with itemsets (supports, items) are stored in a node in a table of elements, denoted further as tblElem. Each ﬁeld of the tblElem table in a node at level l may have pointer to another node at level l + 1. With each element stored in the tblElem table there are three values associated: val pSup, val probSup, val oSup, which are used for calculating pessimistic, estimated and optimistic supports respectively. On Figure 1 a simpliﬁed structure of mT-tree is presented. In the tables are only shown elements and name of a val pSup parameter with itemsets to which this parameter concerns. Items belonging to a path in the mT-tree, beginning from the root and composed of the ﬁelds of the tblElem tables, form a frequent or potentially frequent itemset. The items stored in the root of the mT-tree are frequent 1 itemsets. For instance, the set {a, b, c, d} is represented by the path (n1[a], n2[b], n5[c], n8[d]) on Figure 1, where nN[x] stands for the ﬁeld in tblElem where the item x is stored in a node of number N .

n1 - level 1. ( root of tree ) val_pSup{a} val_pSup{b} val_pSup{c} val_pSup{d} a b c d

n4 - level 2. val_pSup{c,d} d

n2 - level 2. val_pSup{a,b} val_pSup{a,c} val_pSup{a,d} b c d

n3 level 2. val_pSup{b,c} val_pSup{b,d} c d n5 - level 3. val_pSup{a,b,c} val_pSup{a,b,d} c d

n8 - level 4. val_pSup{a,b,c,d} d

n6 - level 3. val_pSup{a,c,d} d

n7 - level 3. val_pSup{b,c,d} d

Fig. 1. mT-tree - simpliﬁed schema

5.2

Algorithm DARIT

The pseudo code of the algorithm DARIT is presented below.

Discovering Association Rules in Incomplete Transactional Databases

319

Algorithm DARIT 1) mTd: mT-tree; 2) Apriori Adapt(kmD, nkD, mTd) 3) for each transaction t ∈ nkD 4) begin 5) Generate Set NZ(t, mTd); 6) mTd.Modify probSup(t,t.NZ); 7) end 8) Generate Rules(mTd); At the beginning the set of potentially frequent itemsets is generated (line 2) by calling the Apriori Adapt procedure. In the procedure for determining potentially frequent itemsets the pessimistic support and minimal pessimistic supports thresholds are used. The minimal pessimistic support, denoted further as min minSup, is an additional parameter of the algorithm. This parameter deﬁnes threshold which should be exceeded by the pessimistic support of each itemset in order to consider the itemset as potentially frequent. It allows for appropriate limitation of number of sets taken into consideration during execution of the procedure Apriori Adapt, especially in the case of signiﬁcant data incompleteness. In the next step of the DARIT algorithm for each incomplete transaction t, based on the sets stored in the mT-tree a set t.N Z is generated; it consists of the sets which may occur in place of the special element null, which indicates missing elements in the transactions (line 5). For each element of the t.N Z set the probability of occurring in the considered transaction is assigned. Based on the results obtained in this step, the values of the probSup support of itemsets represented in the mT-tree are modiﬁed (line 6). At the end of the algorithm the procedure Generate Rules is called – it produces the association rules from the mT-tree using values of estimated support of itemsets. Procedure Apriori Adapt Procedure Apriori Adapt(Set of transaction kmD,nkD; mT-tree mTd) 1) Add Frequent Items(kmD, nkD, mTd) 2) p=2; 3) while(Generate Candidates(p, mTd)> 0) 4) begin 5) for each transaction t ∈ kmD 6) Calculate Support(t, mTd); 7) for each transaction t ∈ nkD 8) Calculate Support Incomplete(t, mTd); 9) mTd.Remove NotFrequent( minSup, min minSup); 10) p=p+1; 11) end The procedure Apriori Adapt starts from adding 1-itemset potentially frequent to the root of the mT-tree (line 1). The function Generate Candidates creates

320

G. Protaziuk and H. Rybinski

candidate sets and returns their number. Candidate sets in a node n at level p are generated by creating child nodes at level p + 1, for each ﬁeld in the table tblElem, except for the last one. In the child node cn created for the j th ﬁeld, the table of elements consists of all those elements from the table tblElem of the parent node, which are stored in the ﬁelds of the index greater than j. The procedure Calculate Support increases value of the optimistic support and of the pessimistic support for those of the candidate sets which are supported by a complete transaction t. The procedure Calculate Support Incomplete (line 8) diﬀers from the procedure Calculate Support in that it increases values of the optimistic support for each candidate set. The method Remove NotFrequent removes the candidate sets which certainly will not be frequent, and for the remaining candidate itemsets it sets value of probSup support to the value of the pessimistic support. Procedure Generate SetNZ Procedure Generate Set NZ(transaction t, mT-tree mTd) 1) k = min(mT d.max length set −1, t.nb known items) 2) while(Stop condition = false and k > 0) 3) begin 4) t.NZ= t.NZ ∪ mTd.Find set nZ(t.knownk ); 5) k = k − 1; 6) Stop condidtion = Check stop condition(); 7) end 8) Calculate Probability(mTd); At the beginning of procedure Generate Set NZ the initial value of k is deﬁned. It is calculated as a minimum of the 2 values (i) number of known items in the transaction t, and (ii) the number of elements in the most numerous potentially frequent set stored in mT-tree. Next, in the method Find set nZ for each incomplete transaction t the (k + j)-itemsets, (denoted as zpc) are looked for in the mT-tree. Formally, zpc = {nZ ∪ t.knownki } and t.knownki is a set consisting of k known elements from the transaction t. First, 1-item nZ sets are found, and then the mT-tree is traversed deeper in order to ﬁnd more numerous nZ sets. The nZ set is added to the t.N Z set if it does not include known elements from the transaction t. With each set nZ the parmProb parameter is associated. Further on, the parameter is used to estimate probability of occurrence of the given set nZ instead of the special element null in the transaction, thus to estimate the value of the probSup support. The value of the parmP rob is equal to the value of minSup(zpc) of the currently examined set zpc, or if the given nZ set is already in the t.N Z set, the value of the parameter is increased by minSup(zpc). This procedure is repeated for k = k − 1 down to k = 0, or until the stop condition is met. The stop condition is fulﬁlled when the sum of the parmProb parameters of 1-item nZ sets included in the t.N Z set exceeds the following value: max nb unknown = max len trans − t.nb known elem

(16)

Discovering Association Rules in Incomplete Transactional Databases

321

where max len trans is the maximal number of the potentially frequent items included in the single transaction, and t.nb known elem is the number of known items contained in the transaction t. The Calculate Probability procedure The way in which the value of the parameter parmP rob of the nZ sets is computed causes that it can not be directly used as probability of occurrence of the given nZ set in a transaction. Generally, we have to deal with the following basic problems: 1) A value of the parameter for a single set or a sum of values for group of sets exceeds thresholds. In the former case it is 1 – the maximal value of probability, in the latter the threshold is associated with number of items included in a transaction, for instance, the sum of probabilities of occurrences of single items in a given transaction cannot be greater than the maximal length of the transaction. This condition may be expressed as: i=1..n prob(itemi , t) ≤ max(|t|), where prob(itemi , t) is the probability of occurring of ith item in the transaction t, max(|t|) is the maximal length of the transaction t, n is the number of items potentially frequent. The solution of this problem is a simple normalization of values to the required level. 2) Values of the parameter are very small. In this case the results obtained by applying the DARIT algorithm is comparable with the results obtained by using the methods in which the incompleteness of data is neglected (pessimistic support is used), but with much greater computational cost. To solve the problem these values are multiplied by certain ratio, which is calculated based of most probable number of items which should be present in the transaction in the place of the element null. The ratio is computed in such a way that the sum of probabilities of occurring of single items in a given transaction is not greater than the possible maximal length of the transaction. The Modify probSup method The M odif y probSup method increases the value of the probSup support for each such zpc set stored in the mT-tree, that zpc = {nZ ∪zZ}, where nZ ∈ t.N Z and zZ ∈ t.known. Note that the set zZ may be empty. The value of the probSup support is increased by the value of the parmP rob for the given nZ set. The Generate Rules procedure The Generate Rules procedure generates association rules from the sets stored in the mT-tree in the following way: for each set cZ for which the minimal support requirements are fulﬁlled, all its supersets nZc are found, such that the value probSup(nZc)/probSup(cZ) is not less than the minimal conﬁdence threshold. Next, the rule cZ → {nZc\cZ} is generated with support equal to probSup(nZc) and conﬁdence equal to probSup(nZc)/probSup(cZ). 5.3

The SJATD Postulates

Theorem: DARIT satisﬁes the SJATD postulates.

322

G. Protaziuk and H. Rybinski

Proof: Postulate 1. Value of the probSup support for each set is of course not less than 0. The way of calculating probSup applied in the Calculate Probabilty procedure ensures that the value of probSup support never exceeds 1. Postulate 2. The method of computing the values of the parmSup parameters - summing the values of pessimistic support, ensures that the value of the parameter parmSup for any set will be not less than parmSup for its supersets. Multiplication of these values by the same factor does not inﬂuence this relation. Additionally, the way of adding the sets nZ to the set t.N Z by traversing mT-tree in depth, and adding more and more numerous sets guarantees that for any given set all its subsets have been taken into consideration. Postulate 3. Fulﬁlling this postulate follows from the method of computing the probSup support described in the proof for Postulate 2 and from the fact that the pessimistic support meets this postulate. Postulate 5 (weakened). The postulate says that sum of the support of nitemsets calculated for a single transaction cannot be greater than the number of k-elements sets which can be created from the items included in the transaction. According to the deﬁnition for an incomplete transaction, support of the set dowZ is equal to: • 1, if dowZ ∈ t.known, • nZ.parmSup, if dowZ = nZ or dowZ = nZ∪zZ, nZ ∈ t.N Z, zZ ∈ t.known, • 0 otherwise where t.known contains all the sets which can be formed from the known elements included in the transaction t. Assuming that the number of elements nb elem t in the transaction t is equal to max(|t.N Zk |) + t.nb known, where t.nb known denotes the number of known elements in the transaction t and t.N Zk denotes k − itemsets from the set t.N Z, then number on n-elements sets, created from the items included in the n transaction t is equal to Cnb elem t . The sum Sm of the support of n-itemsets counted for the transaction t can be split into tree parts: Sm = Sknown + Sunknown + Sjoined

(17)

Of course, not all parts occur in all cases. The partial sums in the equation above means, respectively: • Sknown — the sum of the support of the sets dowZ ∈ t.known. The sum is equal to the number of n-elements sets created from the known items included in the transaction t. • Sunknown — the sum of supports of the sets dowZ ∈ t.N Z. In this case the sum is less than the number of the sets created from max(|t.N Zk |) elements. It is ensured by the way of computing the normalization ratio in the method Calculate Probability. • Sjoined — the sum of supports of the sets dowZ = nZ ∪ zZ, nZ ∈ t.N Z, zZ ∈ t.known, nZ = ∅, zZ = ∅. The sum is smaller than the number of n-element sets created by joining the i-itemsets created from t.nb known

Discovering Association Rules in Incomplete Transactional Databases

323

elements and j-itemsets created from the max(|t.N Zk |) elements for j, i > 0 and j + i = k. The number of such sets can be computed from the following equation:

i=min(n−1,t.nb known) max(t.N Zk )

known Ct.nb ∗ Cn−1 i

(18)

i=1

for i ≤ t.nb known and n − i ≤ max(|t.N Zk |). During the calculation of the sum Sjoined the second factor is replaced by the sum of supports of the sets nZ ∈ t.N Zn−i , which as it follows from explanation presented for Sunknown is smaller than the value of this factor in the original formula.

6

Experiments

percent differenceof supportsof sets

The objective of the performed experiments was to evaluate the practical value of the proposed approach. In order to evaluate the quality of the results obtained by executing the DARIT algorithm the following procedure was applied. First, the sets of association rules and frequent itemsets, further denoted as the referential sets, were generated from a complete transactional database. Next, the incomplete database was created by random removing some elements from 40% 35% d_1

30%

d_2 25% d_5 20%

d_10

15%

ign_1

10%

ign_2

5%

ign_5

0% 0%

ign_10 10%

20% 30% 40% 50% 60% 70% percent of incomlete transactions

80%

90%

100%

Fig. 2. Percent diﬀerence of supports of sets 100% not discoveredsetsinpercent

90% 80%

d_1

70%

d_2

60%

d_5

50%

d_10

40%

ign_1

30% ign_2

20%

ign_5

10% 0% 0%

ign_10 10%

20% 30% 40% 50% 60% 70% percent of incomlete transactions

80%

90%

Fig. 3. Percent of non-discovered sets

100%

324

G. Protaziuk and H. Rybinski 100%

additional setsinpercent

sumof not discoveredand

90% 80%

d_1

70%

d_2

60%

d_5

50%

d_10

40%

ign_1

30% ign_2 20% ign_5

10% 0% 0%

ign_10 10%

20% 30% 40% 50% 60% 70% percent of incomlete transactions

80%

90%

100%

Fig. 4. Erroneous sets as percentage of the size of the referential set

percent differenceof supportsof rules

32% 28% d_1

24%

d_2

20%

d_5 16%

d_10

12%

ign_1

8%

ign_2

4%

ign_5

0% 0%

ign_10 10%

20% 30% 40% 50% 60% 70% percent of incomlete transactions

80%

90%

100%

Fig. 5. Percent diﬀerence of supports of association rules

8%

percent difference

of confidencesof rules

7% d_1

6%

d_2 5% d_5 4%

d_10

3%

ign_1

2%

ign_2 ign_5

1% 0% 0%

ign_10 10%

20% 30% 40% 50% 60% 70% percent of incomlete transactions

80%

90%

100%

Fig. 6. Percent diﬀerence of conﬁdences of association rules

some transactions. The number of incomplete transactions and the number of unknown elements in a single transaction vary in the experiments. The last step in the procedure was to use the DARIT algorithm to discover association rules in the incomplete data. The evaluation of quality of the results was based on the diﬀerence between the referential sets of association rules or frequent sets and the sets obtained from analyzing incomplete data. The diﬀerence was described by the following measures: – percent diﬀerence between the supports of rules and frequent itemsets and between the conﬁdence measures

Discovering Association Rules in Incomplete Transactional Databases

325

not discovered rulesinpercent

110% 100% 90%

d_1

80%

d_2

70%

d_5

60% 50%

d_10

40%

ign_1

30%

ign_2

20%

ign_5

10% 0% 0%

ign_10 10%

20% 30% 40% 50% 60% 70% percent of incomlete transactions

80%

90%

100%

Fig. 7. Percent of non-discovered association rules 110%

additional rulesinpercent

sumof not discoveredand

100% 90%

d_1

80% d_2

70% 60%

d_5

50%

d_10

40%

ign_1

30%

ign_2

20%

ign_5

10% 0% 0%

ign_10 10%

20% 30% 40% 50% 60% 70% percent of incomlete transactions

80%

90%

100%

Fig. 8. Erroneous rules as a percentage of the size of the referential set

– percentage of the sets (rules) presented in the referential set that have not been discovered from incomplete data – number of additional sets (rules) discovered from incomplete data but not present in the referential set, expressed as a percentage of the referential set. The results obtained from DARIT were also compared with the results obtained from the other data mining algorithms, where the pessimistic support estimation is used. In the experiments the synthetic data were used. The set consisted of 592 transactions, the average length of the transactions was 20 elements. It was 1116 diﬀerent items in the database. The test databases were incomplete in the degree between 10% up to 90% of transactions. For each case four situations with diﬀerent number of missing elements in a single transaction were analyzed, namely the situation in which 1, 2, 5, or 10 elements in a single transaction were missing. In all experiments the minimal support threshold was set to 0.06 and the minimal conﬁdence threshold was set 0.9. The tests were carried out for four values of minimal pessimistic support threshold calculated as a percentage of the minimal support value, namely 10%, 20%, 50%, 80% of this value. On the charts for Figure 2 to Figure 8 the average results for the minimal pessimistic support are presented. On the legend d denotes the results obtained from DARIT, ign stands for the ignore method, and the index at d or ign indicates the number of missing elements in a single transaction. The results concerning frequent itemsets are presented on the ﬁgures 2, 3 and 4. The proposed approach is generally much better than the ignore method.

326

G. Protaziuk and H. Rybinski

On Figure 4 the error of the algorithms DARIT and ignore is presented in the form of the sum of numbers of missing and erroneously discovered sets. Such a sum better presents diﬀerences between results obtained by the considered methods, since the ignore method does not produce erroneous sets. The results coming from those two methods are comparable only in the situations where the incompleteness of data is small. The presented results show that by applying the DARIT algorithm one can achieve good results also in the case of quite high incompleteness of data (up to 50% of incomplete transactions). On Figures 5 – 8 we present the results concerning comparisons of the sets of discovered association rules. Also with respect to association rules DARIT is much better than the ignore method, though here the diﬀerence is smaller. The obtained results are generally worse than the results obtained for the frequent sets. The errors in estimating support rather cumulate then eliminate. In the case of association rules only the best conﬁgurations concerning the minimal pessimistic support thresholds allow to obtain good results when the incompleteness of data is higher.

7

Conclusions

In the paper we extended some postulates of legitimate approach to discovering association rules, as deﬁned in [12], so that they may be applied also to transactional data. We have presented in detail a new approach for discovering association rules from incomplete transactional databases. In the presented DARIT algorithm we do not use any knowledge that is external to the dataset, but exploit only relations discovered in the investigated database. We have shown that the approach satisﬁes all the postulates of SJATD. We presented a number of experimental results using synthetic data. The performed experiments show that the proposed algorithm well foresees real values of the support and outperform the methods based on ignoring unknown values.

References 1. Agrawal R., Imielinski T., Swami A.: Mining Associations Rules between Sets of Items in Large Databases. In: Proc. of the ACM SIGMOD Conference on Management of Data, Washington, USA (1993) 207–216 2. Agrawal R., Srikant R.: Fast Algorithms for Mining Association Rules. In: Proc. of the 20th International Conference on Very Large Databases Conference (VLDB), Santiago, Chile, 1994. Morgan Kaufmann (1994) 487–499 3. Bayardo R.J., Agrawal R.: Mining the Most Interesting Rules. In: Proc. of the Fifth ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD), San Diego, CA, USA, 1999. ACM (1999) 145–154 4. Bayardo R.J., Agrawal R., Gunopulos D.: Constraint-Based Rule Mining in Large, Dense Databases. Data Mining and Knowledge Discovery, Vol. 4, No. 2/3 (2000) 217–240 5. Bayardo R. J. Jr.: Eﬃciently Mining Long Patterns from Databases, Proceedings of ACM SIGMOD International Conference on Management of Data, Seattle, (1998).

Discovering Association Rules in Incomplete Transactional Databases

327

6. Breiman L., Friedman J. H., Olshen R. A. Stone C. J.: Classiﬁcation and regression trees, Belmont Wadsworth, (1984). 7. Calders T., Goethals B.: Mining All Non-derivable Frequent Item Sets, Proc. of Principles of Data Mining and Knowledge Discovery, 6th European Conf., Helsinki, (2002) 8. Dardzi´ nska-Gl¸ebocka A., Chase Method Based on Dynamic Knowledge Discovery for Prediction Values in Incomplete Information Systems, PhD thesis, Warsaw, 2004. 9. Friedman H. F., Kohavi R., Yun Y., Lazy decision trees, Proceedings of the 13th National Conference on Artiﬁcial Intelligence, Portland, Oregon, (1996) 10. Grzymala-Busse J. W.: Characteristic Relations for Incomplete Data: A Generalization of the Indiscernibility Relation, Proceedings Rough Sets and Current Trends in Computing, 4th International Conference, Uppsala, (2004). 11. Han J., Pei J., Yin Y.: Mining Frequent Patterns without Candidate Generation. In: Proc. Of the 2000 ACM SIGMOD International Conference on Management of Data, Dallas, Texas, USA, 2000. SIGMOD Record, Vol. 29, No. 2 (2000) 1–12 12. Kryszkiewicz M., Rybinski H.: Legitimate Approach to Association Rules under Incompleteness. In: Foundations of Intelligent Systems. Proc. of 12th International Symposium (ISMIS), Charlotte, USA, 2000. Lecture Notes in Artiﬁcial Intelligence, Vol. 1932. Springer-Verlag (2000) 505–514 13. Kryszkiewicz M.: Probabilistic Approach to Association Rules in Incomplete Databases, Proceedings of Web-Age Information Management, First International Conference, WAIM 2000, Shanghai, (2000). 14. Kryszkiewicz M.: Concise Representation of Frequent Patterns based on Disjunction-Free Generators. In: Proc. of the 2001 IEEE International Conference on Data Mining (ICDM), San Jose, California, USA, 2001. IEEE Computer Society (2001) 305–312 15. Kryszkiewicz M.: Representative Association Rules. In: Research and Development in Knowledge Discovery and Data Mining. Proc. of Second Paciﬁc-Asia Conference (PAKDD). Melbourne, Australia, 1998. Lecture Notes in Computer Science, Vol. 1394. Springer (1998) 198–209 16. Kryszkiewicz M.: Concise Representations of Frequent Patterns and Association Rules Habilitation Thesis, Warsaw University of Technology, (2002) 17. Liu W. Z., White A. P.: Thompson S. G., Bramer M. A.: Techniques for Dealing with Missing Values in Classiﬁcation, Proceedings of Advances in Intelligent Data Analysis, Reasoning about Data, Second International Symposium, London, (1997) 18. Nayak J. R., Cook D. J.: Approximate Association Rule Mining, Proceedings of the Fourteenth International Artiﬁcial Intelligence Research Society Conference, Key West, Florida, (2001) 19. Parsons S.: Current Approach to Handling Imperfect Information in Data and Knowledge Bases, IEEE Transaction on knowledge and data engineering Vol. 8, (1996) 20. Pawlak Z.: Rough Sets. International Journal of Information and Computer Sciences No. 11 (1982) 341–356 21. Pawlak Z.: Rough Sets: Theoretical Aspects of Reasoning about Data. Kluwer Academic Publishers, Vol. 9 (1991) 22. Pasquier N., Bastide Y., Taouil R., Lakhal L.: Discovering Frequent Closed Itemsets for Association Rules. In: Proc. of Database Theory - ICDT ’99. Proc. of 7th International Conference (ICDT), Jerusalem, Israel, 1999. Lecture Notes in Computer Science, Vol. 1540. Springer (1999) 398–416

328

G. Protaziuk and H. Rybinski

23. Protaziuk G., Sodacki P., Gancarz ., Discovering interesting rules in dense data, The Eleventh International Symposium on Intelligent Information Systems, Sopot, (2002). 24. Bastide Y., Pasquier N., Taouil R., Stumme G., Lakhal L.: Mining Minimal Nonredundant Association Rules Using Frequent Closed Itemsets. Comp. Logic (2000) 972–986 25. Quinlan J. R., C4.5 Programs for Machine Learning, San Mateo, California, (1993) 26. Ragel A., Cremilleux B.: Treatment of Missing Values for Association Rules. In: Research and Development in Knowledge Discovery and Data Mining. Proc. of Second Paciﬁc-Asia Conference (PAKDD). Melbourne, Australia, 1998. Lecture Notes in Computer Science, Vol. 1394. Springer (1998) 258–270 27. Srikant R., Vu Q., Agrawal R.: Mining Association Rules with Item Constraints. In: Proc. Of the Third International Conference on Knowledge Discovery and Data Mining (KDD). Newport Beach, California, USA, 1997. AAAI Press (1997) 67–73 28. Stefanowski J., Tsoukias A.: Incomplete Information Tables and Rough Classiﬁcation. Int. Journal of Computational Intelligence, Vol. 17, No 3 (2001) 545–566 29. Stefanowski J.: Algorytmy indukcji regu decyzyjnych w odkrywaniu wiedzy (Algorithms of Rule Induction for Knowledge Discovery). Habilitation Thesis, Poznan University of Technology, No. 361 (2001) 30. Wang G.: Extension of Rough Set under Incomplete Information Systems, Proceedings of the 2002 IEEE International Conf. on Fuzzy Systems, Honolulu, (2002) 31. Zaki M.J.: Generating Non-Redundant Association Rules. In Proc. of 6th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Boston, MA, 2000. ACM Press (2000) 34–43 32. Zhang J., Honavar V.: Learning Decision Tree Classiﬁers from Attribute Value Taxonomies and Partially Speciﬁed Data, Proceedings of the Twentieth International Conference (ICML 2003), Washington, DC, (2003)

On Combined Classiﬁers, Rule Induction and Rough Sets Jerzy Stefanowski Institute of Computing Science Pozna´ n University of Technology, 60-965 Pozna´ n, ul.Piotrowo 2, Poland Jerzy.Stefanowski@cs.put.poznan.pl

Abstract. Problems of using elements of rough sets theory and rule induction to create eﬃcient classiﬁers are discussed. In the last decade many researches attempted to increase a classiﬁcation accuracy by combining several classiﬁers into integrated systems. The main aim of this paper is to summarize the author’s own experience with applying one of his rule induction algorithm, called MODLEM, in the framework of diﬀerent combined classiﬁers, namely, the bagging, n2 –classiﬁer and the combiner aggregation. We also discuss how rough approximations are applied in rule induction. The results of carried out experiments have shown that the MODLEM algorithm can be eﬃciently used within the framework of considered combined classiﬁers.

1

Introduction

Rough sets theory has been introduced by Professor Zdzislaw Pawlak to analyse granular information [25,26]. It is based on an observation that given information about objects described by attributes, a basic relation between objects could be established. In the original Pawlak’s proposal [25] objects described by the same attribute values are considered to be indiscernible. Due to limitations of available information, its natural granulation or vagueness of a representation language some elementary classes of this relation may be inconsistent, i.e. objects having the same descriptions are assigned to diﬀerent categories. As a consequence of the above inconsistency it is not possible, in general, to precisely specify a set of objects in terms of elementary sets of indiscernible objects. Therefore, Professor Zdzislaw Pawlak introduced the concept of the rough set which is a set characterized by a pair of precise concepts – lower and upper approximations constructed from elementary sets of objects. This quite simple, but smart, idea is the essence of the Pawlak’s theory. It is a starting point to other problems, see e.g. [27,20,26,9]. In particular many research eﬀorts have concerned classiﬁcation of objects represented in data tables. Studying relationships between elementary sets and categories of objects (in other terms, target concepts or decision classes in the data table) leads to, e.g., evaluating dependency between attributes and objects classiﬁcation, determining the level of this dependency, calculating importance of attributes for objects J.F. Peters et al. (Eds.): Transactions on Rough Sets VI, LNCS 4374, pp. 329–350, 2007. c Springer-Verlag Berlin Heidelberg 2007

330

J. Stefanowski

classiﬁcation, reducing the set of attributes or generating decision rules from data. It is also said that the aim is to synthesize reduced, approximate models of concepts from data [20]. The transparency and explainability of such models to human is an important property. Up to now rough sets based approaches were applied to many practical problems in diﬀerent domains – see, e.g., their list presented in [20]. Besides ”classical” rough sets, based on the indiscernibility relation, several generalizations have been introduced. Such data properties as, e.g., imprecise attribute values, incompleteness, preference orders, are handled by means of tolerance, similarity, fuzzy valued or dominance relations [9,20,37]. Looking into the previous research on rough sets theory and its applications, we could distinguish two main perspectives: descriptive and predictive ones. The descriptive perspective includes extraction information patterns or regularities, which characterize some properties hidden in available data. Such patterns could facilitate understanding dependencies between data elements, explaining circumstances of previous decisions and generally gain insight into the structure of the acquired knowledge. In this context presentation of results in a human readable form allowing an interpretation is a crucial issue. The other perspective concerns predicting unknown values of some attributes on the basis of an analysis of previous examples. In particular, it is a prediction of classes for new object. In this context rough sets and rules are used to construct a classiﬁer that has to classify new objects. So, the main evaluation criterion is a predictive classiﬁcation accuracy. Let us remind that the predictive classiﬁcation has been intensively studied since many decades in such ﬁelds as machine learning, statistical learning, pattern recognition. Several eﬃcient methods for creating classiﬁers have been introduced; for their review see, e.g., [16,19,23]. These classiﬁers are often constructed with using a search strategy optimizing criteria strongly related to predictive performance (which is not directly present in the original rough sets theory formulation). Requirements concerning interpretability are often neglected in favor of producing complex transformations of input data – an example is an idea of support vector machines. Although in both perspectives we could use the same knowledge representation – rules, since motivation and objectives are distinct, algorithmic strategies as well as criteria for evaluating a set of rules are quite diﬀerent. For instance, the prediction perspective directs an interest to classiﬁcation ability of the complete rules, while in the descriptive perspective each rule is treated individually as a possible representative of an ‘interesting’ pattern evaluated by measures as conﬁdence, support or coverage - for a more exhaustive discussion see, e.g., [42]. In my opinion, basic concepts of the rough sets theory have been rather considered in the way similar to a descriptive analysis of data tables. Nevertheless, several authors have developed their original approaches to construct decision rules from rough approximations of decision classes which joined together with classiﬁcation strategies led to good classiﬁers, see e.g. [1,11,20,34,37]. It seems to me that many authors moved their interest to this direction in the 90’s because of at least two reasons: (1) a research interest to verify whether knowledge

On Combined Classiﬁers, Rule Induction and Rough Sets

331

derived from ”closed world” of the data table could be eﬃciently applied to new objects coming from the ”open world” – not seen in the analysed data table; (2) as a result of working with real life applications. Let us also notice that the majority of research has been focused on developing single classiﬁers – i.e. based on the single set of rules. However, both empirical observations and theoretical works conﬁrm that one cannot expect to ﬁnd one single approach leading to the best results on overall problems [6]. Each learning algorithm has its own area of superiority and it may outperform others for a speciﬁc subset of classiﬁcation problems while being worse for others. In the last decade many researches attempted to increase classiﬁcation accuracy by combining several single classiﬁers into an integrated system. These are sets of learned classiﬁers, whose individual predictions are combined to produce the ﬁnal decision. Such systems are known under names: multiple classiﬁers, ensembles or committees [6,45]. Experimental evaluations shown that these classiﬁers are quite eﬀective techniques for improving classiﬁcation accuracy. Such classiﬁers can be constructed in many ways, e.g., by changing the distributions of examples in the learning set, manipulating the input features, using diﬀerent learning algorithms to the same data, see e.g. reviews [6,45,36]. Construction of integrated classiﬁers has also attracted the interest of some rough sets researchers, see e.g. [2,8,24]. The author and his co-operators have also carried out research, ﬁrst on developing various rule induction algorithms and classiﬁcation strategies (a review is given in [37]), and then on multiple classiﬁers [18,36,38,40,41]. The main aim of this paper is to summarize the author’s experience with applying one of his rule induction algorithm, called MODLEM [35], in the framework of diﬀerent multiple classiﬁers: the popular bagging approach [4], the n2 -classiﬁer [18] – a specialized approach to solve multiple class learning problems, and the combiner approach to merge predictions of heterogeneous classiﬁers including also MODLEM [5]. The second aim is to brieﬂy discuss the MODLEM rule induction algorithm and its experimental evaluation. This paper is organized as follows. In the next section we shortly discuss rule induction using the rough sets theory. Section 3 is devoted to the MODLEM algorithm. In section 4 we brieﬂy present diﬀerent approaches to construct multiple classiﬁers. Then, in the successive three sections we summarize the experience of using rule classiﬁers induced by MODLEM in the framework of three diﬀerent multiple classiﬁers. Conclusions are grouped in section 8.

2 2.1

Rules Generation and Rough Sets Notation

Let us assume that objects – learning examples for rule generation – are represented in decision table DT = (U, A ∪ {d}), where U is a set of objects, A is a set of condition attributes describing objects. The set Va is a domain of a. Let fa (x) denotes the value of attribute a ∈ A taken by x ∈ U ; d ∈ / A is a decision attribute that partitions examples into a set of decision classes {Kj : j = 1, . . . , k}.

332

J. Stefanowski

The indiscernibility relation is the basis of Pawlak’s concept of the rough set theory. It is associated with every non-empty subset of attributes C ⊆ A and ∀x, y ∈ U is deﬁned as xIC y ⇔ {(x, y) ∈ U × U fa (x) = fa (y) ∀a ∈ C }. The family of all equivalence classes of relation I(C) is denoted by U/I(C). These classes are called elementary sets. An elementary equivalence class containing element x is denoted by IC (x). If C ⊆ A is a subset of attributes and X ⊆ U is a subset of objects then the sets: {x ∈ U : IC (x) ⊆ X}, {x ∈ U : IC (x) ∩ X = ∅} are called C-lower and C-upper approximations of X, denoted by CX and CX, respectively. The set BNC (X) = CX − CX is called the C-boundary of X. A decision rule r describing class Kj is represented in the following form: if P then Q, where P = w1 ∧ w2 ∧ . . . wp is a condition part of the rule and Q is decision part of the rule indicating that example satisfying P should be assigned to class Kj . The elementary condition of the rule r is deﬁned as (ai (x) rel vai ), where rel is a relational operator from the set {=, , ≥} and vai is a constant being a value of attribute ai . Let us present some deﬁnitions of basic rule properties. [P ] is a cover of the condition part of rule r in DT , i.e. it is a set of examples, which description satisfy elementary conditions in P . Let B be a set of examples belonging to decision concept (class Kj or its appropriate rough approximation in case of inconsistencies). The rule r is discriminant if it distinguishes positive examples of B from its negative examples, i.e. [P ] = [wi ] ⊆ B. P should be a minimal conjunction of elementary conditions satisfying this requirement. The set of decision rules R completely describes examples of class Kj , if each example is covered by at least one decision rules. Discriminant rules are typically considered in the rough sets literature. However, we can also construct partially discriminant rules that besides positive examples could cover a limited number of negative ones. Such rules are characterized by the accuracy measure being a ratio covered positive examples to all examples covered by the rule, i.e. [P ∩ B]/[P ]. 2.2

Rule Generation

If decision tables contain inconsistent examples, decision rules could be generated from rough approximations of decision classes. This special way of treating inconsistencies in the input data is the main point where the concept of the rough sets theory is used in the rules induction phase. As a consequence of using the approximations, induced decision rules are categorized into certain (discriminant in the sense of the previous deﬁnition) and possible ones, depending on the used lower and upper approximations, respectively. Moreover, let us mention other rough sets approaches that use information on class distribution inside boundary and assign to lower approximation these inconsistent elementary sets where the majority of examples belong to the given class. This is handled in the Variable Precision Model introduced by Ziarko

On Combined Classiﬁers, Rule Induction and Rough Sets

333

[47] or Variable Consistency Model proposed by Greco et al. [10] – both are a subject of many extensions, see e.g. [31]. Rules induced from such variable lower approximations are not certain but partly discriminant ones. A number of various algorithms have been already proposed to induce decision rules – for some reviews see e.g. [1,11,14,20,28,34,37]. In fact, there is no unique ”rough set approach” to rule induction as elements of rough sets can be used on diﬀerent stages of the process of induction and data pre-processing. In general, we can distinguish approaches producing minimal set of rules (i.e. covering input objects using the minimum number of necessary rules) and approaches generating more extensive rule sets. A good example for the ﬁrst category is LEM2, MODLEM and similar algorithms [11,35]. The second approaches are nicely exempliﬁed by Boolean reasoning [28,29,1]. There are also speciﬁc algorithms inducing the set of decision rules which satisfy user’s requirements given a priori, e.g. the threshold value for a minimum number of examples covered by a rule or its accuracy. An example of such algorithms is Explore described in [42]. Let us comment that this algorithm could be further extended to handle imbalanced data (i.e. data set where one class – being particularly important – is under-represented comparing to cardinalities of other classes), see e.g. studies in [15,43].

3

Exemplary Rule Classiﬁer

In our study we will use the algorithm, called MODLEM, introduced by Stefanowski in [35]. We have chosen it because of several reasons. First of all, the union of rules induced by this algorithm with a classiﬁcation strategy proved to provide eﬃcient single classiﬁers, [14,41,37]. Next, it is designed to handle various data properties not included in the classical rough sets approach, as e.g. numerical attributes without its pre-discretization. Finally, it produces the set of rules with reasonable computational costs – what is important property for using it as a component inside combined classiﬁers. 3.1

MODLEM Algorithm

The general schema of the MODLEM algorithm is brieﬂy presented below. More detailed description could be found in [14,35,37]. This algorithm is based on the idea of a sequential covering and it generates a minimal set of decision rules for every decision concept (decision class or its rough approximation in case of inconsistent examples). Such a minimal set of rules (also called local covering [11]) attempts to cover all positive examples of the given decision concept, further denoted as B, and not to cover any negative examples (i.e. U \ B). The main procedure for rule induction scheme starts from creating a ﬁrst rule by choosing sequentially the ‘best’ elementary conditions according to chosen criteria (see the function Find best condition). When the rule is stored, all learning positive examples that match this rule are removed from consideration. The process

334

J. Stefanowski

is repeated while some positive examples of the decision concept remain still uncovered. Then, the procedure is sequentially repeated for each set of examples from a succeeding decision concept. In the MODLEM algorithm numerical attributes are handled during rule induction while elementary conditions of rules are created. These conditions are represented as either (a < va ) or (a ≥ va ), where a denotes an attribute and va is its value. If the same attribute is chosen twice while building a single rule, one may also obtain the condition (a = [v1 , v2 )) that results from an intersection of two conditions (a < v2 ) and (a ≥ v1 ) such that v1 < v2 . For nominal attributes, these conditions are (a = va ) or could be extended to the set of values. Procedure MODLEM (input B - a set of positive examples from a given decision concept; criterion - an evaluation measure; output T – single local covering of B, treated here as rule condition parts) begin G := B; {A temporary set of rules covered by generated rules} T := ∅; while G = ∅ do {look for rules until some examples remain uncovered} begin T := ∅; {a candidate for a rule condition part} S := U ; {a set of objects currently covered by T } while (T = ∅) or (not([T ] ⊆ B)) do {stop condition for accepting a rule} begin t := ∅; {a candidate for an elementary condition} for each attribute q ∈ C do {looking for the best elementary condition} begin new t :=Find best condition(q, S); if Better(new t, t, criterion) then t := new t; {evaluate if a new condition is better than previous one according to the chosen evaluation measure} end; T := T ∪ {t}; {add the best condition to the candidate rule} S := S ∩ [t]; {focus on examples covered by the candidate} end; { while not([T ] ⊆ B } for each elementary condition t ∈ T do if [T − t] ⊆ B then T := T − {t}; {test a rule minimality} T := T ∪ {T }; {store a rule} G := B − T ∈T [T ] ; {remove already covered examples} end; { while G = ∅ } for each T ∈ T do if T ∈T −T [T ] = B then T := T − T {test minimality of the rule set} end {procedure}

Ë

Ë

function Find best condition (input c - given attribute; S - set of examples; output best t - bestcondition) begin best t := ∅; if c is a numerical attribute then

On Combined Classiﬁers, Rule Induction and Rough Sets

335

begin H:=list of sorted values for attribute c and objects from S; { H(i) - ith unique value in the list } for i:=1 to length(H)-1 do if object class assignments for H(i) and H(i + 1) are diﬀerent then begin v := (H(i) + H(i + 1))/2; create a new t as either (c < v) or (c ≥ v); if Better(new t, best t, criterion) then best t := new t ; end end else { attribute is nominal } begin for each value v of attribute c do if Better((c = v), best t, criterion) then best t := (c = v) ; end end {function}.

For the evaluation measure (i.e. a function Better) indicating the best condition, one can use either class entropy measure or Laplacian accuracy. For their deﬁnitions see [14] or [23]. It is also possible to consider a lexicographic order of two criteria measuring the rule positive cover and, then, its conditional probability (originally considered by Grzymala in his LEM2 algorithm or its last, quite interesting modiﬁcation called MLEM). In all experiments, presented further in this paper, we will use the entropy as an evaluation measure. Having the best cut-point we choose a condition (a < v) or (a ≥ v) that covers more positive examples from the concept B. In a case of nominal attributes it is also possible to use another option of Find best condition function, where a single attribute value in the elementary condition (a = vi ) is extended to a multi-valued set (a ∈ Wa ), where Wa is a subset of values from the attribute domain. This set is constructed in the similar way as in techniques for inducing binary classiﬁcation trees. Moreover, the author created MODLEM version with another version of rule stop condition. Let us notice that in the above schema the candidate T is accepted to become a rule if [T ] ⊆ B, i.e. a rule should cover learning examples belonging to an appropriate approximation of the given class Kj . For some data sets – in particular noisy ones – using this stop condition may produce too speciﬁc rules (i.e. containing many elementary conditions and covering too few examples). In such situations the user may accept partially discriminating rules with high enough accuracy – this could be done by applying another stop condition ([T ∩ B]/[T ] ≥ α. An alternative is to induce all, even too speciﬁc rules and to post-process them – which is somehow similar to pruning of decision trees. Finally we can illustrate the use of MODLEM by a simple example. The data table contains examples of 17 decision concerning classiﬁcation of some customers into three classes coded as d, p, r. All examples are described by 5 qualitative and numerical attributes.

336

J. Stefanowski Table 1. A data table containing examples of customer classiﬁcation Age Job Period Income Purpose Decision m sr m st sr m sr m sr st m m sr m st m m

u p p p p u b p p e u b p b p p b

0 2 4 16 14 0 0 3 11 0 0 0 17 0 21 5 0

500 1400 2600 2300 1600 700 600 1400 1600 1100 1500 1000 2500 700 5000 3700 800

K S M D M W D D W D D M S D S M K

r r d d p r r p d p p r p r d d r

This data table is consistent, so lower and upper approximations are the same. The use of MODLEM results in the following set of certain rules (square brackets contain the number of learning examples covered by the rule): rule rule rule rule rule rule rule

1. 2. 3. 4. 5. 6. 7.

if (Income < 1050) then (Dec = r) [6] if (Age = sr) ∧ (P eriod < 2.5) then (Dec = r) [2] if (P eriod ∈ [3.5, 12.5)) then (Dec = d) [2] if (Age = st) ∧ (Job = p) then (Dec = d) [3] if (Age = m) ∧ (Income ∈ [1050, 2550)) then (Dec = p) [2] if (Job = e) then (Dec = p) [1] if (Age = sr) ∧ (P eriod ≥ 12.5) then (Dec = p) [2]

Due to the purpose and page limits of this paper we do not show details of MODLEM working steps while looking for a single rule - the reader is referred to the earlier author’s papers devoted to this topic only. 3.2

Classiﬁcation Strategies

Using rule sets to predict class assignment for an unseen object is based on matching the object description to condition parts of decision rules. This may result in unique matching to rules from the single class. However two other ambiguous cases are possible: matching to more rules indicating diﬀerent classes or the object description does not match any of the rules. In these cases, it is necessary to apply proper strategies to solve these conﬂict cases. Review of diﬀerent strategies is given in [37]

On Combined Classiﬁers, Rule Induction and Rough Sets

337

In this paper we employ two classiﬁcation strategies. The ﬁrst was introduced by Grzymala in LERS [12]. The decision to which class an object belongs to is made on the basis of the following factors: strength and support. The Strength is the total number of learning examples correctly classiﬁed by the rule during training. The support is deﬁned as the sum of scores of all matching rules from the class. The class Kj for which the support, i.e., the following expression Strength f actor(R) matching rules R describing Ki

is the largest is the winner and the object is assigned to Kj . If complete matching is impossible, all partially matching rules are identiﬁed. These are rules with at least one elementary condition matching the corresponding object description. For any partially matching rule R, the factor, called Matching factor (R), deﬁned as a ratio of matching conditions to all conditions in the rule, is computed. In partial matching, the concept Kj for which the following expression is the largest M atching f actor(R) ∗ Strength f actor(R) partially matching rules R

is the winner and the object is classiﬁed as being a member of Kj . The other strategy was introduced in [32]. The main diﬀerence is in solving no matching case. It is proposed to consider, so called, nearest rules instead of partially matched ones. These are rules nearest to the object description in the sense of chosen distance measure. In [32] a weighted heterogeneous metric DR is used which aggregates a normalized distance measure for numerical attributes and {0;1} diﬀerences for nominal attributes. Let r be a nearest matched rule, e denotes a classiﬁed object. Then DR(r, e) is deﬁned as: Dr(r, e) =

1 p 1/p ( da ) m a∈P

where p is a coeﬃcient equal to 1 or 2, m is the number of elementary conditions in P – a condition part of rule r. A distance da for numerical attributes is equal to |a(e) − vai |/|va−max − va−min |, where vai is the threshold value occurring in this elementary condition and va−max , va−min are maximal and minimal values in the domain of this attribute. For nominal attributes present in the elementary condition, distance da is equal to 0 if the description of the classiﬁed object e satisﬁes this condition or 1 otherwise. The coeﬃcient expressing rule similarity (complement of the calculated distance, i.e. 1−DR(r, e)) is used instead of matching factor in the above formula and again the strongest decision Kj wins. While computing this formula we can use also heuristic of choosing the ﬁrst k nearest rules only. More details on this strategy the reader can ﬁnd in papers [32,33,37]. Let us consider a simple example of classifying two objects e1 = {(Age = m), (Job = p), (P eriod = 6), (Income = 3000), (P urpose = K)} and e2 = {(Age = m), (Job = p), (P eriod = 2), (Income = 2600), (P urpose = M )}. The

338

J. Stefanowski

ﬁrst object is completely matched by to one rule no. 3. So, this object is be assigned to class d. The other object does not satisfy condition part of any rules. If we use the ﬁrst strategy for solving no matching case, we can notice that object e2 is partially matched to rules no. 2, 4 and 5. The support for class r is equal to 0.5·2 = 1. The support for class d is equal to 0.5·2 + 0.5·2 = 2. So, the object is assigned to class d. 3.3

Summarizing Experience with Single MODLEM Classiﬁers

Let us shortly summarize the results of studies, where we evaluated the classiﬁcation performance of the single rule classiﬁer induced by MODLEM. There are some options of using this algorithm. First of all one can choose as decision concepts either lower or upper approximations. We have carried out several experimental studies on benchmark data sets from ML Irvine repository [3]. Due to the limited size of this paper, we do not give precise tables but conclude that generally none of approximations was better. The diﬀerences of classiﬁcation accuracies were usually not signiﬁcant or depended on the particular data at hand. This observation is consistent with previous experiments on using certain or possible rules in the framework of LEM2 algorithm [13]. We also noticed that using classiﬁcation strategies while solving ambiguous matching was necessary for all data sets. Again the diﬀerence of applied strategies in case of non-matching (either Grzymala’s proposal or nearest rules) were not signiﬁcant. Moreover, in [14] we performed a comparative study of using MODLEM and LEM2 algorithms on numerical data. LEM2 was used with preprocessing phase with the good discretization algorithm. The results showed that MODLEM can achieved good classiﬁcation accuracy comparable to best pre-discretization and LEM2 rules. Here, we could comment that elements of rough sets are mainly used in MODLEM as a kind of preprocessing, i.e. approximations are decision concepts. Then, the main procedure of this algorithm follows rather the general inductive principle which is common aspect with many machine learning algorithms – see e.g. a discussion of rule induction presented in [23]. Moreover, the idea of handling numerical attributes is somehow consistent with solutions also already present in classiﬁcation tree generation. In this sense, other rule generation algorithms popular in rough sets community, as e.g. based on Boolean reasoning, are more connected with rough sets theory. It is natural to compare performance of MODLEM induced rules against standard machine learning systems. Such a comparative study was carried out in [37,41] and showed that generally the results obtained by MODLEM (with nearest rules strategies) were very similar to ones obtained by C4.5 decision tree.

4

Combined Classiﬁers – General Issues

In the next sections we will study the use of MOLDEM in the framework of the combined classiﬁers. Previous theoretical research (see, e.g., their summary in [6,45]) indicated that combining several classiﬁers is eﬀective only if there is

On Combined Classiﬁers, Rule Induction and Rough Sets

339

a substantial level of disagreement among them, i.e. if they make errors independently with respect to one another. In other words, if they make errors for a given object they should indicate diﬀerent class assignments. Diversiﬁed base classiﬁers can be generated in many ways, for some review see, e.g. [6,36,45]. In general, either homogeneous or heterogeneous classiﬁers are constructed. In the ﬁrst category, the same learning algorithm is used over diﬀerent samples of the data set. The best-known examples are either bagging and boosting techniques which manipulate set of examples by including or weighting particular examples, or methods that manipulate set of attributes, e.g. randomly choosing several attribute subsets. Moreover, multiple classiﬁers could be trained over diﬀerent samples or partitions of data sets. In the second category, diﬀerent learning algorithms are applied to the same data set, and the diversity of results comes from heterogeneous knowledge representations or diﬀerent evaluation criteria used to construct them. The stacked generalization or meta-learning belong to this category. In section 7 we study the combiner as one of these methods. Combining classiﬁcation predictions from single classiﬁers is usually done by group or specialized decision making. In the ﬁrst method all base classiﬁers are consulted to classify a new object while the other method chooses only these classiﬁers whose are expertised for this object. Voting is the most common method used to combine single classiﬁers. The vote of each classiﬁer may be weighted, e.g., by an evaluation of its classiﬁcation performance. Moreover, looking into the rough sets literature one can notice a growing research interest in constructing more complex classiﬁcation system. First works concerned rather an intelligent integration of diﬀerent algorithms into hybrid system. For instance, some researchers tried to reﬁne rule classiﬁers by analysing relationships with neural networks [44]. More related works included an integration of k - nearest neighbor with rough sets rule generation, see e.g. RIONA system, which oﬀered good classiﬁcation performance [8]. Yet another approach comprises two level knowledge representation: rules induced by Explore representing general patterns in data and case base representing exceptions [36], which worked quite well for the diﬃcult task of credit risk prediction [43]. Recently Skowron and his co-operators have been developing hierarchical classiﬁers which attempt at approximating more complex concepts [2]. Classiﬁers on different hierarchy level correspond to diﬀerent levels of pattern generalization and seems to be a speciﬁc combination of multiple models, which could be obtained in various ways, e.g. using a special lattice theory [46] or leveled rule generation. Nguyen et al. described in [24] an application concerning detecting sunspots where hierarchical classiﬁer is constructed with a domain knowledge containing an ontology of considered concepts.

5

Using MODLEM Inside the Bagging

Firstly, we consider the use of MODLEM induced classiﬁer inside the most popular homogeneous multiple classiﬁers [38].

340

J. Stefanowski

This approach was originally introduced by Breiman [4]. It aggregates classiﬁers generated from diﬀerent bootstrap samples. The bootstrap sample is obtained by uniformly sampling with replacement objects from the training set. Each sample has the same size as the original set, however, some examples do not appear in it, while others may appear more than once. For a training set with m examples, the probability of an example being selected at least once is 1−(1−1/m)m. For a large m, this is about 1 - 1/e. Given the parameter R which is the number of repetitions, R bootstrap samples S1 , S2 , . . . , SR are generated. From each sample Si a classiﬁer Ci is induced by the same learning algorithm and the ﬁnal classiﬁer C ∗ is formed by aggregating these R classiﬁers. A ﬁnal classiﬁcation of object x is built by a uniform voting scheme on C1 , C2 , . . . , CR , i.e. is assigned to the class predicted most often by these sub-classiﬁers, with ties broken arbitrarily. For more details and theoretical justiﬁcation see e.g. [4]. Table 2. Comparison of classiﬁcation accuracies [%] obtained by the single MODLEM based classiﬁer and the bagging approach; R denotes the number of component classiﬁers inside bagging Name of data set

Single classiﬁer

Bagging

R

bank buses zoo hepatitis hsv iris automobile segmentation glass bricks vote bupa election urology german crx pima

93.81 ± 0.94 97.20 ± 0.94 94.64 ± 0.67 78.62 ± 0.93 54.52 ± 1.05 94.93 ± 0.5 85.23 ± 1.1 85.71 ± 0.71 72.41 ± 1.23 90.32* ± 0.82 92.67 ± 0.38 65.77 ± 0.6 88.96± 0.54 63.80 ± 0.73 72.16 ± 0.27 84.64 ± 0.35 73.57 ± 0.67

95.22 ± 1.02 99.54 ± 1.09 93.89* ± 0.71 84.05 ± 1.1 64.78 ± 0.57 95.06* ± 0.53 83.00 ±0.99 87.62 ± 0.55 76.09 ± 0.68 91.21* ± 0.48 96.01 ± 0.29 76.28 ± 0.44 91.66 ± 0.34 67.40 ± 0.46 76.2 ± 0.34 89.42 ± 0.44 77.87 ± 0.39

7 5 7 5 7 5 5 7 10 7 10 5 7 7 5 10 7

In this paper we shortly summarize main results obtained in the extensive computational study [38]. The MODLEM algorithm was applied to generate base classiﬁers in the bagging combined classiﬁer. In table 2 we present the comparison of the classiﬁcation accuracy obtained for the best variant of the bagging against the single rule classiﬁer (also induced by MODLEM). The experiments were carried out on several data sets coming mainly from ML Irvine repository [3]. For each data set, we show the classiﬁcation accuracy obtained by a single classiﬁer over the 10 cross-validation loops. A standard deviation is also given. An asterisk

On Combined Classiﬁers, Rule Induction and Rough Sets

341

indicates that the diﬀerence for these compared classiﬁers and a given data set is not statistically signiﬁcant (according to two-paired t-Student test with α=0.05). The last column presents the number of R component classiﬁers inside the bagging - more details on tuning this value are described in [38]. We conclude that results of this experiment showed that the bagging significantly outperformed the single classiﬁer on 14 data sets out of total 18 ones. The diﬀerence between classiﬁers were non-signiﬁcant on 3 data sets (those which were rather easy to learn as, e.g. iris and bricks - which were characterized by a linear separation between classes). Moreover, we noticed the slightly worse performance of the bagging for quite small data (e.g. buses, zoo - which seemed to be too small for sampling), and signiﬁcantly better for data sets containing a higher number of examples. For some of these data sets we observed an substantial increase of predictive accuracy, e.g. for hsv – over 10%, bupa – around 10% and hepatitis – 5.43%. However, we should admit that this good performance was expected as we know that there are many previous reports on successful use of decision trees in bagging or boosting.

6

On Solving Multiclass Problems with the n2 -Classiﬁer

One can say the bagging experiment has been just a variant of a standard approach. Now we will move to more original approach, called the n2 -classiﬁer, which was introduced by Jelonek and author in [18,36]. This kind of a multiple classiﬁer is a specialized approach to solve multiple class learning problems. The n2 -classiﬁer is composed of (n2 − n)/2 base binary classiﬁers (where n is a number of decision classes; n > 2). The main idea is to discriminate each pair of the classes: (i, j), i, j ∈ [1..n], i = j , by an independent binary classiﬁer Cij . Each base binary classiﬁer Cij corresponds to a pair of two classes i and j only. Therefore, the speciﬁcity of the training of each base classiﬁer Cij consists in presenting to it a subset of the entire learning set that contains only examples coming from classes i and j. The classiﬁer Cij yields a binary classiﬁcation indicating whether a new example x belongs to class i or to class j. Let us denote by Cij (x) the classiﬁcation of an example x by the base classiﬁer Cij . The complementary classiﬁers: Cij and Cji (where i, j ∈ < 1 . . . n >; i = j) solve the same classiﬁcation problem – a discrimination between class i-th and j-th. So, they are equivalent (Cij ≡ Cji ) and it is suﬃcient to use only (n2 - n)/2 classiﬁers Cij (i < j), which correspond to all combinations of pairs of n classes. An algorithm providing the ﬁnal classiﬁcation assumes that a new example x is applied to all base classiﬁers Cij . As a result, their binary predictions Cij (x) are computed. The ﬁnal classiﬁcation is obtained by an aggregation rule, which is based on ﬁnding a class that wins the most pairwise comparisons. The more sophisticated approach includes a weighted majority voting rules, where the vote of each classiﬁer is modiﬁed by its credibility, which is calculated as its classiﬁcation performance during learning phase; more details in [18].

342

J. Stefanowski

We have to remark that the similar approach was independently studied by Friedman [7] and by Hastie and Tibshirani [17] – they called it classiﬁcation by pairwise coupling. The experimental studies, e.g. [7,17,18], have shown that such multiple classiﬁers performed usually better than the standard classiﬁers. Previously the author and J.Jelonek have also examined the inﬂuence of a learning algorithm on the classiﬁcation performance of the n2 -classiﬁer.

Table 3. Comparison of classiﬁcation accuracies [%] and computation times [s] for the single MODLEM based classiﬁer and the n2 -classiﬁer also based on decision rules induced by MODLEM algorithm Accuracy of Name of single data set MODLEM (%) automobile 85.25 ± 1.3 cooc 55.57 ± 2.0 ecoli 79.63 ± 0.8 glass 72.07 ± 1.2 hist 69.36 ± 1.1 meta-data 47.2 ± 1.3 iris 94.2 ± 0.6 soybean-large 91.09 ± 0.9 vowel 81.81 ± 0.5 yeast 54.12 ± 0.7 zoo 94.64 ± 0.5

Accuracy of n2M ODLEM (%) 87.96 ± 1.5 59.30 ± 1.4 81.34 ± 1.7 74.82 ± 1.4 73.10 ± 1.4 49.83 ± 1.9 95.53* ± 1.2 91.99* ± 0.8 83.79 ± 1.2 55.74 ± 0.9 94.46* ± 0.8

Time of comput. MODLEM 15.88 ± 0.4 4148,7 ± 48.8 27.53 ± 0.5 45.29 ± 1.1 3563.79 ± 116.1 252.59 ± 78.9 0.71 ± 0.04 26.38 ± 0.3 3750.57 ± 30.4 1544.3 ± 13.2 0.30 ± 0.02

Time of comput. n2M ODLEM 5.22 ± 0.3 431.51 ± 1.6 11.25 ± 0.7 13.88 ± 0.4 333.96 ± 0.8 276.71 ± 5.21 0.39 ± 0.04 107.5 ± 5.7 250.63 ± 0.7 673.82 ± 9.4 0.34 ± 0.12

Here, we summarize these of our previous results, where the MODLEM was applied to generate base classiﬁers inside the n2 -classiﬁer [38]. In table 3 we present classiﬁcation accuracies obtained by the n2 -classiﬁer and compare them against the single rule classiﬁer induced by MODLEM on 11 data sets, all concerning multiple-class learning problems, with a number of classes varied from 3 up to 14. The second and third columns are presented in a similar way as in Table 2. These results showed that the n2 -classiﬁer signiﬁcantly (again in the sense of paired t test with a signiﬁcance level α = 0.05) outperformed the single classiﬁer on 7 out of 11 problems, e.g. for hist – over 3.7%, glass – around 2.7%, automobile – 2.5% and meta-data – 2.6%. These improvements were not so high as in the bagging but still they occurred for many diﬃcult multi-class problems. Again, the multiple classiﬁer was not useful for easier problems (e.g. iris). Moreover, we noticed that its performance was better for data sets with a higher number of examples. Coming back to our previous results for the n2 -classiﬁer [18] we can again remark that the comparable classiﬁcation improvements were observed for the case of using decision trees. Then, let us focus our attention on interesting phenomena concerning computation costs of using the MODLEM in a construction of the n2 -classiﬁer. Table 3 (two last columns) contains computation times (in seconds calculated as average

On Combined Classiﬁers, Rule Induction and Rough Sets

343

values over 10 folds with standard deviations). We can notice that generally constructing a combined classiﬁers does not increase the computation time. What is even more astonishing, for some data sets constructing the n2 -classiﬁer requires even less time than training the standard single classiﬁer. Here, we have to stress that in our previous works [18,37] we noticed that the increase of classiﬁcation accuracy (for other learning algorithms as e.g. decision trees, k-nearest neighbor or neural networks) was burden with increasing the computational costs (sometimes quite high). In [38] we attempted to explain the good performance of MODLEM inside the n2 -classiﬁer. Shortly speaking, the n2 -classiﬁer should be rather applied to solving diﬃcult (”complex”) classiﬁcation tasks, where examples of decision classes are separated by non-linear decision borders – these are often diﬃcult concepts to be learned by standard classiﬁers, while pairwise decision boundaries between each pair of classes may be simpler and easier to be learned with using a smaller number of attributes. Here, MODLEM could gain its performance thanks to his sequential covering and greedy heuristic search. It generates rules distinguishing smaller number of learning examples (from two classes only) than in the multiple class case and, above all, testing a smaller number of elementary conditions. To verify hypothesis we inspect syntax of rule sets induced by the single classiﬁer and the n2 -classiﬁer. Rules for binary classiﬁers were using less attributes and covered more learning example on average than rules from the single set generated in the standard way [38].

7

Combining Predictions of Heterogeneous Classiﬁers

In two previous sections we described the use of MODLEM based classiﬁers inside the architecture of homogeneous classiﬁers. In these solutions, the MODLEM was the only algorithm applied to create base classiﬁers inside multiple classiﬁers and could directly inﬂuence their ﬁnal performance. Diversiﬁcation of base classiﬁers is one of the conditions for improving classiﬁcation performance of the ﬁnal system. Let us repeat that in previously considered solutions it was achieved by changing the distribution of examples in the input data. Another method to obtain component classiﬁer diversity is constructing, so called, heterogeneous classiﬁers. They are generated from the same input data by diﬀerent learning algorithms which use diﬀerent representation language and search strategies. These base classiﬁers could be put inside a layered architecture. At the ﬁrst level base classiﬁers receive the original data as input. Their predictions are then aggregated at the second level into the ﬁnal prediction of the system. This could be done in various ways. In one of our studies we used a solution coming from Chan & Stolfo [5], called a combiner. The combiner is based on an idea of merging predictions of base classiﬁers by an additional classiﬁer, called meta-classiﬁer. This is constructed in an extra meta-learning step, i.e. ﬁrst base classiﬁers are learned, then their predictions made on a set of extra validation examples, together with correct decision labels, form a meta-level training set. An extra learning algorithm is applied to this set to discover how to merge base classiﬁer predictions into a ﬁnal decision.

344

J. Stefanowski Table 4. Classiﬁcation accuracies [%] for diﬀerent multiple classiﬁers Data set

Bagging n2 -classiﬁer Combiner

Automobile Bank Bupa Ecoli Glass HSV Meta-data Pima Voting Yeast Zoo

83.00 95.22 76.28 85.70 74.82 64.75 48.11 75.78 93.33 58.18 93.89

87.90 – – 81.34 74.82 – 49.80 – – 55.74 94.46

84.90 95.45 69.12 85.42 71.50 59.02 51.33 74.78 94.67 58.36 95.05

In [41] we performed a comparative study of using a combiner approach against the single classiﬁers learned by these algorithms which were applied to create its component classiﬁers. In this study base classiﬁers were induced by k-NN, C4.5 and MODLEM. The meta-classiﬁer was a Naive Bayes. This comparative study was performed on 15 data sets. However, the obtained results showed that the combiner did not improve classiﬁcation accuracy in so many cases as previously studied homogeneous classiﬁers. Only in 33% data we observed a signiﬁcant improvement comparing against single classiﬁers. In table 4 we present only some of these results concerning the ﬁnal evaluation of the combiner compared also against the previous multiple classiﬁers. However, while comparing these classiﬁers we should be cautious as the number of the results on common data sets was limited. Moreover, MODLEM is only one of three component classiﬁers inside the combiner that inﬂuences the ﬁnal result. We could also ask a question about other elements of the architecture of heterogeneous classiﬁer, e.g. number of component classiﬁers or the aggregation techniques. In recent experiments we focus our interest on testing two other techniques instead of the meta-combiner: – a simple aggregation performed by means of a majority voting rule (denoted as MV in table 4), – using a quite sophisticated approach – SCANN; It was introduced by Merz [22] and uses a mechanism of the correspondence analysis to discover hidden relationships between the learning examples and the classiﬁcation done by the component classiﬁers. Results from ongoing experiments are given in Table 5. There is also a diﬀerence to previous architecture, i.e. adding an additional, forth component classiﬁers Naive Bayesian at the ﬁrst level. We can remark that the more advanced aggregation technique could slightly increase the classiﬁcation accuracy comparing to simpler one. On the other hand they are much time consuming.

On Combined Classiﬁers, Rule Induction and Rough Sets

345

Table 5. Comparison of diﬀerent methods producing the ﬁnal decision inside the heterogeneous classiﬁers - classiﬁcation accuracies [%] Data set credit-a 86.2 glass 68.5 ecoli 86.1 zoo 95

8

MV ± 0.6 ± 0.3 ± 0.9 ± 0.9

SCANN Combiner 87 70.1 81.5 92.2

± 0.7 ± 0.2 ± 0.8 ± 0.7

86.6 70.5 84.5 95.1

± 0.4 ± 0.6 ± 0.5 ± 0.4

Discussion of Results and Final Remarks

As Professor Zdzislaw Pawlak wrote in the introductory chapter of his book on rough sets [26] knowledge of human beings and other species is strictly connected with their ability to classify objects. Finding classiﬁcation patterns of sensor signals or data form fundamental mechanisms for very living being. In his point of view it was then connected with a partition (classiﬁcation) operation leading to basic blocks for constructing knowledge. Many researchers followed the Pawlak’s idea. One of the main research directions includes constructing approximations of knowledge from tables containing examples of decisions on object classiﬁcation. Rules were often induced as the most popular knowledge representation. They could be used either to describe the characteristics of available data or as the basis for supporting classiﬁcation decisions concerning new objects. Up to now several eﬃcient rule classiﬁers have been introduced. In this study we have attempted to brieﬂy describe the current experience with using the author’s rule induction algorithm MODLEM, which induces either certain or possible rules from appropriate rough approximations. This is the main point where elements of the rough sets theory is applied in this algorithm. Given as an input learning examples from approximations, the rule generation phase follows the general idea of sequential covering, which is somehow in common with machine learning paradigms. The MODLEM produces a minimal set of rules covering examples from rough approximations. This rule sets should be joined with classiﬁcation strategies for solving ambiguous matching of the new object description to condition parts of rules. An extra property of this algorithm is it ability to handle directly numerical attributes without prior discretization. The current experience with comparative studies on benchmark data sets and real life applications showed that the classiﬁcation performance of this approach was comparable to other symbolic classiﬁers, in particular to decision trees. Although the MODLEM classiﬁer and other machine learning approaches are eﬃcient for many classiﬁcation problems, they do not always lead to satisfactory classiﬁcation accuracy for more complex and diﬃcult problems. This is our motivation to consider new approaches for increasing classiﬁcation accuracy by combining several classiﬁers into an integrated system. Several proposals of

346

J. Stefanowski

constructing such multiple classiﬁers are already proposed. Most of them are general approaches, where many diﬀerent algorithms could be applied to induce the component classiﬁers. Thus, our main research interest in this study is to summarize our experiments with using MODLEM induced rule classiﬁers inside the framework of three diﬀerent multiple classiﬁers, namely the bagging, the n2 -classiﬁer and the combiner. A classiﬁcation accuracy for the multiple classiﬁer has been compared against the standard classiﬁers – also induced by MODLEM. These results and their detailed discussion has been given in the previous sections. Firstly we could notice that using MODLEM inside the bagging was quite eﬀective. However, it was a kind of standard approach and we could expect such good performance as MODLEM performs similarly to decision trees (which have been extensively studied in the bagging) and could be seen as unstable learning algorithm - i.e. an algorithm whose output classiﬁer undergoes changes in response to small changes in the training data. This kind of algorithm may produce base classiﬁers diversiﬁed enough (but not too much, see e.g. discussion of experimental study by Kuncheva and Whitaker [21]) which is a necessary condition for their eﬀective aggregation. Following the same arguments we also suspect that MODLEM should nicely work inside the boosting classiﬁer. Further on, we could hypothesize that slightly worse improvements of the classiﬁcation accuracy in the combiner approach may result from insuﬃcient diversiﬁcation of component heterogeneous classiﬁers. This has been veriﬁed by analysing distributions of wrong decisions for base classiﬁers, presented in [41]. It showed the correlation of errors for some data sets, where ﬁnally we did not notice the improvement of the classiﬁcation accuracy. The most original methodological approach is Jelonek and author’s proposal of the n2 -classiﬁer which is in fact a specialized approach to learning multiple class problems. The n2 -classiﬁer is particularly well suited for multiple class data where exist ”simpler” pairwise decision boundaries between pair of classes. MODLEM seems to be a good choice to be used inside this framework as it leads to an improvement of classiﬁcation performance and does not increase computational costs - reasons for this have been discussed in section 7. Let us notice that using other learning algorithms inside the n2 -classiﬁer and applying MODLEM in two other multiple classiﬁer requires an extra computation eﬀorts comparing to learning the single, standard classiﬁer [38]. Comparing results of all together multiple classiﬁers ”head to head” we should be cautious as we had a limited number of common data sets. It seems that the n2 -classiﬁer is slightly better for these data. While the standard multiple classiﬁers, as bagging or combiner, are quite eﬃcient for simpler data and are easier to be implemented. To sum up, the results of our experiments have shown that the MODLEM algorithm can be eﬃciently used within the framework of three multiple classiﬁers for data sets concerning more ”complex” decision concepts. However, the relative merits of these new approaches depends on the speciﬁes of particular problems and a training sample size.

On Combined Classiﬁers, Rule Induction and Rough Sets

347

Let us notice that there is a disadvantage of the multiple classiﬁers - loosing a simple and easy interpretable structure of knowledge represented in a form decision rules. These are ensembles of diversiﬁed rule sets specialized for predictive aims not one set of rules in a form for a human inspection. As to future research directions we could consider yet another way of obtaining diversiﬁed data – i.e. selecting diﬀerent subsets of attributes for each component classiﬁers. The author has already started research on extending bootstrap samples inside the bagging by applying additionally attribute selection [39,40]. In this way each bootstrap is replicated few times, each of them using diﬀerent subset of attributes. We have considered the use of diﬀerent selection techniques and observed that besides random choice or wrapper model, techniques which use either entropy based measures or correlation merits are quite useful. The results of comparative experiments carried out in [40] have showed that the classiﬁcation accuracy of such a new extended bagging is higher than for standard one. In this context one could come back to the classical rough sets topic of reducts, which relates to ﬁnding an ensemble of few attribute subsets covering diﬀerent data properties and constructing in this way a set of diversiﬁed examples for an integrated system. However, we are not limited to ”classical” meaning of pure rough sets reducts but rather to approximate ones, where the entropy measure is also considered [30]. Acknowledgment. The author would like to thank his colleagues Jacek Jelonek, Slawomir Nowaczyk and his M.Sc. students Michal Bro´ nczyk, Ryszard Gizelski, Maciej L uszczy´ nski who have worked with him on the software implementations of the classiﬁers or took part in some experiments.

References 1. Bazan J.: A comparison of dynamic and non-dynamic rough set methods for extracting laws from decision tables. In Polkowski L., Skowron A. (eds.), Rough Sets in Data Mining and Knowledge Discovery vol. 1, Physica-Verlag, 1998, 321–365. 2. Bazan J., Nguyen Hung Son, Skowron A.: Rough sets methods in approximation of hierarchical concepts. In Proc. of the Conference on Rough Sets and New Trends in Computing, RSCTC – 2004, LNAI 2066, Springer Verlag, 2004, 346–355. 3. Blake C., Koegh E., Mertz C.J.: Repository of Machine Learning, University of California at Irvine (1999). 4. Breiman L.: Bagging predictors. Machine Learning, 24 (2), 1996, 123–140. 5. Chan P.K., Stolfo S.: On the accuracy of meta-learning for scalable data mining. Journal of Intelligent Information Systems, 8, (1), 1997, 5-28. 6. Dietrich T.G.: Ensemble methods in machine learning. In Proc. of 1st Int. Workshop on Multiple Classiﬁer Systems, 2000, 1–15. 7. Friedman J.: Another approach to polychotomous classiﬁcation, Technical Report, Stanford University, 1996. 8. G´ ora G., Wojna A.: RIONA: a new classiﬁcation system combining rule induction and instance based learning. Fundamenta Informaticae 51 (4), 2002, 369-390. 9. Greco S., Matarazzo B., Slowi´ nski R.: The use of rough sets and fuzzy sets in MCDM. In Gal T., Stewart T., Hanne T. (eds), Advances in Multiple Criteria Decision Making, Kluwer, chapter 14, 1999, pp. 14.1-14.59.

348

J. Stefanowski

10. Greco S., Matarazzo B., Slowi´ nski R., Stefanowski J.: Variable consistency model of dominance-based rough set approach. In Proc. 2nd Int. Conference on Rough Sets and New Trends in Computing, RSCTC – 2000, LNAI 2005, Springer Verlag, 2001,170–181. 11. Grzymala-Busse J.W. LERS - a system for learning from examples based on rough sets. In Slowinski R. (ed.), Intelligent Decision Support, Kluwer Academic Publishers, 1992, 3–18. 12. Grzymala-Busse J.W.: Managing uncertainty in machine learning from examples. In Proc. 3rd Int. Symp. in Intelligent Systems, Wigry, Poland, IPI PAN Press, 1994, 70–84. 13. Grzymala-Busse J.W. Zou X.: Classiﬁcation strategies using certain and possible rules. In Proceedings of the 1th Rough Sets and Current Trends in Computing Conference, RSCTC–98 , LNAI 1424, Springer Verlag, 1998, 37-44. 14. Grzymala-Busse J.W., Stefanowski J.: Three approaches to numerical attribute discretization for rule induction. International Journal of Intelligent Systems, 16 (1), (2001) 29–38. 15. Grzymala-Busse J.W., Stefanowski J. Wilk Sz.: A comparison of two approaches to data mining from imbalanced data. In Proc. of the KES 2004 - 8-th Int. Conf. on Knowledge-based Intelligent Information & Engineering Systems, Springer LNCS vol. 3213, 2004, 757-763. 16. Han J., Kamber M.: Data mining: Concepts and techniques, San Francisco, Morgan Kaufmann, 2000. 17. Hastie T., Tibshirani R.: Classiﬁcation by pairwise coupling. In Jordan M.I. (ed.) Advances in Neural Information Processing Systems: 10 (NIPS-97), MIT Press, 1998, 507-513. 18. Jelonek J., Stefanowski J.: Experiments on solving multiclass learning problems by the n2 -classiﬁer. In Proceedings of 10th European Conference on Machine Learning ECML 98, Springer LNAI no. 1398, 1998, 172–177. ˙ 19. Klosgen W., Zytkow J.M. (eds.): Handbook of Data Mining and Knowledge Discovery, Oxford Press, 2002. 20. Komorowski J., Pawlak Z., Polkowski L. Skowron A.: Rough Sets: tutorial. In Pal S.K., Skowron A. (eds) Rough Fuzzy Hybridization. A new trend in decision making, Springer Verlag, Singapore, 1999, 3–98. 21. Kuncheva L., Whitaker C.J.: Measures of diversity in classiﬁer ensembles and their relationship with the ensemble accuracy. Machine Learning, 51, 2003, 181–207. 22. Merz C.: Using correspondence analysis to combine classiﬁers. Machine Learning, 36 (1/2), 1999, 33–58. 23. Mitchell Tom M.: Machine learning, McGraw Hill, 1997. 24. Nguyen Sinh Hoa, Trung Tham Nguyen, Nguyen Hung Son: Rough sets approach to sunspot classiﬁcation problem. In Proc. of the Conference RSFDGrC – 2005, vol 2, LNAI 3642, Springer Verlag, 2005, 263-272. 25. Pawlak Z.: Rough sets. Int. J. Computer and Information Sci., 11, 1982, 341–356. 26. Pawlak Z.: Rough sets. Theoretical aspects of reasoning about data. Kluwer Academic Publishers, Dordrecht, 1991. 27. Pawlak Z., Grzymala-Busse J., Slowinski R., Ziarko W.: Rough sets. Communications of the ACM, vol. 38, no. 11, 1995, 89-95. 28. Skowron A.: Boolean reasoning for decision rules generation. In Komorowski J., Ras Z. (des.) Methodologies for Intelligent Systems, LNAI 689, Springer-Verlag, 1993, 295–305.

On Combined Classiﬁers, Rule Induction and Rough Sets

349

29. Skowron A., Rauszer C.: The discernibility matrices and functions in information systems. In Slowinski R. (ed.), Intelligent Decision Support. Handbook of Applications and Advances of Rough Set Theory. Kluwer Academic Publishers, 1992, 331–362. 30. Slezak D.: Approximate entropy reducts. Fundamenta Informaticae 53 (3/4), 2002, 365-387. 31. Slowinski R., Greco S.: Inducing Robust Decision Rules from Rough Approximations of a Preference Relation. In Rutkowski L. et al. (eds): Artiﬀcial Intelligence and Soft Computing, LNAI 3070, Springer-Verlag, 2004, 118-132. 32. Stefanowski J.: Classiﬁcation support based on the rough sets. Foundations of Computing and Decision Sciences, vol. 18, no. 3-4, 1993, 371-380. 33. Stefanowski J.: Using valued closeness relation in classiﬁcation support of new objects. In Lin T. Y., Wildberger (eds) Soft computing: rough sets, fuzzy logic, neural networks uncertainty management, knowledge discovery, Simulation Councils Inc., San Diego CA, 1995, 324–327. 34. Stefanowski J.: On rough set based approaches to induction of decision rules. In Polkowski L., Skowron A. (eds), Rough Sets in Data Mining and Knowledge Discovery, vol. 1, Physica-Verlag, 1998, 500–529. 35. Stefanowski J.: The rough set based rule induction technique for classiﬁcation problems. In Proceedings of 6th European Conference on Intelligent Techniques and Soft Computing EUFIT 98, Aachen 7-10 Sept., 1998, 109–113. 36. Stefanowski J.: Multiple and hybrid classiﬁers. In Polkowski L. (ed.) Formal Methods and Intelligent Techniques in Control, Decision Making, Multimedia and Robotics, Post-Proceedings of 2nd Int. Conference, Warszawa, 2001, 174–188. 37. Stefanowski J.: Algorithims of rule induction for knowledge discovery. (In Polish), Habilitation Thesis published as Series Rozprawy no. 361, Poznan Univeristy of Technology Press, Poznan (2001). 38. Stefanowski J.: The bagging and n2-classiﬁers based on rules induced by MODLEM. In Proceedings of the 4th Int. Conference Rough Sets and Current Trends in Computing, RSCTC – 2004, LNAI 3066, Springer-Verlag, 2004, 488-497. 39. Stefanowski J.: An experimental study of methods combining multiple classiﬁers diversiﬁed both by feature selection and bootstrap sampling. In K.T. Atanassov, J. Kacprzyk, M. Krawczak, E. Szmidt (eds), Issues in the Representation and Processing of Uncertain and Imprecise Information, Akademicka Oﬁcyna Wydawnicza EXIT, Warszawa, 2005, 337-354. 40. Stefanowski J., Kaczmarek M.: Integrating attribute selection to improve accuracy of bagging classiﬁers. In Proc. of the AI-METH 2004. Recent Developments in Artiﬁcial Intelligence Methods, Gliwice, 2004, 263-268. 41. Stefanowski J., Nowaczyk S.: On using rule induction in multiple classiﬁers with a combiner aggregation strategy. In Proc. of the 5th Int. Conference on Intelligent Systems Design and Applications - ISDA 2005, IEEE Press, 432-437. 42. Stefanowski J., Vanderpooten D.: Induction of decision rules in classiﬁcation and discovery-oriented perspectives. International Journal of Intelligent Systems 16 (1), 2001, 13–28. 43. Stefanowski J., Wilk S.: Evaluating business credit risk by means of approach integrating decision rules and case based learning. International Journal of Intelligent Systems in Accounting, Finance and Management 10 (2001) 97–114. 44. Szczuka M: Reﬁning classiﬁers with neural networks. International Journal of Intelligent Systems 16 (1), 2001, 39–56.

350

J. Stefanowski

45. Valentini G., Masuli F.: Ensambles of learning machines. In R. Tagliaferri, M. Marinaro (eds), Neural Nets WIRN Vietri-2002, Springer-Verlag LNCS, vol. 2486, 2002 , 3–19. 46. Wang H., Duntsch I., Gediga G., Skowron A.: Hyperrelations in version space. International Journal of Approximate Reasoning, 23, 2000, 111–136. 47. Ziarko W.: Variable precision rough sets model. Journal of Computer and Systems Sciences, vol. 46. no. 1, 1993, 39–59.

Approximation Spaces in Multi Relational Knowledge Discovery Jaroslaw Stepaniuk Department of Computer Science, Bialystok University of Technology Wiejska 45a, 15-351 Bialystok, Poland jstepan@ii.pb.bialystok.pl

Abstract. Pawlak introduced approximation spaces in his seminal work on rough sets more than two decades ago. In this paper, we show that approximation spaces are basic structures for knowledge discovery from multi-relational data. The utility of approximation spaces as fundamental objects constructed for concept approximation is emphasized. Examples of basic concepts are given throughout this paper to illustrate how approximation spaces can be beneﬁcially used in many settings. The contribution of this paper is the presentation of an approximation space-based framework for doing research in various forms of knowledge discovery in multi relational data. Keywords: rough sets, approximation spaces, multi-relational data mining, rough inclusion, uncertainty function.

1

Introduction

Approximation spaces are fundamental structures for the rough set approach [7,8,10]. In this paper we present a generalization of the original approximation space model. Using such approximation spaces we show how the rough set approach can be used for approximation of concepts assuming that only partial information on approximation spaces is available. Hence, searching for concept approximation, i.e., the basic task in machine learning and pattern recognition can be formulated as searching for relevant approximation spaces. Rough set approach has been used in a lot of applications aimed at description of concepts. In most cases, only approximate descriptions of concepts can be constructed because of incomplete information about them. In learning approximations of concepts, there is a need to choose a description language. This choice may limit the domains to which a given algorithm can be applied. There are at least two basic types of objects: structured and unstructured. An unstructured object is usually described by attribute-value pairs. For objects having an internal structure ﬁrst order logic language is often used. Attribute-value languages have the expressive power of propositional logic. These languages sometimes do not allow for proper representation of complex structured objects and relations among objects or their components. The background knowledge that can be J.F. Peters et al. (Eds.): Transactions on Rough Sets VI, LNCS 4374, pp. 351–365, 2007. c Springer-Verlag Berlin Heidelberg 2007

352

J. Stepaniuk

used in the discovery process is of a restricted form and other relations from the database cannot be used in the discovery process. Using ﬁrst-order logic (or FOL for short) has some advantages over propositional logic [1,2,4]. First order logic provides a uniform and very expressive means of representation. The background knowledge and the examples, as well as the induced patterns, can all be represented as formulas in a ﬁrst order language. Unlike propositional learning systems, the ﬁrst order approaches do not require that the relevant data be composed into a single relation but, rather they can take into account data organized in several database relations with various connections existing among them. The paper is organized as follows. In Section 2 we recall the deﬁnition of approximation spaces. Next, we describe a constructive approach for computing values of uncertainty and rough inclusion functions. These functions are the basic components of approximation spaces. Parameters of the uncertainty and rough inclusion functions are tuned in searching for relevant approximation spaces. Among such parameters we distinguish sensory environments and their extensions. These parameters are used for constructive deﬁnition of uncertainty and rough inclusion functions. In Section 3 we discuss notions of relational learning. In Sections 4 and 5 we consider application of rough set methods to discovery of interesting patterns expressed in a ﬁrst order language. In Section 4 rough set methodology is used in the process of translating ﬁrst–order data into attribute– value data. Some properties of this algorithm were presented in [13]. In Section 5 rough set methodology is used in the process of selecting literals which may be a part of a rule. The criterion of selecting a literal is as follows: only such a literal is selected which added to the rule makes the rule discerning most of the examples which were indiscernible so far. Some properties of this algorithm were presented in [14,15].

2

Approximation Spaces

In this section we recall the deﬁnition of an approximation space from [10,13,11]. Deﬁnition 1. A parameterized approximation space is a system AS#,$ = (U, I# , ν$ ), where – U is a non-empty set of objects, – I# : U → P (U ) is an uncertainty function, where P (U ) denotes the power set of U , – ν$ : P (U ) × P (U ) → [0, 1] is a rough inclusion function, and #, $ denote vectors of parameters (the indexes #, $ will be omitted if it does not lead to misunderstanding). 2.1

Uncertainty Function

The uncertainty function deﬁnes for every object x, a set of objects described similarly to x. The set I(x) is called the neighborhood of x (see, e.g., [8,10]).

Approximation Spaces in Multi Relational Knowledge Discovery

353

We assume that the values of the uncertainty function are deﬁned using a sensory environment [11], i.e., a pair (Σ, · U ), where Σ is a set of formulas, called the sensory formulas, and ·U : Σ −→ P (U ) is the sensory semantics. We assume that for any sensory formula α and any object x ∈ U the information whether x ∈ αU holds is available. The set {α : x ∈ αU } is called the signature of x in AS and is denoted by InfAS (x). For any x ∈ U, the set NAS (x) of neighborhoods of x in AS is deﬁned by {αU : x ∈ αU } and from this set the neighborhood I(x) is constructed. For example, I(x) is deﬁned by selecting an element from the set {αU : x ∈ αU } or by I(x) = NAS (x). Observe that any sensory environment (Σ, · U ) can be treated as a parameter of I from the vector # (see Deﬁnition 1). Let us consider two examples. Any decision table DT = (U, A, d) [8] deﬁnes an approximation space ASDT = (U, IA , νSRI ), where, as we will see, IA (x) = {y ∈ U : a(y) = a(x) for all a ∈ A}. Any sensory formula is a descriptor, i.e., a formula of the form a = v where a ∈ A and v ∈ Va with the standard semantics a = vU = {x ∈ U : a(x) = v}. Then, for any x ∈ U its signature Inf ASDT (x) is equal to {a = a(x) : a ∈ A} and the neighborhood IA (x) is equal to NASDT (x). Another example can be obtained assuming that for any a ∈ A there is given a tolerance relation τa ⊆ Va × Va (see, e.g., [10]). Let τ = {τa }a∈A . Then, one can consider a tolerance decision table DTτ = (U, A, d, τ ) with tolerance descriptors a =τa v and their semantics a =τa vU = {x ∈ U : vτa a(x)}. Any such tolerance decision table DTτ = (U, A, d, τ ) deﬁnes the approximation signature InfASDTτ (x) = {a =τa a(x) : a ∈ A} and the space ASDTτ with the neighborhood IA (x) = NASDTτ (x) for any x ∈ U . The fusion of NASDTτ (x) for computing the neighborhood of x can have many diﬀerent forms, the intersection is only an example. For example, to compute the value of I(x) some subfamilies of NAS (x) may ﬁrst be selected and the family consisting of intersection of each such a subfamily is next taken as the value of I(x). 2.2

Rough Inclusion Function

One can consider general constraints which the rough inclusion functions should satisfy. Searching for such constraints initiated investigations resulting in creation and development of rough mereology (see, the bibliography in [9]). In this subsection, we present some examples of rough inclusion functions only. The rough inclusion function ν$ : P (U ) × P (U ) → [0, 1] deﬁnes the degree of inclusion of X in Y , where X, Y ⊆ U . In the simplest case it can be deﬁned by (see, e.g., [10,8]): card(X∩Y ) card(X) if X = ∅ νSRI (X, Y ) = 1 if X = ∅. This measure is widely used by the data mining and rough set communities. It is worth mentioning that Jan L ukasiewicz [3] was the ﬁrst one who used this idea

354

J. Stepaniuk

to estimate the probability of implications. However, rough inclusion can have a much more general form than inclusion of sets to a degree (see, e.g., [9]). Another example of rough inclusion is used for relation approximation [12] and in the variable precision rough set approach [16]. 2.3

Lower and Upper Approximations

The lower and the upper approximations of subsets of U are deﬁned as follows. Deﬁnition 2. For any approximation space AS#,$ = (U, I# , ν$ ) and any subset X ⊆ U , the lower and upper approximations are deﬁned by LOW AS#,$ , X = {x ∈ U : ν$ (I# (x) , X) = 1} , U P P AS#,$ , X = {x ∈ U : ν$ (I# (x) , X) > 0}, respectively. The lower approximation of a set X with respect to the approximation space AS#,$ is the set of all objects, which can be classiﬁed with certainty as objects of X with respect to AS#,$ . The upper approximation of a set X with respect to the approximation space AS#,$ is the set of all objects which can possibly be classiﬁed as objects of X with respect to AS#,$ . Several known approaches to concept approximation can be covered using the approximation spaces discussed here, e.g., the approach given in [8] or tolerance (similarity) rough set approximations (see, e.g., references in [10]). We recall the notions of the positive region and the quality of approximation of classiﬁcation in the case of generalized approximation spaces [13]. Deﬁnition 3. Let AS#,$ = (U, I# , ν$ ) be an approximation space and let r > 1 be a given naturalnumber and let {X1 , . . . , Xr } be a classiﬁcation of objects (i.e. X1 , . . . , Xr ⊆ U , ri=1 Xi = U and Xi ∩ Xj = ∅ for i = j, where i, j = 1, . . . , r). 1. The positive region of the classiﬁcation {X1 , . . . , Xr } with respect to the approximation space AS#,$ is deﬁned by P OS AS#,$ , {X1 , . . . , Xr } = ri=1 LOW AS#,$ , Xi . 2. The quality of approximation of the classiﬁcation {X1 , . . . , Xr } in the approximation space AS#,$ is deﬁned by card(P OS (AS#,$ ,{X1 ,...,Xr })) γ AS#,$ , {X1 , . . . , Xr } = . card(U) The quality of approximation of the classiﬁcation coeﬃcient expresses the ratio of the number of all AS#,$ -correctly classiﬁed objects to the number of all objects in the data table.

3

Relational Data Mining

Knowledge discovery is the process of discovering particular patterns over data. In this context data is typically stored in a database. Approaches using ﬁrst order logic (FOL, for short) languages for the description of such patterns oﬀer data mining the opportunity of discovering more complex regularities which may be out of reach for attribute-value languages.

Approximation Spaces in Multi Relational Knowledge Discovery

3.1

355

Didactic Example

In this section we present an example inspired by [2]. Example 1. There are two information systems: ISCustomer = (UCustomer , ACustomer ) where the set of objects UCustomer = {x1 , . . . , x7 }, and the set of attributes ACustomer = {N ame, Gender, Income, BigSpender} (see Table 1) and ISMarriedT o = (UMarriedT o , AMarriedT o ) where UMarriedT o = {y1 , y2 , y3 }, and AMarriedT o = {Spouse1, Spouse2} (see Table 2). Table 1. An Information System ISCustomer UCustomer x1 x2 x3 x4 x5 x6 x7

Name Mary Eve Kate Meg Jim Tom Henry

Gender Female Female Female Female Male Male Male

Income BigSpender 70000 yes 120000 yes 80000 no 80000 yes 100000 yes 100000 yes 60000 no

Table 2. An Information System ISM arriedT o UM arriedT o Spouse1 Spouse2 y1 M ary Jim y2 M eg T om y3 Kate Henry

Using attribute–value language we obtain for example the following decision rules: if Income ≥ 100000 then BigSpender = yes if Income ≤ 75000 then BigSpender = yes (May be this rule is not intuitive.) if N ame = M eg then BigSpender = yes (This rule is generally not applicable to new objects.) Using ﬁrst order language one can obtain the following two rules: BigSpender(var1 , var2 , var3 ) ← var3 ≥ 100000 BigSpender(var1 , var3 , var3 ) ← M arriedT o(var1 , var1 ) and Customer(var1 , var2 , var3 , var4 ) and var3 ≥ 100000 which involve the predicates Customer and M arriedT o. It predicts a person to be a big spender if the person is married to somebody with high income (compare this to the rule that states a person is a big spender if he/she has high

356

J. Stepaniuk

income, listed above the relational rules). Note that the two persons var1 and var1 are connected through the relation MarriedTo. Relational patterns are typically expressed in subsets of ﬁrst-order logic (also called predicate or relational logic). Essentials of predicate logic include predicates (M arriedT o) and variables (var1 , var1 ), which are not present in propositional logic (attribute–value language). Relational patterns are thus more expressive than the propositional ones. Knowledge discovery based on FOL has other advantages as well. Complex background knowledge provided by experts can be encoded as ﬁrst order formulas and be used in the discovery task. The expressiveness of FOL enables the discovered patterns to be described in a concise way, which in most cases increases readability of the output. Multiple relations can be naturally handled without explicit (and expensive) joins. 3.2

Relational Learning

Before moving on to the algorithm for learning of a set of rules, let us introduce some basic terminology from relational learning. Relational learning algorithms learn classiﬁcation rules for a concept [2] (for relational methods and their applications in computer science see also [5]). The program typically receives a large collection of positive and negative examples from real-world databases as well as background knowledge in the form of relations. Let p be a target predicate of arity m and r1 , . . . , rl be background predicates, where m, l > 0 are given natural numbers. We denote the constants by con1 , . . . , conn , where n > 0. A term is either a variable or a constant. An atomic formula is of the form p (t1 , . . . , tm ) or ri (t1 , . . .) where the t s are terms and i = 1, . . . , l. A literal is an atomic formula or its negation. If a literal contains a negation symbol (¬), we call it a negative literal, otherwise it is a positive literal. A clause is any disjunction of literals, where all variables are assumed to be universally quantiﬁed. The learning task for relational learning systems is as follows: Input + − of positive and a set Xtarget of negative training examples (exa set Xtarget pressed by literals without variables) for the target relation, background knowledge (or BK for short) expressed by literals without variables and not including the target predicate. Output p ) a set of ξ ← λ rules, where ξ is an atomic formula of the form p (var1p , . . . , varm with the target predicate p and λ is a conjunction of literals over background predicates r1 , . . . , rl , such that the set of rules satisﬁes the positive examples relatively to background knowledge. Example 2. Let us consider the data set related to document understanding. The learning task involves identifying the purposes served by components of single-page letters such as that in Figure 1.

Approximation Spaces in Multi Relational Knowledge Discovery

357

Background predicates describe properties of components such as their width and height, and relationships such as horizontal and vertical alignment with other components. Target predicates describe whether a block is one of the ﬁve predetermined types: sender, receiver, logo, reference, and date. For example, for letter presented in Figure 1, we obtain the following predicate data: date (c8 ), logo (c3 ), receiver (c21 ), on top (c8 , c21 ), on top (c21 , c14 ), on top (c5 , c24 ), on top (c3 , c5 ), aligned only lef t col (c1 , c3 ), aligned only right col (c5 , c21 ), . . . We consider generation of rules of the form: sender (var1 ) ← on top (var1 , var2 ) and logo (var2 ). We will adopt the lower and the upper approximations for subsets of the set of target examples. First, we deﬁne the coverage of a rule. Deﬁnition 4. The coverage of Rule, written Coverage(Rule), is the set of examples such that there exists a substitution giving values to all variables appearing in the rule and all literals of the rule are satisﬁed for this substitution. The set of the positive (negative) examples covered by Rule is denoted by Coverage+ (Rule), Coverage− (Rule), respectively. Remark 1. For any literal L, we obtain Coverage(h ← b) = Coverage(h ← b ∧ L) ∪ Coverage(h ← b ∧ ¬L). + − ∪ Xtarget and Rule Set = {Rule1, . . . , Rulen}. Let U = Xtarget

Deﬁnition 5. For the set of rules Rule Set and any example x ∈ U the uncertainty function is deﬁned by IRule

Set (x)

= {x} ∪

n

{Coverage(Rulei ) : x ∈ Coverage(Rulei )} .

i=1

The lower and upper approximations may be deﬁned as earlier but in this case they are equal to the forms presented in Remark 2. Remark 2. For an approximation space ASRule Set = (U, IRule Set , νSRI ) and any subset X ⊆ U the lower and the upper approximations are deﬁned by LOW (ASRule Set , X) = {x ∈ U : IRule U P P (ASRule respectively.

Set , X)

= {x ∈ U : IRule

Set (x)

Set (x)

⊆ X} ,

∩ X = ∅} ,

358

J. Stepaniuk

c1 (sender)

c8 (date)

c21 (receiver)

c3 (logo) c14 (reference)

c5

c7

c24

Fig. 1. Sample Letter Showing Components

4

Translating First–Order Data into Attribute–Value Form

In this section we discuss the approach based on two steps. First, the data is transformed from ﬁrst-order logic into decision table format by the iterative checking whether a new attribute adds any relevant information to the decision table. Next, the reducts and rules from reducts [8,10,13] are computed from the decision table obtained. Data represented as a set of formulas can be transformed into attribute– value form. The idea of translation was inspired by LINUS and DINUS systems

Approximation Spaces in Multi Relational Knowledge Discovery

359

(see, e.g., [2]). We start with a decision table directly derived from the positive and negative examples of the target relation. Assuming that we have mary target predicate, the set U of objects in the decision table is a subset of {con1 , . . . , conn }m . Decision attribute dp : U → {+, −} is deﬁned by the target predicate with possible values ” + ” or ” − ”. All positive and negative examples of the target predicate are now put into the decision table. Each example forms a separate row in the table. Then background knowledge is applied to the decision table. We determine all the possible applications of the background predicates to the arguments of the target relation. Each such application introduces a new Boolean attribute. To analyze the complexity of the obtained data table, let us consider the number of condition attributes. Let Ari be a set of attributes constructed for every predicate symbol ri , where i = 1, . . . , l. The number of condition attributes l in constructed data table is equal to i=1 card (Ari ) resulting from the possible applications of the l background predicates on the variables of the target relation. The cardinality of Ari depends on the number of arguments of target predicate p (denoted by m) and the arity of ri . Namely, card (Ari ) is equal to mar(ri ) , where ar (ri ) is the arity of the predicate ri . The number of condition attributes in obtained data table is polynomial in the arity m of the target predicate p and the number l of background knowledge predicates, but its size is usually so large that its processing is unfeasible. Therefore, one can check interactively if a new attribute is relevant, i.e., if it adds any information to the decision table and, next we add to the decision table only relevant attributes. Two conditions for testing if a new attribute a is relevant are proposed: 1. γ ASB∪{a} , {X+ , X− } > γ (ASB , {X+ , X− }) , where X+ and X− denote the decision classes corresponding to the target concept. An attribute a is added to the decision table if this results in a growth of the positive region with respect to the attributes selected previously. 2. QDIS (a) = νSRI (X+ × X− , {(x, y) ∈ X+ × X− : a (x) = a (y)}) ≥ θ, where θ ∈ [0, 1] is a given real number. An attribute a is added to the decision table if it introduces some discernibility between objects belonging to diﬀerent non-empty classes X+ and X− . Each of these conditions can be applied to a single attribute before it is introduced to the decision table. If this attribute does not meet a condition, it should not be included into the decision table. The received data table is then analyzed by a rough set based systems. First, reducts are computed. Next, decision rules are generated. Example 3. The problem with three binary predicates r1 , r3 , p and one unary predicate r2 can be used to demonstrate the transformation of relational learning problem into attribute–value form. Suppose that there are the following positive and negative examples of a target predicate p : + − = {p(1, 2), p(4, 1), p(4, 2)}, Xtarget = {¬p(6, 2), ¬p(3, 5), ¬p(1, 4)}. Xtarget

360

J. Stepaniuk

Consider the background knowledge about relations, r1 , r2 , and r3 : r1 (5, 1), r1 (1, 2), r1 (1, 4), r1 (4, 1), r1 (3, 1), r1 (2, 6), r1 (3, 5), r1 (4, 2), r2 (1), r2 (2), r2 (3), r2 (4), r2 (6), r3 (2, 1), r3 (1, 4), r3 (2, 4), r3 (2, 5), r3 (3, 2), r3 (3, 5), r3 (5, 1), r3 (5, 3), r3 (2, 6), r3 (4, 2). We then transform the data into attribute–value form (decision table). In Table 3, a quality index QDIS of potential attributes is presented. Table 3. Quality QDIS of Potential Attributes Symbol Attribute QDIS (•) a1 r2 (var1 ) 0 a2 r2 (var2 ) 0.33 a3 r1 (var1 , var1 ) 0 a4 r1 (var1 , var2 ) 0.33 a5 r1 (var2 , var1 ) 0.56 a6 r1 (var2 , var2 ) 0 a7 r3 (var1 , var1 ) 0 a8 r3 (var1 , var2 ) 0.56 a9 r3 (var2 , var1 ) 0.33 a10 r3 (var2 , var2 ) 0

Using conditions introduced in this section some attributes will not be included in the resulting decision table. For example, the second condition with QDIS (•) ≥ θ = 0.3 would permit the following attribute set into the decision table: A0.3 = {a2 , a4 , a5 , a8 , a9 }. Therefore, DT0.3 = (U, A0.3 ∪ {d}) ﬁnally. We obtain two decision classes: X+ = {(1, 2) , (4, 1) , (4, 2)} and X− = {(6, 2) , (3, 5) , (1, 4)} . For the obtained decision table we construct an approximation space ASA0.3 = (U, IA0.3 , νSRI ) such that the uncertainty function and the rough inclusion are deﬁned in Table 4. Then, we can compute reducts and decision rules.

5

The Rough Set Relational Learning Algorithm

In this section we introduce and investigate the RSRL (Rough Set Relational Learning) algorithm. Some preliminary versions of this algorithm were presented in [14,15]. 5.1

RSRL Algorithm

To select the most promising literal from the candidates generated at each step, RSRL considers the performance of the rule over the training data. The evaluation function card(R(L, N ewRule)) used by RSRL to estimate the utility of adding a new literal is based on the numbers of discernible positive and negative examples before and after adding the new literal (see, Figure 2).

Approximation Spaces in Multi Relational Knowledge Discovery

361

Table 4. Resulting Decision Table DT0.3 , Uncertainty Function and Rough Inclusion (var1 , var2 ) (1, 2) (4, 1) (4, 2) (6, 2) (3, 5) (1, 4)

a2 true true true true false true

a4 true true true false true true

a5 false true false true false true

a8 false false true false true true

a9 true true true true true false

dp + + + -

IA0.3 (•) νSRI (•, X+ ) νSRI (•, X− ) {(1, 2)} 1 0 {(4, 1)} 1 0 {(4, 2)} 1 0 {(6, 2)} 0 1 {(3, 5)} 0 1 {(1, 4)} 0 1

Some modiﬁcation of the algorithm RSRL were presented in [15]. The modiﬁed algorithm generates rules as the original RSRL but its complexity is lower because it performs operations on the cardinalities of sets without computing the sets. 5.2

Illustrative Example

Let us illustrate the RSRL algorithm on a simple problem of learning a relation. Example 4. The task is to deﬁne the target relation p(var1 , var2 ) in terms of the background knowledge relations r1 and r3 . Let BK = {r1 (1, 2), r1 (1, 3), r1 (2, 4), r3 (5, 2), r3 (5, 3), r3 (4, 6), r3 (4, 7)}. There are two positive and three negative examples of the target relation: + − = {e1 , e2 } and Xtarget = {e3 , e4 , e5 }, where Xtarget

e1 = p(1, 4), e2 = p(2, 6), e3 = ¬p(5, 4), e4 = ¬p(5, 3) and e5 = ¬p(1, 2). Let us see how the algorithm generates rules for h = p(var1 , var2 ), app = lower. The successive steps of the algorithm: P os = {e1 , e2 }, Learned rules = ∅. P os = ∅. R = {(e1 , e3 ), (e1 , e4 ), (e1 , e5 ), (e2 , e3 ), (e2 , e4 ), (e2 , e5 )}. R = ∅. We obtain the following candidates: ri (var1 , var1 ), ri (var1 , var2 ), ri (var1 , var3 ), ri (var2 , var1 ), ri (var2 , var2 ), ri (var2 , var3 ), ri (var3 , var1 ), ri (var3 , var2 ), where i = 1, 3. For ever1. candidate, we compute R(L, N ewRule) and we obtain the best result for r3 (var1 , var3 ). In the ﬁrst step, every example is covered either by the rule p(var1 , var2 ) ← r3 (var1 , var3 ) or by p(var1 , var2 ) ← ¬r3 (var1 , var3 ). We obtain: e1 , e2 , e5 ∈ Coverage+ (h ← ¬Best literal) ∪ Coverage− (h ← ¬Best literal), e3 , e4 ∈ Coverage+ (h ← Best literal) ∪ Coverage− (h ← Best literal). From the intersection of R and the set (Coverage+ (h ← Best literal) × Coverage− (h ← ¬Best literal))∪

362

J. Stepaniuk

Coverage − (h ← b ∧ L )

Coverage − (h ← b ∧ ¬L )

Coverage + (h ← b ∧ ¬L )

Coverage + (h ← b ∧ L )

Fig. 2. The set R(L, h ← b) is equal to the union of two Cartesian products

(Coverage+ (h ← ¬Best literal) × Coverage− (h ← Best literal)), we obtain R(Best literal, N ewRule) = {(e1 , e3 ), (e1 , e4 ), (e2 , e3 ), (e2 , e4 )} = ∅. Since the value of the coverage of p(var1 , var2 ) ← ¬r3 (var1 , var3 ) is greater than the value of the coverage of p(var1 , var2 ) ← r3 (var1 , var3 ), Best literal = ¬r3 (var1 , var3 ). app = upper. b = ¬r3 (var1 , var3 ). Coverage− (N ewRule) = ∅.

Approximation Spaces in Multi Relational Knowledge Discovery

363

We ﬁnd new R considering a general case: b = b1 = ¬r3 (var1 , var3 ), L1 = ¬r3 (var1 , var3 ). R1 = Coverage+ (h ← b1 ) × Coverage− (h ← b1 ) ∪ S1 . Coverage+ (h ← b1 ) = {e1 , e2 }, Coverage− (h ← b1 ) = {e5 }, Coverage+ (h ← b0 ∧ ¬L1 ) = ∅, Coverage− (h ← b0 ∧ ¬L1 ) = {e3 , e4 }, S1 = Coverage+ (h ← b0 ∧ ¬L1 ) × Coverage− (h ← b0 ∧ ¬L1 ) = ∅. Hence, we obtain R = R1 = {(e1 , e5 ), (e2 , e5 )}. The second step in the second loop: R = ∅. Algorithm 1. RSRL Algorithm + − input : T arget predicate, BK, Xtarget ∪ Xtarget , app //where T arget predicate + is a target predicate with a set Xtarget of positive examples and a set − Xtarget of negative examples, BK is a background knowledge, app is a type of approximation (app ∈ {lower, upper}). output: Learned rules //where Learned rules is a set of rules for ”positive decision class”. + ; P os ←− Xtarget Learned rules ←− ∅; while P os = ∅ do Learn a N ewRule; N ewRule ←− most general rule possible; − ; R ←− P os × Xtarget while R = ∅ do Candidate literals ←− generated candidates; // RSRL generates candidate specializations of N ewRule by considering a new literal L that ﬁts one of the following forms: – r(var1 , . . . , vars ), where at least one of the variable vari in the created literal must already exist in the positive literals of the rule; – the negation of the above form of literal;

Best literal ←− arg max L∈Candidate literals card(R(L, N ewRule)); // (the explanation of R(L, Rule) is in Figure 2 given) if R(Best literal, N ewRule) = ∅ or (app = upper and (N ewRule = most general rule possible) and Coverage+ (N ewRule) = Coverage+ (N ewRule ∧ Best literal)) then exit while; end Add Best literal to N ewRule preconditions; //Add a new literal to specialize N ewRule; if Coverage−(N ewRule) = ∅ then exit while; end R := R \ R(Best literal, N ewRule); end Learned rules ←− Learned rules ∪ {N ewRule} ; P os ←− P os \ Coverage+ (N ewRule); end

364

J. Stepaniuk

We generate new candidates. We obtain the best result for the candidate r1 (var1 , var2 ) thus Best literal = r1 (var1 , var2 ). Now b = b2 = ¬r3 (var1 , var3 ) ∧ r1 (var1 , var2 ), L2 = r1 (var1 , var2 ). We compute the following sets: Coverage+ (h ← b2 ) = ∅, Coverage− (h ← b2 ) = {e5 }, Coverage+ (h ← b1 ∧ ¬L2 ) = {e1 , e2 }, Coverage− (h ← b1 ∧ ¬L2 ) = ∅. Coverage+ (h ← b2 ) × Coverage− (h ← b1 ∧ ¬L2 )∪ Coverage+ (h ← b1 ∧ ¬L2 ) × Coverage− (h ← b2 ) = {(e1 , e5 ), (e2 , e5 )}. We obtain R(Best literal, N ewRule) = {(e1 , e5 ), (e2 , e5 )} = ∅. Since the value of the coverage of p(var1 , var2 ) ← ¬r3 (var1 , var3 ) ∧ ¬r1 (var1 , var2 ) is greater than the value of the coverage of p(var1 , var2 ) ← ¬r3 (var1 , var3 ) ∧ r1 (var1 , var2 ) then Best literal = ¬r1 (var1 , var2 ). app = upper. b = ¬r3 (var1 , var3 ) ∧ ¬r1 (var1 , var2 ). Coverage− (N ewRule) = ∅. The end of the second loop. Learned rules = {p(var1 , var2 ) ← ¬r3 (var1 , var3 ) ∧ ¬r1 (var1 , var2 )}. Coverage+ (p(var1 , var2 ) ← ¬r3 (var1 , var3 ) ∧ ¬r1 (var1 , var2 )) = {e1 , e2 } = P os, hence P os = P os \ Coverage+ (N ewRule) = ∅. The end of the algorithm proceeding. In each step of the algorithm we obtain Coverage+ (N ewRule) = Coverage+ (N ewRule ∧ Best literal). Hence, if app = upper then we obtain the same rules as for app = lower. Hence, the lower and the upper approximations + of Xtarget are equal in our example. Let us compute the above sets to compare them. We have Rule Set = {p(var1 , var2 ) ← ¬r3 (var1 , var3 ) ∧ r1 (var1 , var2 ), + p(var1 , var2 ) ← ¬r3 (var1 , var3 ) ∧ ¬r1 (var1 , var2 )} and Xtarget = {e1 , e2 }. We obtain the uncertainty function IRule Set (e1 ) = IRule Set (e2 ) = {e1 , e2 }, IRule Set (e3 ) = {e3 }, IRule Set (e4 ) = {e4 } and IRule Set(e5 ) = {e5 }. + + Hence, LOW ASRule Set , Xtarget = {e1 , e2 } = U P P ASRule Set , Xtarget .

6

Conclusions

The ﬁrst approach presented in this paper transforms input ﬁrst-order logic formulas into decision table form, then uses reducts to select only meaningful data. The second approach is based on the algorithm RSRL for the ﬁrst order rules generation. We showed that approximation spaces are basic structures for knowledge discovery from multi-relational data. Furthermore, our approach can be treated as a step towards the understanding of rough set methods in the ﬁrst order rules generation.

Acknowledgements The author wishes to thank the anonymous reviewers for their many helpful comments.

Approximation Spaces in Multi Relational Knowledge Discovery

365

References 1. Bonchi, F., Boulicaut, J. F. (Eds.): Knowledge Discovery in Inductive Databases, Lecture Notes in Computer Science 3933, Springer–Verlag, Berlin Heidelberg, 2006. 2. Dzeroski, S., Lavrac, N. (Eds.): Relational Data Mining, Springer-Verlag, Berlin, 2001. 3. L ukasiewicz, J.: Die logischen Grundlagen der Wahrscheinlichkeitsrechnung, Krak´ ow 1913. In: Borkowski, L. (ed.), Jan L ukasiewicz - Selected Works. North Holland, Amstardam, Polish Scientiﬁc Publishers, Warsaw, 1970. 4. Milton, R. S., Maheswari V. U., Siromoney A.: Rough Sets and Relational Learning, Transactions on Rough Sets I, Lecture Notes in Computer Science 3100, Springer, 2004, 321–337. 5. Orlowska, E., Szalas, A. (Eds.): Relational Methods for Computer Science Applications, Physica–Verlag, Heidelberg, 2001. 6. Pal, S.K., Polkowski, L., Skowron, A. (Eds.): Rough-Neural Computing: Techniques for Computing with Words. Springer-Verlag, Berlin, 2004. 7. Pawlak, Z.: Rough sets, International J. Comp. Inform. Science 11, 1982, 341–356. 8. Pawlak, Z.: Rough Sets. Theoretical Aspects of Reasoning about Data, Kluwer Academic Publishers, Dordrecht, 1991. 9. Polkowski, L., Skowron, A. (Eds.): Rough Sets in Knowledge Discovery 1 and 2. Physica-Verlag, Heidelberg, 1998. 10. Skowron, A., Stepaniuk, J.: Tolerance Approximation Spaces, Fundamenta Informaticae, 27, 1996, 245–253. 11. Skowron A., Stepaniuk J., Peters J. F., Swiniarski R.: Calculi of Approximation Spaces, Fundamenta Informaticae vol. 72(1–3), 2006, 363–378. 12. Stepaniuk, J.: Rough relations and logics. In: L. Polkowski, A. Skowron (Eds.), Rough Sets in Knowledge Discovery 1. Methodology and Applications, Physica Verlag, Heidelberg, 1998, 248–260. 13. Stepaniuk, J.: Knowledge Discovery by Application of Rough Set Models, L. Polkowski, S. Tsumoto, T.Y. Lin, (Eds.) Rough Set Methods and Applications. New Developments in Knowledge Discovery in Information Systems, Physica– Verlag, Heidelberg, 2000, 137–233. 14. Stepaniuk, J., G´ oralczuk, L.: An Algorithm Generating First Order Rules Based on Rough Set Methods, (ed.) J. Stepaniuk, Zeszyty Naukowe Politechniki Bialostockiej Informatyka nr 1, 2002, 235–250. [in Polish] 15. Stepaniuk, J., Honko, P.: Learning First–Order Rules: A Rough Set Approach Fundamenta Informaticae, 61(2), 2004, 139–157. 16. Ziarko, W., Variable precision rough set model, Journal of Computer and System Sciences 46, 1993, 39–59.

Finding Relevant Attributes in High Dimensional Data: A Distributed Computing Hybrid Data Mining Strategy Julio J. Vald´es and Alan J. Barton National Research Council Canada, M50, 1200 Montreal Rd., Ottawa, ON K1A 0R6 julio.valdes@nrc-cnrc.gc.ca, alan.barton@nrc-cnrc.gc.ca http://iit-iti.nrc-cnrc.gc.ca

Abstract. In many domains the data objects are described in terms of a large number of features (e.g. microarray experiments, or spectral characterizations of organic and inorganic samples). A pipelined approach using two clustering algorithms in combination with Rough Sets is investigated for the purpose of discovering important combinations of attributes in high dimensional data. The Leader and several k-means algorithms are used as fast procedures for attribute set simplification of the information systems presented to the rough sets algorithms. The data described in terms of these fewer features are then discretized with respect to the decision attribute according to different rough set based schemes. From them, the reducts and their derived rules are extracted, which are applied to test data in order to evaluate the resulting classification accuracy in crossvalidation experiments. The data mining process is implemented within a high throughput distributed computing environment. Nonlinear transformation of attribute subsets preserving the similarity structure of the data were also investigated. Their classification ability, and that of subsets of attributes obtained after the mining process were described in terms of analytic functions obtained by genetic programming (gene expression programming), and simplified using computer algebra systems. Visual data mining techniques using virtual reality were used for inspecting results. An exploration of this approach (using Leukemia, Colon cancer and Breast cancer gene expression data) was conducted in a series of experiments. They led to small subsets of genes with high discrimination power.

1 Introduction As a consequence of the information explosion and the development of sensor and observation technologies, it is now common in many domains to have data objects characterized by an increasingly larger number of attributes, leading to high dimensional databases in terms of the set of fields. A typical example is a gene expression experiment, where the genetic content of samples of tissues are obtained with high throughput technologies (microchips). Usually, thousands of genes are investigated in such experiments. In other bio-medical research contexts, the samples are characterized by infrared, ultraviolet, and other kinds of spectra, where the absorption properties, with respect to a large number of wavelengths, are investigated. The same situation occurs in other domains, and the common denominator is to have a set of data objects of a very high dimensional nature. J.F. Peters et al. (Eds.): Transactions on Rough Sets VI, LNCS 4374, pp. 366–396, 2007. c Springer-Verlag Berlin Heidelberg 2007

Finding Relevant Attributes in High Dimensional Data

367

This paper investigates one, of the possibly many approaches to the problem of finding relevant attributes in high dimensional datasets. The approach is based on a combination of clustering and rough sets techniques in a high throughput distributed computing environment, with low dimensional virtual reality data representations aiding data analysis understanding. The goals are: i) to investigate the behavior of the combination of these techniques in a knowledge discovery process ii) to perform preliminary comparisons of the experimental results from the point of view of the discovered relevant attributes, applied to the example problem of finding relevant genes

2 Datasets In this study publicly available datasets were considered. They result from gene expression experiments in genomics, and appear in numerous studies about data mining and machine learning in bioinformatics. All of them share a feature typical of that kind of information: the data consist of a relatively small number of samples, described in terms of a large collection of attributes. Besides genomics, this situation is found in other fields as well, like experimental physics and astronomy. When infrared, ultraviolet or other spectral properties are used to describe the sampled objects, hundreds or thousands of energy intensity values for radiation emission or absorption at different wavelengths are used as sample attributes. The techniques investigated here are of a general nature, that is, not specific or tailored to any particular domain. The datasets considered for this study were: – Leukemia ALL/AML dataset: (72 samples described in terms of 7129 genes [15]). – Breast Cancer (24 samples described in terms of 12, 625 genes [7]). – Colon Cancer: (62 samples described in terms of 2000 genes [1]). The Leukemia dataset is that of [15], and consists of 7129 genes, where patients are separated into i) a training set containing 38 bone marrow samples: 27 acute lymphoblastic leukemia (ALL) and 11 acute myeloid leukemia (AML), obtained from patients at the time of diagnosis, and ii) a testing set containing 34 samples (24 bone marrow and 10 peripheral blood samples), where 20 are ALL and 14 AML. The test set contains a much broader range of biological samples, including those from peripheral blood rather than bone marrow, from childhood AML patients, and from different reference laboratories that used different sample preparation protocols. In the present study, however, the dataset will not be divided into training and test samples, because crossvalidation is used, as explained below. The breast cancer data selected [7] was that provided by the Gene Expression Omnibus (GEO) (See www.ncbi.nlm.nih.gov/projects/geo/gds/ gds browse.cgi?gds=360). It consists of 24 core biopsies taken from patients found to be resistant (greater than 25% residual tumor volume, of which there are 14) or sensitive (less than 25% residual tumor volume, of which there are 10) to

368

J.J. Vald´es and A.J. Barton

docetaxel treatment. The number of genes (probes) placed onto (and measured from) the microarray is 12, 625, and two classes are recognized: resistant and sensitive. The Colon cancer data correspond to tumor and normal colon tissues probed by oligonucleotide arrays [1].

3 Foundational Concepts 3.1 Clustering Methods Clustering with classical partition methods constructs crisp (non overlapping) subpopulations of objects or attributes. Two such classical algorithms were used in this study: the Leader algorithm [17], and several variants of k-means [2]. Leader Algorithm. The leader algorithm operates with a dissimilarity or similarity measure and a preset threshold. A single pass is made through the data objects, assigning each object to the first cluster whose leader (i.e. representative) is close enough (or similar enough) to the current object w.r.t. the specified measure and threshold. If no such matching leader is found, then the algorithm will set the current object to be a new leader; forming a new cluster. This technique is very fast; however, it has several negative properties: i) the first data object always defines a cluster and therefore, appears as a leader. ii) the partition formed is not invariant under a permutation of the data objects. iii) the algorithm is biased, as the first clusters tend to be larger than the later ones since they get first chance at “absorbing” each object as it is allocated. Variants of this algorithm with the purpose of reducing bias include: a) reversing the order of presentation of a data object to the list of currently formed leaders. b) selecting the absolute best leader found (thus making the object presentation order irrelevant). The highest quality is obtained using b), but at a higher computational cost because the set of leaders (whose cardinality increases as the process progresses), has to be completely explored for every data object. Nevertheless, even with this extra computational overhead, the technique is still very fast, and large datasets can be clustered very quickly. Usually the partitions generated by this method are used as initial approximations to more elaborated methods. K-Means. The k-means algorithm is actually a family of techniques based on the concept of data reallocation. A dissimilarity or similarity measure is supplied, together with an initial partition of the data, and the goal is to alter cluster membership so as to obtain a better partition w.r.t. the chosen measure. The modification of membership is performed by reallocating the data objects to a different group w.r.t. the one in which it was a member. Different variants very often give different partition results. However,

Finding Relevant Attributes in High Dimensional Data

369

in papers dealing with gene expression analysis, very seldom are the specificities of the k-means algorithm described. For the purposes of this study, the following k-means variants were used: Forgy’s, Jancey’s, convergent, and MacQueen’s [13], [20], [24], [2]. Let nc be the number of clusters desired. The definition of an initial partition follows basically two schemes: i) direct specification of a set of nc initial centroids (seed points), or ii) specification of nc initial disjoint groups such that they cover the entire dataset, and compute from them initial centroids to start the process. There are many variations of these two schemes, and the following variants for defining an initial partition were considered in this paper: 1. Select nc data objects and use them as initial centroids. 2. Divide the total number of objects into nc consecutive clusters, compute the centroid of each, and use them as initial centroids. 3. Arbitrary nc centroids are given externally. 4. Take the first nc data objects and use them as initial centroids. The classical k-means clustering is a simple algorithm with the following sequence of steps. 1. Allocate each data unit to the cluster with the nearest seed point (if a dissimilarity measure is used), or to the cluster with the most similar seed point (if a similarity measure is used). 2. Compute new seed points as the centroids of the newly formed clusters 3. If (termination criteria = true) then stop else goto 2 Several termination criteria (or a combination of them) can be established, which provide better control on the conditions under which a k-means process concludes. Among them are the folowing: 1. 2. 3. 4.

A preset number of object reallocations have been performed. A preset number of iterations has been reached. A partition quality measure has been reached. The partition quality measure does not change in subsequent steps.

There are several variants of the general k-means scheme. That is why it is necessary to specify explicitly the specific variant applied. In this paper, several of them were used. K-Means: Forgy’s Variant. The classical Forgy’s k-means algorithm [13] consists of the following steps: i) Begin with any desired initial configuration. Go to (ii) if beginning with a set of seed objects, or go to (iii) if beginning with a partition of the dataset. ii) Allocate each object to the cluster with the nearest (most similar) seed object (centroid). The seed objects remain fixed for a full cycle through the entire dataset. iii) Compute new centroids of the clusters. iv) Alternate (ii) and (iii) until the process converges (that is, until no objects change their cluster membership).

370

J.J. Vald´es and A.J. Barton

K-Means: Jancey’s Variant. In Jancey’s variant [20], the process is similar to Forgy’s, but the first set of cluster seed objects is either given, or computed as the centroids of clusters in the initial partition. Then, at all succeeding stages, each new seed point is found by reflecting the old one through the new centroid for the cluster (a heuristic which tries to approximate the direction of the gradient of the error function). K-Means: MacQueen’s Variant. MacQueen’s method [24] is another popular member of the k-means family, and is composed of the following steps: i) Take the first k data units as clusters of one member each. ii) Assign each of the remaining objects to the cluster with the nearest (most similar) centroid. After each assignment, recompute the centroid of the gaining cluster. iii) After all objects have been assigned in step ii), take the existing cluster centroids as fixed points and make one more pass through the dataset assigning each object to the nearest (most similar) seed object. K-Means: Convergent Variant. The so called convergent k-means [2] is a variant defined by the following steps: i)

Begin with an initial partition like in Forgy’s and Jancey’s methods (or the output of MacQueen’s method). ii) Take each object in sequence and compute the distances (similarities) to all cluster centroids; if the nearest (most similar) is not that of the object’s parent cluster, reassign the object and update the centroids of the losing and gaining clusters. iii) Repeat steps ii) and iii) until convergence is achieved (that is, until there is no change in cluster membership). Similarity Measure. The Leader and the k-means algorithms were used with a similarity measure rather than with a distance. In particular Gower’s general coefficient was used [16], where the similarity between objects i and j is given by Eq-1: Sij =

p k=1

sijk /

p

wijk ,

(1)

k=1

where the weight of the attribute (wijk ) is set equal to 0 or 1 depending on whether the comparison is considered valid for attribute k. If vk (i), vk (j) are the values of attribute k for objects i and j respectively, an invalid comparison occurs when at least one them is missing. In this situation wijk is set to 0. For quantitative attributes (like the ones in the datasets used in this paper), the scores sijk are assigned as in Eq-2: sijk = 1 − |Xik − Xjk |/Rk ,

(2)

where Xik is the value of attribute k for object i (similarly for object j), and Rk is the range of attribute k. For symbolic attributes (nominal), the scores sijk are assigned as in Eq-3 1 if Xik = Xjk sijk = (3) 0 otherwise .

Finding Relevant Attributes in High Dimensional Data

371

3.2 Rough Sets The Rough Set Theory [31] bears on the assumption that in order to define a set, some knowledge about the elements of the dataset is needed. This is in contrast to the classical approach where a set is uniquely defined by its elements. In the Rough Set Theory, some elements may be indiscernible from the point of view of the available information and it turns out that vagueness and uncertainty are strongly related to indiscernibility. Within this theory, knowledge is understood to be the ability of characterizing all classes of the classification. More specifically, an information system is a pair A = (U, A) where U is a non-empty finite set called the universe and A is a non-empty finite set of attributes such that a : U → Va for every a ∈ A . The set Va is called the value set of a. For example, a decision table is any information system of the form A = (U, A ∪ {d}), where d ∈ A is the decision attribute and the elements of A are the condition attributes. Implicants. It has been described [28] that an m-variable function f : Bm → B is called a Boolean function if and only if it can be expressed by a Boolean formula. An implicant of a Boolean function f is a term p such that p f , where is a partial order called the inclusion relation. A prime implicant is an implicant of f that ceases to be so if any of its literals are removed. An implicant p of f is a prime implicant of f in case, for any term q, the implication of Eq-4 holds. pqf ⇒p=q .

(4)

General Boolean Reasoning Solution Scheme. It has been described [28] that following the presentation of earlier work, the general scheme of applying Boolean reasoning to solve a problem P can be formulated as follows: 1. Encode problem P as a system of simultaneously-asserted Boolean equations as in Eq-5, where the gi and hi are Boolean functions on B. ⎧ ⎪ ⎨ g 1 = h1 .. P = . (5) . ⎪ ⎩ g k = hk 2. Reduce the system to a single Boolean equation (e.g. fp = 0) as in Eq-6. fp =

k g i · hi + g i · hi .

(6)

i=1

3. Compute Blake’s Canonical Form (BCF (fp )), the prime implicants of fp . 4. Solutions to P are then obtained by interpreting the prime implicants of fp . Discernibility Matrices. An information system A defines a matrix MA called a discernibility matrix. Each entry MA (x, y) ⊆ A consists of the set of attributes that can be used to discern between objects x, y ∈ U according to Eq-7. MA (x, y) = {a ∈ A : discerns (a, x, y)} . Where, discerns (a, x, y) may be tailored to the application at hand.

(7)

372

J.J. Vald´es and A.J. Barton

Indiscernibility Relations and Graphs. A discernibility matrix MA defines a binary relation RA ⊆ U 2 . The relation RA is called an indiscernibility relation [28] (See Eq-8) with respect to A, and expresses which pairs of objects that we cannot discern between. xRA y ⇔ MA (x, y) = ∅ .

(8)

An alternative way to represent RA is via an indiscernibility graph (IDG), which is a graph GA = (U, RA ) with vertex set U and edge set RA . It has been stated [28] that GA is normally only interesting to consider when RA is a tolerance relation, in which case GA may be used for the purpose of clustering or unsupervised learning. Discernibility Functions. A discernibility function [28] is a function that expresses how an object or a set of objects can be discerned from a certain subset of the full universe of objects. It can be constructed relative to an object x ∈ U from a discernibility matrix MA according to Eq-9.

(9) fA (x) = a∗ : a ∈MA (x, y) and MA (x, y) = ∅ . y∈U

The function fA (x) contains |A| Boolean variables, where variable a∗ corresponds to attribute a. Each conjunction of fA (x) stems from an object y ∈ U from which x can be discerned and each term within that conjunction represents an attribute that discerns between those objects. The prime implicants of fA (x) reveal the minimal subsets of A that are needed to discern object x from the objects in U that are not members of RA (x). In addition to defining discernibility relative to a particular object, discernibility can also be defined for the information system A as a whole. The full discernibility function gA (U ) (See Eq-10) expresses how all objects in U can be discerned from each other. The prime implicants of gA (U ) reveal the minimal subsets of A we need to discern all distinct objects in U from each other.

gA (U ) = fA (x) . (10) x∈U

Reducts. If an attribute subset B ⊆ A preserves the indiscernibility relation RA then the attributes A\B are said to be dispensable. An information system may have many such attribute subsets B. All such subsets that are minimal (i.e. that do not contain any dispensable attributes) are called reducts. The set of all reducts of an information system A is denoted RED(A). In particular, minimum reducts (those with a small number of attributes), are extremely important, as decision rules can be constructed from them [4]. However, the problem of reduct computation is NP-hard, and several heuristics have been proposed [43]. Rough Clustering. Based on the concept of a rough set, modifications to the classical family of k-means algorithms have been introduced in [22] and [23] observing that

Finding Relevant Attributes in High Dimensional Data

373

in data mining it is not possible to provide an exact representation of each class in the partition. For example, an approximation image classification method has been reported in [32]. Rough sets enable such representation using upper and lower bounds. In the case of rough k-means clustering, the centroids of the clusters have to be modified to include the effects of lower and upper bounds. The modified centroid calculations for a distance-based clustering would be as shown in Eq-11 [23]: ⎧ ⎨ wlower × ν∈A(x) vj + wupper × ν∈(A(x)−A(x)) vj if A(x) − A(x) = φ |A(x)| |A(x)−A(x)| x= , ⎩ wlower × ν∈A(x) vj otherwise

|A(x)|

(11) where 1 ≤ j ≤ m (the number of clusters). The parameters wlower and wupper control the importance of the lower and upper bonds. Equation 11 generalizes the corresponding k-means centroids update. If the lower and upper bounds are equal, conventional crisp clusters would be obtained (the boundary region A(x) − A(x) is empty). The object membership w.r.t. the lower or upper bound of a cluster is determined in the following way: Let v be an object and xi , xj the centroids of clusters Xi , Xj respectively, where xi is the closest centroid to object v, and xj an arbitrary other centroid. Let d(v, xi ), d(v, xj ) be the distances from object v to the corresponding centroids, and let T be a threshold value. If d(v, xi ) − d(v, xj ) ≤ T , then v ∈ A(xi ), and v ∈ A(xj ) (i.e. v is not part of any lower bound). Otherwise, v ∈ A(xi ) and clearly v ∈ A(xi ). This algorithm depends on three parameters wlower , wupper , and T . 3.3 Virtual Reality Representation of Relational Structures The role of visualization techniques in the knowledge discovery process is well known. Several reasons make Virtual Reality (VR) a suitable paradigm: Virtual Reality is flexible, in the sense that it allows the choice of different representation models to better accommodate different human perception preferences. In other words, allows the construction of different virtual worlds representing the same underlying information, but with a different look and feel. Thus, the user can choose the particular representation that is most appealing. VR allows immersion. VR creates a living experience. The user is not merely a passive observer or an outsider, but an actor in the world. VR is broad and deep. The user may see the VR world as a whole, and/or concentrate the focus of attention on specific details of the world. Of no less importance is the fact that in order to interact with a Virtual World, no mathematical knowledge is required, but only minimal computer skills. A virtual reality based visual data mining technique, extending the concept of 3D modeling to relational structures, was introduced [40], [41] (see also http://www.hybridstrategies.com). It is oriented to the understanding of large heterogeneous, incomplete and imprecise data, as well as symbolic knowledge. The notion of data is not restricted to databases, but includes logical relations and other forms of both structured and non-structured knowledge. In this approach, the data objects are considered as tuples from a heterogeneous space [39]. Different

374

J.J. Vald´es and A.J. Barton

information sources are associated with the attributes, relations and functions, and these sources are associated with the nature of what is observed (e.g. point measurements, signals, documents, images, directed graphs, etc). They are described by mathematical sets of the appropriate kind called source sets (Ψi ), constructed according to the nature of the information source to represent. Source sets also account for incomplete information. A heterogeneous domain is a Cartesian product of a collection of source sets: ˆ n = Ψ1 × · · · × Ψn , where n > 0 is the number of information sources. H A virtual reality space is the tuple Υ =< O, G, B, m , go , l, gr , b, r >, where O is a relational structure (O =< O, Γ v > , the O is a finite set of objects, and Γ v is a set of relations), G is a non-empty set of geometries representing the different objects and relations. B is a non-empty set of behaviors of the objects in the virtual world. is the set of real numbers and m ⊂ Rm is a metric space of dimension m (Euclidean or not) which will be the actual virtual reality geometric space. The other elements are mappings: go : O → G, l : O → m , gr : Γ v → G, b : O → B. Of particular importance is the mapping l. If the objects are in a heterogeneous space, ˆ n → m . Several desiderata can be considered for building a VR-space. One may l:H be to preserve one or more properties from the original space as much as possible (for example, the similarity structure of the data [6]). From an unsupervised perspective, the role of l could be to maximize some metric/non-metric structure preservation criteria [5], or to minimize some measure of information loss. From a supervised point of view l could be chosen as to emphasize some measure of class separability over the objects in O [41]. Hybrid requirements are also possible. For example, if δij is a dissimilarity measure between any two i, j ∈ U (i, j ∈ [1, N ], where n is the number of objects), and ζiv j v is another dissimilarity measure defined on objects iv , j v ∈ O from Υ (iv = ξ(i), j v = ξ(j), they are in one-to-one correspondence). An error measure frequently used is shown in Eq-12 [35]:

2 1 i<j (δij − ζij ) . (12) Sammon error = δij i<j δij Typically, classical algorithms have been used for directly optimizing measures of this type, like Steepest descent, Conjugate gradient, Fletcher-Reeves, Powell, LevenbergMarquardt, and others. The l mappings within this paper were obtained using the method of Fletcher-Reeves [33]. The new nonlinear features are a form of dimensionality reduction and new attribute creation. 3.4 Gene Expression Programming Pattern matching and function approximation are very important operations within data mining and data analysis. Typical examples of general function approximators are neural networks and fuzzy systems. While their performance is unquestioned, their interpretation is still awkward, sometimes extremely difficult in human terms. In the case of a neural network, the understanding of its performance is obscured by the intricacies of its architecture and its weights, some times very many. In the case of a fuzzy system,

Finding Relevant Attributes in High Dimensional Data

375

the set of fuzzy rules might be large in number and complexity. Moreover, the number of linguistic variables required and the collection of membership functions, might be large as well. Therefore, either a neural network or a fuzzy model may have an excellent performance, but interpretability issues might make a human user reluctant to use them. Analytic functions, have a relation with physical systems in general, which has a long history in science. They are easier to understand by humans, the preferred building blocks of modeling, and a highly condensed form of knowledge. Regression is an example where the family of functions is restricted to a few (typically just one), and the problem reduces to finding a set of parameters or coefficients which makes the function fulfill some desirable approximation property (for example, minimizing a least square error or other model quality measure). However, direct discovery of general analytic functions poses enormous challenges because of the (in principle) infinite size of the search space. This important knowledge discovery problem can be approached from a computational intelligence perspective via evolutionary computation, and the solutions obtained are relevant to a large number of disciplines and domains. In particular genetic programming techniques aim at evolving computer programs, which ultimately are functions. Among this subfield of evolutionary computation, gene expression programming (GEP) is appealing [12]. Gene expression programming (GEP), like genetic algorithms (GAs), evolution strategies (ES) and genetic programming (GP), is an evolutionary algorithm as it uses populations of individuals, selects them according to fitness, and introduces genetic variation using one or more genetic operators. The fundamental difference between these techniques resides in the nature of the individuals. Different from GA, ES and GP, GEP individuals are nonlinear entities of different sizes and shapes (expression trees) encoded as strings of fixed length. For the interplay of the GEP chromosomes and the expression trees (ET), GEP uses a translation system to transfer the chromosomes into expression trees and vice versa [12]. The set of operators applied to GEP chromosomes always produces valid ETs. The chromosomes in GEP itself are composed of genes structurally organized into a head and a tail [11]. The head contains symbols that represent both functions (from a function set F) and terminals (from a terminal set T), whereas the tail contains only terminals. Two different alphabets occur at different regions within a gene. For each problem, the length of the head h is chosen, whereas the length of the tail t is a function of h and the number of arguments of the function with the largest arity. As an example, consider a gene composed of the function set F={Q, +, −, ∗, /}, where Q represents the square root function, and the terminal set T={a, b}. Such a gene (the tail is shown in bold) is: *Q-b++a/-bbaabaaabaab, and encodes the ET which corresponds to the mathematical √equation √ f (a, b) = b · a + ab − ((a − b) + b) , which simplifies to f (a, b) = b·a b . GEP chromosomes are usually composed of more than one gene of equal length. For each problem the number of genes as well as the length of the head has to be chosen. Each gene encodes a sub-ET and the sub-ETs interact with one another forming more complex multi-subunit ETs through a connection function. To evaluate GEP chromosomes, different fitness functions can be used.

376

J.J. Vald´es and A.J. Barton

3.5 Distributed Computing and the Grid Distributed computing can be defined in different ways, and there is no universally accepted formulation of the concept. It can be understood as an environment where idle CPU cycles and storage space of tens, hundreds, or thousands of networked systems can be harnessed to work together on a particular processing-intensive problem. The growth of such processing models has been limited, however, due to a lack of compelling applications and by bandwidth bottlenecks, combined with significant security, management, and standardization challenges. However, in the last years the interest has grown to the extent of making the technology an emergent fact. Increasing desktop CPU power and communications bandwidth have also helped to make distributed computing a more practical approach. The numbers of real applications are still somewhat limited, and the challenges (particularly standardization) are significant. Grid computing is a form of distributed computing that involves coordinating and sharing computing, application, data, storage, or network resources across dynamic and geographically dispersed organizations. As previously stated, there is no universally accepted definition, but a consensus exists in that a Grid is a type of parallel and distributed system that enables the sharing, selection, and aggregation of geographically distributed “autonomous” resources dynamically at runtime depending on their availability, capability, performance, cost, and users’ quality-of-service requirements. Grid technologies promise to change the way complex computational problems are approached and solved. However, the vision of large scale resource sharing is not yet a reality in many areas. Grid computing is an evolving area of computing, where standards and technology are still being developed to enable this new paradigm. The grid computing concept aims to promote the development and advancement of technologies that provide seamless and scalable access to wide-area distributed resources. Computational Grids enable the sharing, selection, and aggregation of a wide variety of geographically distributed computational resources (such as supercomputers, compute clusters, storage systems, data sources, instruments, people) and presents them as a single, unified resource for solving large-scale compute and data intensive computing applications (e.g, molecular modelling for drug design, brain activity analysis, and high energy physics). The idea is analogous to electric power networks (grids) where power generators are distributed, but the users are able to access electric power without bothering about the source of energy and its location. Grids aim at exploiting synergies that result from cooperation–ablity to share and agreegrate distributed computational capabilities and deliver them as service. The use of grid technologies for data mining is an obvious choice for many exploratory data analysis tasks within the knowledge discovery process. The identification of the research issues and their potential priorities for the years 2003-2010, as well as the formulation of proposal of suitable means for implementation, has been addressed by several groups of experts [14], [3]. Among distributed computing systems for delivering high throughput computing, the Condor system stands out [9], [36], [37], [38],(http://www.cs.wisc.edu/ condor/). Condor is a specialized workload management system for computeintensive jobs in a distributed computing environment, developed by the Condor

Finding Relevant Attributes in High Dimensional Data

377

Research Project at the University of Wisconsin-Madison (UW-Madison). Like other full-featured batch systems, Condor provides a job queueing mechanism, scheduling policy, priority scheme, resource monitoring, and resource management. Users submit their serial or parallel jobs to Condor, Condor places them into a queue, chooses when and where to run the jobs based upon a policy, carefully monitors their progress, and ultimately informs the user upon completion. While providing functionality similar to that of a more traditional batch queueing system, Condor’s novel architecture allows it to succeed in areas where traditional scheduling systems fail. Condor can be used to manage a cluster of dedicated compute nodes (such as a “Beowulf” cluster), possibly mixed with individual nodes. In addition, unique mechanisms enable Condor to effectively harness wasted CPU power from otherwise idle desktop workstations. For instance, Condor can be configured to only use desktop machines where the keyboard and mouse are idle. Should Condor detect that a machine is no longer available (such as a key press detected), in many circumstances Condor is able to transparently produce a checkpoint and migrate a job to a different machine which would otherwise be idle. Condor does not require a shared file system across machines - if no shared file system is available, Condor can transfer the job’s data files on behalf of the user, or Condor may be able to transparently redirect all of the job’s I/O requests back to the submit machine. As a result, Condor can be used to seamlessly combine all of an organization’s computational power into one resource. 3.6 Implementation A detailed perspective of data mining procedures provides insight into additional important issues to consider (e.g. storage/memory/communication/management/time/etc) when evaluating a computational methodology consisting of combined techniques. This study presents one possible implementation, from which more software development may occur in order to integrate better and/or different tools. In addition, all of these issues become even more pronounced when, as in this study, a complex problem is investigated. The implementation of the distributed pipeline is shown in Alg.1. It consists of two pieces; a sequential portion, and a distributed portion. For the sequential portion, a specific machine (usually the local host) is used to perform some preliminary processing on the data (as it only needs to be performed once) and then distributes the data via a specific distribution mechanism to a set of waiting computing nodes, which may include the distributing machine. Once all of the computations have completed, the sequential portion of the pipeline may then proceed to collect the results from all of the files that have been placed onto the distributing machine (again via the distribution mechanism, but this time from compute node (e.g. remote host) to distributing host). The resultant databases may then be queried for the purpose of analysis. The specific distribution mechanism used, is a high throughput pipeline (Fig. 2) consisting of many co-operating programs. Such a pipeline structure is generated automatically in order to ease the proper configuration of each participating program within the

378

J.J. Vald´es and A.J. Barton

Algorithm 1. Abstract Conceptualization of the Distributed Pipeline Input : A Data Matrix, DInput Output: A Set of Relevant Attributes. From a Specific Host, Sequentially do GenerateAndConfigurePipeline() ; DRandom ←− ShuffleObjects(i)(Opt(i) , DInput ) ; DistributeToComputeNodes(j)(Opt(j) , DRandom ) ; StartPipelineExecution() ; BlockedWaitForResults() ; // Monitor Each Job’s Progress // Store All Completed Job Results Locally ResultsRules ←− ReformatAndCollectRules(k)() ; ResultsStatistics ←− ReformatAndCollectStats(l)() ; AnalyzeResults(DRandom, ResultsRules , ResultsStatistics) ; end On Each Compute Node Run A Job And do DLeaders ←− ConstructLeaders(m)(Opt(m) , DRandom ) ; DSubsets ←− SubsetSelection(n)(Opt(n) , DLeaders ) ; // e.g. create DSubsets by 10-fold cross-validation forall (DTi r , DTi e ) ∈ DSubsets do r (DTi,discr , CutsTi r ) ←− Discretize(o)(OptT(o)r , DTi r ) ; r ); ReductsTi r ←− FormReducts(p) (OptT(p)r , DTi,discr Tr Tr Rulesi ←− GenerateRules(q)(Opt(q) , ReductsTi r ) ; e ←− Discretize(o)(OptT(o)e , DTi e , CutsTi r ) ; DTi,discr e ); RuleSetMeriti ←− Classify(r) (Opt(r) , DTi,discr Record(RuleSetMeriti) end end

pipeline. In this paper, the automatically generated pipeline was facilitated via i) a file generation program (written in Python and running on the local host) and ii) the Condor tool described in section 3.5. The initial preprocessing stage of the pipeline, occurring on the distributing host after generation of files, involves shuffling the input data records as described previously and in Fig.1. The shuffled data is stored on the distributing host’s disk, in order to provide the same randomized data to the next stage of processing, which occurs on the computing hosts (Fig.2). A Condor submission program, which was also automatically generated, is used to specify all of the data and configuration files for the programs that will execute on the remote host. The submission process enables Condor to: i) ii) iii) iv)

schedule jobs for execution check point them (put a job on hold) transfer all data to the remote host transfer all generated data back to the local host (submitting machine)

Finding Relevant Attributes in High Dimensional Data

379

Fig. 1. Data processing strategy combining clustering, Rough Sets analysis and crossvalidation

The final postprocessing stage of the pipeline involves collecting all of the results (parsing the files) and reporting them in a database.

4 Experimental Methodology The datasets consist of information systems with an attribute set composed of ratio and interval variables, and a nominal or ordinal decision attribute. More general information systems have been described in [39]. The general idea is to construct subsets of relatively similar attributes, such that a simplified representation of the data objects is obtained by using the corresponding attribute subset representatives. The attributes of these simplified information systems are explored from the point of view of their reducts. From them, rules are learned and applied systematically to testing data subsets not involved in the learning process (Fig.1). The whole procedure can be seen as a pipeline. In a first step, the objects in the dataset are shuffled using a randomized approach in order to reduce the possible biases introduced within the learning process by data chunks sharing the same decision attribute. Then, the attributes of the shuffled dataset are clustered using the two families of fast clustering algorithms described in previous sections (the leader, and k-means). Each of the formed clusters of attributes is

380

J.J. Vald´es and A.J. Barton

Fig. 2. Automatically generated high throughput pipeline oriented towards the Condor distributed computing environment

represented by exactly one of the original data attributes. By the nature of the leader algorithm, the representative is the leader (called an l-leader), whereas for a k-means algorithm, a cluster is represented by the most similar object w.r.t. the centroid of the corresponding cluster (called a k-leader). This operation can be seen as a filtering of the attribute set of the original information system. As a next step, the filtered information system undergoes a segmentation with the purpose of learning classification rules, and testing their generalization ability in a cross-validation framework. N-folds are used as training sets; where the numeric attributes present are converted into nominal attributes via a discretization process (many possibilities exist), and from them, reducts are constructed. Finally, classification rules are built from the reducts, and applied to a discretized version of the test fold (according to the cuts obtained previously), from which the generalization ability of the generated rules can be evaluated. Cross-validation is used in order to create a statistically meaningful estimate of the classification accuracy of the rules generated from the reducts for a particular experiment. The final database of experimental results was then sorted by minimum mean classification accuracy. The best mean accuracy experiments were then selected in order to extract the attributes from within the computed reducts for further analysis. Besides the numeric descriptors associated with the application of classification rules to data, use of visual data mining techniques, like the virtual reality representation (section 3.3), enables structural understanding of the data described in terms of the selected

Finding Relevant Attributes in High Dimensional Data

381

subset of attributes and/or the rules learned from them. This technique can be applied at a pre- and/or post-processing stage. In this paper, all of the applications were made in the unsupervised mode. That is, the existing class information was not used during the process, and is incorporated within the resulting visualization only to enable thecomparison between the data structure provided by the predictor variables with the class distribution that is known to exist. Each stage of the process feeds its results to the next stage of processing, yielding a pipelined data analysis stream, with partial outputs that can be used for other kinds of analysis. 4.1 ROSETTA The ROSETTA Software [28], [30] was used within this study with the algorithms that are described within the following sections. Discretization: NaiveScaler. The heuristic implemented by Rosetta [29] was used and was described under the assumption that all condition attributes A are numerical. For each condition attribute a, sort its value set Va to obtain the ordering indicated by Eq-13. va1 < · · · vai < · · · < va|Va | . (13) Then let Ca denote the set of all cuts for attribute a generated in a naive fashion according to equations Eq-14, Eq-15, and Eq-16. (14) Xai = x ∈ U : a (x) = vai , Δia = v ∈ Vd : ∃x ∈ Xai such that d (x) = v , and Ca =

vai + vai+1 i > 1 or Δia = Δi+1 : Δa > 1 or Δi+1 a a 2

(15) .

(16)

The set Ca consists of all cuts midway between two observed attribute values, except for the cuts that are clearly not needed due to the fact that the objects have the same decision value. Hence, such a cut would not discern the objects. If no cuts are found for an attribute, NaiveScaler leaves the attribute unprocessed. Missing values are ignored in the search for cuts. In the worst case, each observed value is assigned its own interval. Discretization: SemiNaiveScaler. The discretization algorithm as implemented within Rosetta [29] is similar to the NaiveScaler but has more logic to handle the case where value-neighboring objects belong to different decision classes. This algorithm typically results in fewer cuts than the simpler NaiveScaler, but may still produce more cuts than are desired. In Eq-17, the set Dai collects the dominating decision values for the objects in Xai . If there are no ties, Dai is a singleton. The rationale used within Rosetta for not adding a cut if the sets of dominating decisions define an inclusion is that it is hoped (although it is stated that the implementation does not check)

382

J.J. Vald´es and A.J. Barton

that a cut will be added for another attribute (different from a) such that the objects in Xai and Xai+1 can be discerned. i i . (17) Da = v ∈ Vd : v = argmax x ∈ Xa : d (x) = v Ca =

v

vai + vai+1 : Dai Dai+1 and Dai+1 Dai 2

.

(18)

Discretization: RSESOrthogonalScaler. This algorithm is an efficient implementation [25] of the Boolean reasoning algorithm [27] within the Rough Set Exploration System (RSES) (See http://logic.mimuw.edu.pl/˜rses/). It is mentioned [29] that this algorithm is functionally similar to BROrthogonalScaler but much faster. Approximate solutions are not supported. If a(x) is missing, object x is not excluded from consideration when processing attribute a, but is instead treated as an infinitely large positive value. If no cuts are found for an attribute, all entries for that attribute are set to 0. Discretization: BROrthogonalScaler. The Rosetta implementation [29] of a previously outlined algorithm [27] was used, which is based on the combination of the NaiveScaler algorithm previously presented and a Boolean reasoning procedure for discarding all but a small subset of the generated cuts. Construct set of candidate cuts Ca according to Eq-16. Then construct a boolean function f from the set of candidate cuts according to Eq-19.

c∗ : c ∈Ca and a (x) < c < a (y) and ∂A (x) = ∂A (y) . (19) f= (x,y) a

Then compute the prime implicant of f using a greedy algorithm [21] (see JohnsonReducer). This Boolean reasoning approach to discretization may result in no cuts being deemed necessary (because they do not aid discernibility) for some attributes. The Rosetta implementation does not alter such attributes. Discretization: EntropyScaler. The Rosetta implementation [29] of the algorithm [8] is based on recursively partitioning the value set of each attribute so that a local measure of entropy is optimized. The minimum description length principle defines a stopping criterion for the partitioning process. Rosetta ignores missing values in the search for cuts and Rosetta does not alter attributes for which no cuts were found. Reduct Computation: RSESExhaustiveReducer. The RSES algorithm included within Rosetta [29] computes all reducts by brute force. Computing reducts is NP-hard, so information systems of moderate size are suggested to be used within Rosetta. Reduct Computation: Holte1RReducer. Rosetta’s [29] algorithm creates all singleton attribute sets, which was inspired by a paper in Machine Learning [18]. The set of all 1R rules, (i.e. univariate decision rules) are thus directly constructed from the attribute sets.

Finding Relevant Attributes in High Dimensional Data

383

Reduct Computation: RSESJohnsonReducer. Rosetta [29] invokes the RSES implementation of the greedy algorithm [21] for reduct computation. No support is provided for IDGs, boundary region thinning or approximate solutions. Reduct Computation: JohnsonReducer. Rosetta [29] invokes a variation of a greedy algorithm to compute a single reduct [21]. The algorithm (See Alg.2) has a natural bias towards finding a single prime implicant of minimal length. The reduct R is found by executing the following algorithm, where w(X) denotes a weight for set X ∈ S that is computed from the data. Support for computing approximate solutions is provided by aborting the loop when enough sets have been removed from S, instead of requiring that S has to be fully emptied.

Algorithm 2. Johnson Reducer Input : A Data Matrix, DInput Output: One Reduct R←∅; // Reduct has no attributes within it S ← {S1 , S2 , . . . , Sn } ; repeat // A contains all attributes that maximizes w(X), // where the sum is taken over all sets X ∈ S that // contain a. w(X)}} ; A ← {a : maximal { {X∈S:a∈X}

// The Rosetta implementation resolves ties arbitrarily a ← RandomElementFromSet (A) ; R ← R ∪ {a} ; // Add attribute to growing reduct S ← {X ∈ S : a ∈ / X} ; // Stop considering sets containing a until S = ∅ ; // No more attributes left for consideration return R ;

Rule Generation: RSESRuleGenerator. Rosetta [29] invokes the RSES implementation of an algorithm to generate rules from a set of reducts. Conceptually performed by overlaying each reduct in the reduct set over the reduct set’s parent decision table and reading off the values.

5 Results 5.1 Leukemia Gene Expression Data The example high dimensional dataset selected is that of [15], and consists of 7129 genes where patients are separated into i) a training set containing 38 bone marrow samples: 27 acute lymphoblastic leukemia (ALL) and 11 acute myeloid leukemia (AML),

384

J.J. Vald´es and A.J. Barton

obtained from patients at the time of diagnosis, and ii) a testing set containing 34 samples (24 bone marrow and 10 peripheral blood samples), where 20 are ALL and 14 AML. The test set contains a much broader range of biological samples, including those from peripheral blood rather than bone marrow, from childhood AML patients,and from different reference laboratories that used different sample preparation protocols. Further, the dataset is known to have two types of ALL, namely B-cell and T-cell. For the purposes of investigation, only the AML and ALL distinction was made. The dataset distributed by [15] contains preprocessed intensity values, which were obtained by rescaling such that overall intensities for each chip are equivalent (A linear regression model using all genes was fit to the data). In this paper no explicit preprocessing of the data was performed, in order to not introduce bias and to be able to expose the behavior of the data processing strategy, the methods used, and their robustness. That is, no background subtraction, deletions, filtering, or averaging of samples/genes were applied, as is typically done in gene expression experiments. In a preprocessing stage, a virtual reality representation of the opriginal dataset in a 3-dimensional space as described in section 3.3 was computed. Gower similarity was used for the original space, and normalized Euclidean distance for the target space. Steepest descent was used for optimizing Sammon’s error. The purpose was to appreciate the relationship of the structure of the existing classes and the collection of original attributes. As shown in (Fig.3, the two Leukemia classes appear completely mixed, as approximated with the original set of attributes. Noisy attributes do not allow a resolution of the classes. The pipeline (Fig.1) was investigated through the generation of 480 k-leader and 160 l-leader for a total of 640 experiments (Table-1). The discretization, reduct computation and rule generation algorithms are those included in the Rosetta system [30]. This approach leads to the generation of 74 files per experiment, with 10-fold cross-validation. From the experiments completed so far, one was chosen which illustrates the kind of results obtained with the explored methodology. It corresponds to a leader clustering algorithm with a similarity threshold of 0.99 (leading to 766 l-leader attributes), used as input to the data processing pipeline containing 38 samples. The results of the best 10 fold cross-validated experiment has a mean accuracy of 0.925 and a standard deviation Table 1. The set of parameters and values used in the experiments using the distributed pipeline environment Algorithm/Parameter Leader Leader Similarity Threshold K-Means Cross-validation Discretization Reduct Computation Rule Generation

Values ReverseSearch, ClosestSearch 0.7, 0.8, 0.9, 0.95, 0.99, 0.999, 0.9999, 0.99999 Forgy, Jancey, Convergent, MacQueen 10 folds BROrthogonalScaler, EntropyScaler, NaiveScaler, SemiNaiveScaler JohnsonReducer, Holte1RReducer RSESRuleGenerator

Finding Relevant Attributes in High Dimensional Data

385

Fig. 3. Snapshot of the Virtual Reality representation of the original Leukemia data (training set with 38 samples + test set with 34, both with 7129 genes). Dark objects= ALL class, Light objects=AML class. Spheres = training, Cubes = test. Representation error = 0.143, Sammon error = 3.56e − 6.

of 0.168. This experiment led to 766 reducts (all of them singleton attributes), which was consistent across each of the 10 folds. The obtained classification accuracy represents a slight improvement over those results reported in [42] (0.912). It was conjectured in that study that the introduction of a cross-validated methodology could improve the obtained classification accuracies, which is indeed the case. It is interesting to observe that all of the 7 relevant attributes (genes) reported in [42] are contained (subsumed) within the single experiment mentioned above. Moreover, they were collectively found using both the leader and k-means algorithms, with different dissimilarity thresholds and number of clusters, whereas with the present approach, a single leader clustering input was required to get the better result. Among the relevant attributes (genes) obtained, many coincide with those reported by [15], [10], and [42]. At a post-processing stage, a virtual reality space representation of the above mentioned experiment is shown in Fig.4. Due to the limitations of representing an interactive virtual world on static media, a snapshot from an appropriate perspective is presented. Sammon’s error [35] was used as criteria for computing the virtual reality space, and Gower’s similarity was used for characterizing the data in the space of the 766 selected genes. After 200 iterations a satisfactory error level of 0.0998 was obtained. It is interesting to see that the ALL and AML classes can be clearly differentiated.

386

J.J. Vald´es and A.J. Barton

Fig. 4. Snapshot of the Virtual Reality representation of the union of all of the reducts obtained from 10 fold cross-validation input (38 samples with 766 genes) for the Leukemia data. The leader clustering algorithm was used with a similarity threshold of 0.99. The ALL and the AML classes are perfectly separated. Representation error = 0.0998.

5.2 Breast Cancer Data A visual representation of the original data in terms of the 12, 625 genes obtained using Gower similarity in the original space, Sammon error, and steepest descent optimization [34] is shown in Fig.5. The sensitive and resistant classes are shown for comparison, and semi-transparent convex hulls wrap the objects from the corresponding classes. There is a little overlap between the two sets, indicating the classification potential of the whole set of attributes, but complete class resolution is not obtained with the nonlinear coordinates of the VR space. Rough k-means [23] was used to cluster the 24 samples into 2 groups using the whole set of original attributes (genes) in order to illustrate the difficulties involved in using all of the original attributes. The particular algorithm parameters (wlower = 0.9, wupper = 0.1, distanceT hreshold = 1). The rough k-means result for the 24 samples using the 12, 625 original attributes, and requesting 2 classes is shown in Fig.6. In the VR space, 2 classes are clearly well differentiated, but one of them contains 5 objects and the other contains the rest. Moreover, when the smaller class is investigated, it contains a mixture of samples from the resistant and sensitive class. Therefore, even the more elaborated rough k-means clustering can not resolve the two known classes from the point of view of all of the original attributes used at the same time. It is also interesting, that for this dataset that no boundary cases were obtained with the clustering parameters used.

Finding Relevant Attributes in High Dimensional Data

387

Fig. 5. Visual representation of 24 breast cancer samples with 12, 625 genes. Convex hulls wrap the resistant(size= 14) and sensitive(size= 10) classes. Absolute Error = 7.33 · 10−2 . Relative Mapping Error = 1.22 · 10−4 .

The experimental settings used in the investigation of breast cancer data with the distributed pipeline are reported in Table 2. For each experiment, the discretization, reduct computation and rule generation algorithms are those included in the Rosetta system [30]. The leader algorithm variants were described in section 3.1. This approach leads to the generation of 84 files per experiment, with 10-fold cross-validation. From the series of l-leader Breast Cancer experiments performed, 4 experiments (Exp-81, Exp-82, Exp-145, Exp-146) were found to be equivalent when analyzing the mean (0.73), median (0.67), standard deviation (0.25), minimum (0.5) and maximum (1.0) of the 10-fold cross validated classification accuracy of the produced rules. For the l-leader algorithm a similarity threshold of 0.7 was used by all experiments, with Exp81 and Exp-145 using closest placement criteria and Exp-82 and Exp-146 using reverse search criteria. The discretization algorithm as provided by the Rosetta system was the RSESOrthogonalScaler for Exp-81 and Exp-82 and the BROrthogonalScaler for Exp145 and Exp-146. The reduct algorithm (RSESExhaustiveReducer) was the same for all 4 experiments with full discernibility and all selection. The rule generation algorithm (RSESRuleGenerator), was also the same for all 4 experiments. In a postprocessing stage, the gene expression programming was applied to selected pipeline results. The idea was to try a simple function set (F = {+, −, ∗}) without the use of numeric constants, in order to reduce the complexity of the assembled functions as much as possible. In particular, experiments 81, 82, 145 and 146 all found a subset

388

J.J. Vald´es and A.J. Barton

Fig. 6. Visual representation of 24 breast cancer samples with 12, 625 genes. Convex hulls wrap the RC1(size= 19) and RC2(size= 5) classes built by rough set based k-means. Absolute Error = 7.06 · 10−2 . Relative Mapping Error = 5.21 · 10−5 . RC1, RC2 stand for the two rough clusters obtained.

of only 3 l-leader attributes with a mean crossvalidation error of 0.73. The subset of attributes was {2, 139, 222}, corresponding to genes {31307 at, 31444 s at, 31527 at}. They represent {12359, 27, 239} data objects respectively. Accordingly, the terminal set defined by the attributes found by the pipeline was set to T = {v2 , v139 , v222 }. The resulting class membership function emerging from the GEP process is as shown in Eq-20: f (v2 , v139 , v222 ) = ((v222 + v139 ) + v139 ) +

(20)

(((v222 − v139 ) ∗ (v2 ∗ v2 )) ∗ ((v139 + v139 ) − v139 )) + (v222 − (((v2 − v222 ) ∗ (v139 − v2 )) ∗ ((v222 + v139 ) + v2 ))) . This analytic expression was simplified with the Yacas computer algebra system http://www.xs4all.nl/˜apinkus/yacas.html which resulted in the expression in Eq-21: 2 2 2 f (v2 , v139 , v222 ) = v222 ∗ v139 − v222 ∗ v2 + v222 ∗ v139 +

v222 ∗ v139 ∗ v22 − v222 ∗ v139 ∗ v2 + 2 ∗ v222 + 2 2 ∗ v2 + v139 ∗ v22 ) + v23 . 2 ∗ v139 − (v139

(21)

Finding Relevant Attributes in High Dimensional Data

389

Table 2. The set of parameters and values used in the experiments with the Breast Cancer dataset using the distributed pipeline environment Algorithm/Parameter Leader Leader Similarity Threshold Cross-validation Discretization Reduct Computation Rule Generation

Values ReverseSearch, ClosestSearch 0.7, 0.8, 0.9, 0.95, 0.99, 0.999, 0.9999, 0.99999 10 folds BROrthogonalScaler, EntropyScaler, NaiveScaler, RSESOrthogonalScaler, SemiNaiveScaler RSESExhaustiveReducer, RSESJohnsonReducer RSESRuleGenerator

The classification rule associated with Eq-21 is shown in Eq-22. The classification accuracy on the original dataset was 91.67%, and it should be noted that only 3 genes out of the 12, 625 original ones are used. Moreover, the resulting model is relatively simple. IF f (v2 , v139 , v222 ) ≥ 0.5) −→ class = sensitive

(22)

otherwise −→ class = resistant . 5.3 Colon Cancer Data A virtual reality space representation of the dataset in terms of the original 2000 attributes was computed for an initial assesment of the structure of the data. Sammon error was used as structure measure, with normalized Euclidean distance as dissimilarity, and Powell’s method for error optimization [34]. In 50 iterations an extremely low mapping error obtained (1.067x10−6) is shown in Fig.7.

Fig. 7. Visual representation (3 dimensions) of 62 colon cancer samples with 2000 genes. Darker objects belong to the tumor class, and lighter objects to the normal class. After 50 iterations: Absolute Error = 1.067 · 10−6 . Relative Mapping Error = 0.0488.

390

J.J. Vald´es and A.J. Barton

Table 3. Two Breast Cancer dataset experiments and their associated reducts for each of the 10 cross-validated folds. The GEP encodings are also reported. Experiment

81

82

Fold 1 2 3 4 5 6 7 8 9 10 1 2 3 4 5 6 7 8 9 10

Reducts {31307 at, 31444 s at, 31527 {31307 at, 31527 at} {31307 at, 31527 at} {31307 at, 31444 s at, 31527 {31307 at, 31444 s at, 31527 {31307 at, 31444 s at, 31527 {31307 at, 31444 s at, 31527 {31307 at, 31444 s at, 31527 {31307 at, 31444 s at, 31527 {31307 at, 31527 at} {31307 at, 31444 s at, 31527 {31444 s at, 31527 at} {31307 at, 31527 at} {31307 at, 31444 s at, 31527 {31307 at, 31444 s at, 31527 {31307 at, 31444 s at, 31527 {31307 at, 31527 at} {31307 at, 31444 s at, 31527 {31307 at, 31527 at} {31307 at, 31527 at}

at}

at} at} at} at} at} at} at}

at} at} at} at}

GEP Encoding {v2 , v139 , v222 } {v2 , v222 } {v2 , v222 } {v2 , v139 , v222 } {v2 , v139 , v222 } {v2 , v139 , v222 } {v2 , v139 , v222 } {v2 , v139 , v222 } {v2 , v139 , v222 } {v2 , v222 } {v2 , v139 , v222 } {v139 , v222 } {v2 , v222 } {v2 , v139 , v222 } {v2 , v139 , v222 } {v2 , v139 , v222 } {v2 , v222 } {v2 , v139 , v222 } {v2 , v222 } {v2 , v222 }

Such a small mapping error indicates that the VR space is a very accurate portrait of the 2000 dimensional original space. The most interesting feature in the VR space is the existence of an intrinsic unidimensionality in the data from the point of view of preserving the distance structure. Although the right hand side of the projected data line in Fig.7 predominantly contains objects of the tumor class, and the left half objects of the normal classes, they are mixed, and therefore, the space does not resolve the classes. Nevertheless, this result is an indication about the potential of finding relatively small subsets of attributes with reasonable classification power and about the large redundancy within the original attributes. The experimental settings used in the investigation of colon cancer data with the distributed pipeline are reported in Table 4.For each experiment, the discretization, reduct computation and rule generation algorithms are those included in the Rosetta system [30]. The leader algorithm variants were described in section 3.1. This approach leads to the generation of 84 files per experiment, with 10-fold cross-validation. From the series of 320 l-leader Colon Cancer experiments, 5 are selected for illustration: Exp-113, Exp-304, Exp-195, Exp-180, and Exp-178. They were found to be equivalent when analyzing the mean (0.73), median (0.67), standard deviation (0.25), minimum (0.5), and maximum (1.0) of the 10-fold cross validated classification accuracy.From the series of l-leader experiments, 2 were selected for illustrating the number of created rules per fold. The i) rules and their respective reducts from which they were generated are shown in Table 6 and ii) the original attribute names as well as related genes as found

Finding Relevant Attributes in High Dimensional Data

391

Table 4. The set of parameters and values used in the experiments with the Colon Cancer dataset using the distributed pipeline environment Algorithm/Parameter Leader Leader Similarity Threshold Cross-validation Discretization Reduct Computation Rule Generation

Values ReverseSearch, ClosestSearch 0.7, 0.8, 0.9, 0.95, 0.99, 0.999, 0.9999, 0.99999 10 folds BROrthogonalScaler, EntropyScaler, NaiveScaler, RSESOrthogonalScaler, SemiNaiveScaler JohnsonReducer, Holte1RReducer, RSESExhaustiveReducer, RSESJohnsonReducer RSESRuleGenerator

Table 5. Selected l-leader Colon Cancer experiments sorted by minimum 10-fold cross-validated classification accuracy. The last column shows the resultant number of pipeline selected attributes (from 2, 000) for each experiment. Experiment Mean Median Standard Deviation Min. Max. Sim. No. Attr. 178 0.562 0.500 0.261 0.167 1.0 0.7 3 180 0.629 0.646 0.219 0.333 1.0 0.8 9

within the source reference [1] are shown in Table 7. In these cases, all reducts were composed of singleton attributes found from the original 2000. They are presented due to their cross-validated minimum and maximum accuracies of [0.167-1.0] and [0.3331.0] respectively. Experiment 178 contains the same 3 singleton reducts in each of the 10 folds, from which [59 − 73] rules were obtained. Whereas, Experiment 180 contains 9 singleton reducts in each of the 10 folds, from which [209 − 241] rules were obtained. For the found l-leaders in these experiments, it can be seen that perfect classification has been made for some of the folds which is a sign of interestingness. In a postprocessing stage, gene expression programming was applied to selected pipeline results. The idea was to try a simple function set (F = {+, −, ∗, sin, cos, log}) in order to reduce the complexity of the assembled nonlinear functions as much as possible. In particular, experiment 178 found a subset of only 3 l-leader attributes. The subset of attributes was {1, 2, 12}, corresponding to genes {H55933, R39465, H86060}. They represent {1, 11, 1988} data objects respectively. Accordingly, the terminal set defined by the attributes found by the pipeline was set to T = {v1 , v2 , v12 }. The resulting class membership function emerging from the GEP process is shown in Eq-23: f (v1 , v2 , v12 ) = ((v1 ∗ (v1 ∗ cos((v12 ∗ v12 )))) − v2 ) + (v1 ∗ (cos(log(v1 )) ∗ v2 )) + ((v2 ∗ (cos((v2 + v2 )) ∗ v2 )) + v1 ) .

(23)

392

J.J. Vald´es and A.J. Barton

When Eq-23 is simplified, the resultant equation is that as shown in Eq-24: 2 ) ∗ v12 + v1 ∗ v2 ∗ cos(log(v1 )) + f (v1 , v2 , v12 ) = cos(v12 2 v1 + v2 ∗ cos(2 ∗ v2 ) − v2 .

(24)

The classification rule associated to Eq-24 is shown in Eq-25, which has a classification accuracy on the original dataset of 88.7%, and it should be noted that only 3 genes out of the 2000 original ones were used. IF (f (v1 , v2 , v12 ) ≥ 0.5) −→ class = normal otherwise −→ class = tumor .

(25)

In a second application of the gene expression programming method, the attributes found by experiment 180 {1, 2, 4, 5, 6, 22, 27, 119, 878} were used (again l-leaders). The terminal set was allowed to have numeric constants. The resulting class membership function emerging from the GEP process is as shown in Eq-26: f (v1 , v2 , v4 , v5 , v6 , v22 , v27 , v119 , v878 ) = (((((v878 − v22 ) ∗ v5 ) ∗ (v4 ∗ v878 )) ∗ ((v6 + v5 ) + v5 )) + v1 ) + (((((v119 − v878 ) ∗ v119 ) ∗ v1 ) ∗ ((v27 + v878 ) ∗ (v1 + v22 ))) + v1 ) + (v27 − (v5 ∗ (k1 − (v119 + ((v119 + v1 ) + v5 ))))) + (k2 ∗ (((v1 − (v878 + v5 )) + v4 ) ∗ v6 )) ,

(26)

where k1 = 3.55777, k2 = −7.828919. After simplification, the resulting function is as shown in Eq-27: f (v1 , v2 , v4 , v5 , v6 , v22 , v27 , v119 , v878 ) = 2 2 ∗ v878 ∗ v52 ∗ v4 + 2 2 2 ∗ v5 ∗ v4 ∗ v6 − v878 ∗ v12 ∗ v119 − v878 ∗ v22 ∗ v1 ∗ v119 + v878

(−2) ∗ v878 ∗ v22 ∗ v52 ∗ v4 − v878 ∗ v22 ∗ v5 ∗ v4 ∗ v6 + 2 − v878 ∗ v22 ∗ v1 ∗ v119 ∗ v27 + v878 ∗ v22 ∗ v1 ∗ v119 2 v878 ∗ v12 ∗ v119 − v878 ∗ v12 ∗ v119 ∗ v27 − v878 ∗ v6 ∗ k2 + 2 ∗ v27 + v22 ∗ v1 ∗ v119

v52 + v5 ∗ v1 + 2 ∗ v5 ∗ v119 − v5 ∗ k1 − v5 ∗ v6 ∗ k2 + v4 ∗ v6 ∗ k2 + v6 ∗ v1 ∗ k2 + 2 ∗ v27 + v12 ∗ v119 2 ∗ v1 + v27 .

(27)

Finding Relevant Attributes in High Dimensional Data

393

The classification rule associated to Eq-27 is as shown in Eq-28: IF (f (v1 , v4 , v5 , v6 , v22 , v27 , v119 , v878 ) ≥ 0.5) −→ class = normal otherwise −→ class = tumor ,

(28)

Table 6. Two Colon Cancer dataset experiments. Exp. 178 has 3 reducts that are the same in all 10 folds. Exp. 180 has 9 reducts that are the same in all 10 folds. Cross Validation Fold No. Rules in Exp. 178 No. Rules in Exp. 180

1 2 3 4 5 6 7 8 9 10 67 72 73 72 68 59 66 72 68 68 230 236 233 234 227 209 216 241 228 227

Table 7. Discovered attributes for 2 Colon Cancer dataset experiments. Exp. 180 found one attribute (v6 =R02593) that was also previously reported. Exp. 178

Exp. 180

Encoding: Original: Encoding: Original: Encoding Original: Compared to [1]:

v1 H55933 v1 H55933 v4 R85482 (R85464)

v2 R39465 v2 R39465 v5 U14973 (U14971)

v12 H86060 v22 J02763 v6 R02593 (Same)

v27 v119 H86060 T72175 v878 M87789

and has a classification accuracy on the original dataset of 91.9%. From the point of view of classification accuracy, it is only slightly better than the one obtained with only 3 attributes. On the other hand, despite the fact that most of the individual terms are relatively simple (addition is the root of the expression tree), the expression as a whole is very complex, and certainly much more than the previous model. Likely such an expression is not an arguable replacement for a neural network or a set of fuzzy relations in terms of simplicity or understandability, and moreover, the situation can be even worse if the function set is extended with other nonlinear functions like ex , ln(x), transcendental functions, numeric constants, etc., as is required in complex function approximation tasks. However, despite these difficulties, genetic programming, and particularly GEP allows an explicit assessment of the role of predictor variables. It also provides analytic means to perform sensitivity analysis directly, through the study of the partial derivatives and the multidimensional gradient of the generated functions. The approach is promising, and new developments in genetic programming including meta-function approximation and the incorporation of more intelligent techniques may overcome the above mentioned difficulties. It is interesting to observe that gene 2 is not found in Eq-28, indicating that it was irrelevant (boldfaced in Eq-27). However, despite the increased complexity, this independent technique showed that the set of genes suggested by the data mining pipeline has an important classification power.

394

J.J. Vald´es and A.J. Barton

6 Conclusions Good results were obtained with the proposed high throughput pipeline based on the combination of clustering and rough sets techniques for the discovery of relevant attributes in high dimensional data. The use of several clustering and rough set analysis techniques, and their combination as a virtual data mining machine implemented in a grid and high throughput computing environment, proved to be a promising way to address complex knowledge discovery tasks. In particular, the introduction of a fast attribute reduction procedure aided rough set reduct discovery in terms of computational time, of which the former is further improvable via its amenability for parallel and distributed computing. Cross-validated experiments using three different sets of gene expression data demonstrate the possibilities of the proposed approach. With the hybrid methodology presented, in all cases it was possible to find subsets of attributes of size much smaller than the original set, retaining a high classification accuracy. The pre- and post-processing stages of visual data mining using multidimensional space mappings and genetic programming techniques like gene expression programming (in combination with computer algebra systems), are effective elements within the data processing strategy proposed. The analytic functional models obtained for evaluating class memberships and ultimately for classification via gene expression programming, allowed a better understanding of the role of the different attributes in the classification process, as well as an explicit explanation of their influence. More thorough studies are required to correctly evaluate the impact of the experimental settings on the data mining effectiveness, and further experiments with this approach are necessary.

Acknowledgements This research was conducted within the scope of the BioMine project (National Research Council Canada (NRC), Institute for Information Technology (IIT)). The authors would like to thank Robert Orchard and Marc Leveille from the Integrated Reasoning Group (NRC-IIT), and to Ratilal Haria and Roger Impey from the High Performance Computing Group (NRC-IIT).

References 1. Alon, U., Barkai, N., Notterman, D.A., Gish, K., Ybarra, S., Mack, D., Levine, A.J.: Broad patterns of gene expression revealed by clustering analysis of tumor and normal colon tissues probed by oligonucleotide arrays. In: Proceedings National Academy of Science. USA v96 (1999) 67456750 2. Anderberg, M.: Cluster Analysis for Applications. Academic Press (1973) 3. Bal, H., de Laat, C., Haridi, S., Labarta, J., Laforenza, D., Maccallum, P., Mass, J., Matyska, L., Priol, T., Reinefeld, A., Reuter, A., Riguidel, M., Snelling, D., van Steen, M.: Next Generation Grid(s) European Grid Research 2005 - 2010 Expert Group Report, (2003) 4. Bazan, J.G., Skowron, A., Synak, P.: Dynamic Reducts as a Tool for Extracting Laws from Decision Tables. In: Proceedings of the Symp. on Methodologies for Intelligent Systems. Charlotte, NC, Oct. 16-19 1994. Lecture Notes in Artificial Intelligence 869, Springer-Verlag (1994) 346-355

Finding Relevant Attributes in High Dimensional Data

395

5. Borg, I., Lingoes, J.: Multidimensional similarity structure analysis. Springer-Verlag, New York, NY (1987) 6. Chandon, J.L., Pinson, S.: Analyse typologique. Thorie et applications. Masson, Paris (1981) 7. Chang, J.C. et al.: Gene expression profiling for the prediction of therapeutic response to docetaxel in patients with breast cancer. Mechanisms of Disease. The Lancet, vol 362 (2003) 8. Dougherty, J., Kohavi, R., Sahami,M.: Supervised and unsupervised discretization of continuous features. In A. Prieditis and S. Russell, editors, Proc. Twelfth International Conference on Machine Learning, Morgan Kaufmann (1995) 194-202 9. Epema, D.H.J., Livny, M., van Dantzig, R., Evers, X., Pruyne, J.: A worldwide flock of Condors: Load sharing among workstation clusters. Journal of Future Generation Computer Systems, (1996) 53-65 10. Famili, F., Ouyang, J.: Data mining: understanding data and disease modeling. In: Proceedings of the 21st IASTED International Conference, Applied Informatics, Innsbruck, Austria, Feb. 10-13, (2003) 32-37 11. Ferreira C.: Gene Expression Programming: A New Adaptive Algorithm for Problem Solving. Journal of Complex Systems ˇ13, 2, (2001) 87-129 12. Ferreira C.: Gene Expression Programming: Mathematical Modeling by an Artificial Intelligence. Angra do Heroismo, Portugal (2002) 13. Forgy, E.W.: Cluster analysis of multivariate data: Efficiency versus interpretability of classifications. Biometric Soc. Meetings, Riverside, California (abstract in Biometrics, v.21, no. 3 (1965) 768 14. Foster, I., Kesselman, C., Tuecke, S.: The Anatomy of the Grid: Enabling Scalable Virtual Organizations. International. Journal of Supercomp. App., v.15(3):20 (2001) 222-237 15. Golub, T.R., et al.: Molecular classification of cancer: class discovery and class prediction by gene expression monitoring. Science, vol. 286 (1999) 531-537 16. Gower, J.C.: A general coefficient of similarity and some of its properties. Biometrics, v.1, no. 27, (1973) 857-871 17. Hartigan, J.: Clustering Algorithms. John Wiley & Sons (1975) 18. Holte, R.C.: Very simple classification rules perform well on most commonly used datasets. Machine Learning, 11(1) April (1993) 63-91 19. Jain, A.K., Mao, J.: Artificial Neural Networks for Nonlinear Projection of Multivariate Data. In: Proceedings 1992 IEEE Joint Conf. on Neural Networks (1992) 335-340 20. Jancey, R.C.: Multidimensional group analysis. Australian Journal of Botany, v.14, no. 1 (1966) 127-130 21. Johnson, D.S.: Approximation algorithms for combinatorial problems. Journal of Computer and System Sciences, 9 (1974) 256-278 22. Lingras, P.: Unsupervised Rough Classification using GAs. Journal of Intelligent Information Systems v16, 3 Springer-Verlag (2001) 215-228 23. Lingras, P., Yao, Y.: Time Complexity of Rough Clustering: GAs versus K-Means. Third. Int. Conf. on Rough Sets and Current Trends in Computing RSCTC 2002. Alpigini, Peters, Skowron, Zhong (Eds.) Lecture Notes in Computer Science (Lecture Notes in Artificial Intelligence Series) LNCS 2475. Springer-Verlag (2002) 279-288 24. MacQueen, J.B.: Some methods for classification and analysis of multivariate observations. In: Proceedings of the 5-th Symposium on Math. Statist. and Probability. Berkeley. AD669871 Univ. of California Press, Berkeley. v.1 (1967) 281–297 25. Nguyen, H.S., Nguyen, S.H.: Some efficient algorithms for rough set methods. In: Proceedings Fifth Conference on Information Processing and Management of Uncertainty in Knowledge-Based Systems (IPMU’96), Granada, Spain, July (1996) 1451–1456 26. Nguyen, H.S., Nguyen, S.H.: Discretization Methods in Data Mining. In: L. Polkowski, A. Skowron (eds.): Rough Sets in Knowledge Discovery. Physica-Verlag, Heidelberg (1998) 451-482

396

J.J. Vald´es and A.J. Barton

27. Nguyen, H.S., Skowron, A.: Quantization of real-valued attributes. In: Proceedings Second International Joint Conference on Information Sciences, Wrightsville Beach, NC, September (1995) 34-37 28. Øhrn, A.: Discernibility and Rough Sets in Medicine: Tools and Applications. PhD thesis, Norwegian University of Science and Technology, Department of Computer and Information Science, December NTNU report 1999:133. [http://www.idi.ntnu.no/˜aleks/thesis/] (1999) 29. Øhrn, A.: Rosetta Technical Reference Manual. Department of Computer and Information Science, Norwegian University of Science and Technology, Trondheim, Norway (2001) 30. Øhrn, A., Komorowski, J.: Rosetta- A Rough Set Toolkit for the Analysis of Data. In: Proceedings of Third Int. Join Conf. on Information Sciences (JCIS97), Durham, NC, USA, March 1-5 (1997) 403-407 31. Pawlak, Z.: Rough sets: Theoretical aspects of reasoning about data. Kluwer Academic Publishers, Dordrecht, Netherlands (1991) 32. Peters, J.F., Borkowski, M.: K-means Indiscernibility Relation over Pixels, Fourth. Int. Conf. on Rough Sets and Current Trends in Computing RSCTC 2004. Tsumoto, Slowinski, Komorowki, Grzymala-Busse (Eds.) Lecture Notes in Computer Science (Lecture Notes in Artificial Intelligence Series) LNAI 3066, Springer-Verlag (2004) 580-585 33. Press, W.H., Flannery, B.P., Teukolsky, S.A., Vetterling, W.T.: Numerical Recipes in C, Cambridge University Press, New York (1986) 34. Press, W.H., Teukolsky, S.A., Vetterling, W.T., Flannery, B.P.: Numerical Recipes in C. The Art of Scientific Computing. Cambridge Univ. Press (1992) 35. Sammon, J.W.: A non-linear mapping for data structure analysis. IEEE Trans. on Computers C18 (1969) 401-409 36. Tannenbaum, T., Wright, D., Miller, K., Livny, M.: Condor – A Distributed Job Scheduler. In Thomas Sterling (Ed.) Beowulf Cluster Computing with Linux. MIT Press (2001) 37. Thain, D., Tannenbaum, T., Livny, M.: Condor and the Grid. In Fran Berman and Geoffrey Fox and Tony Hey (Eds.) Grid Computing: Making the Global Infrastructure a Reality. John Wiley & Sons (2002) 38. Thain, D., Tannenbaum, T., Livny, M.: Distributed Computing in Practice: The Condor Experience. Journal of Concurrency and Computation: Practice and Experience (2004) 39. Vald´es, J.J.: Similarity-Based Heterog

Editorial Board David Hutchison Lancaster University, UK Takeo Kanade Carnegie Mellon University, Pittsburgh, PA, USA Josef Kittler University of Surrey, Guildford, UK Jon M. Kleinberg Cornell University, Ithaca, NY, USA Friedemann Mattern ETH Zurich, Switzerland John C. Mitchell Stanford University, CA, USA Moni Naor Weizmann Institute of Science, Rehovot, Israel Oscar Nierstrasz University of Bern, Switzerland C. Pandu Rangan Indian Institute of Technology, Madras, India Bernhard Steffen University of Dortmund, Germany Madhu Sudan Massachusetts Institute of Technology, MA, USA Demetri Terzopoulos University of California, Los Angeles, CA, USA Doug Tygar University of California, Berkeley, CA, USA Moshe Y. Vardi Rice University, Houston, TX, USA Gerhard Weikum Max-Planck Institute of Computer Science, Saarbruecken, Germany

4374

James F. Peters Andrzej Skowron Ivo Düntsch Jerzy Grzymała-Busse Ewa Orłowska Lech Polkowski (Eds.)

Transactions on Rough Sets VI Commemorating the Life and Work of Zdzisław Pawlak, Part I

13

Editors-in-Chief James F. Peters University of Manitoba, Winnipeg, Manitoba R3T 5V6, Canada E-mail: [email protected] Andrzej Skowron Warsaw University, Banacha 2, 02-097 Warsaw, Poland E-mail: [email protected] Volume Editors Ivo Düntsch Brock University, St. Catharines, Ontario L2S 3A1, Canada E-mail: [email protected] Jerzy Grzymała-Busse University of Kansas, Lawrence, KS 66045, USA E-mail: [email protected] Ewa Orłowska National Institute of Telecommunications, ul. Szachowa 1, 04-894 Warsaw, Poland E-mail: [email protected] Lech Polkowski University of Warmia and Mazury and Polish-Japanese Institute of Information Technology Warsaw 10560 Olsztyn, Poland E-mail: [email protected] Library of Congress Control Number: 2007922187 CR Subject Classification (1998): F.4.1, F.1, I.2, H.2.8, I.5.1, I.4 LNCS Sublibrary: SL 1 – Theoretical Computer Science and General Issues ISSN 0302-9743 (Lecture Notes in Computer Science) ISSN 1861-2059 (Transactions on Rough Sets) ISBN-10 3-540-71198-8 Springer Berlin Heidelberg New York ISBN-13 978-3-540-71198-8 Springer Berlin Heidelberg New York This work is subject to copyright. All rights are reserved, whether the whole or part of the material is concerned, specifically the rights of translation, reprinting, re-use of illustrations, recitation, broadcasting, reproduction on microfilms or in any other way, and storage in data banks. Duplication of this publication or parts thereof is permitted only under the provisions of the German Copyright Law of September 9, 1965, in its current version, and permission for use must always be obtained from Springer. Violations are liable to prosecution under the German Copyright Law. Springer is a part of Springer Science+Business Media springer.com © Springer-Verlag Berlin Heidelberg 2007 Printed in Germany Typesetting: Camera-ready by author, data conversion by Scientific Publishing Services, Chennai, India Printed on acid-free paper SPIN: 12028375 06/3142 543210

Preface

Volume VI of the Transactions on Rough Sets (TRS) commemorates the life and work of Zdzislaw Pawlak (1926-2006)1. His legacy is rich and varied. Professor Pawlak’s research contributions have had far-reaching implications inasmuch as his works are fundamental in establishing new perspectives for scientiﬁc research in a wide spectrum of ﬁelds. From a very early age, Zdzislaw Pawlak devoted his life to scientiﬁc research. The pioneering work by Prof. Pawlak included research on the design of computers, information retrieval, modeling conﬂict analysis and negotiation, genetic grammars, and molecular computing. His research led to the introduction of knowledge representation systems during the early 1970s and the discovery of rough sets during the early 1980s. Added to that was Prof. Pawlak’s lifelong interest in painting, photography, and poetry. During his lifetime, he nurtured worldwide interest in approximation, approximate reasoning, and rough set theory and its applications2 . Evidence of the inﬂuence of Prof. Pawlak’s work can be seen in the growth in the rough-set literature that now includes over 4000 publications by more than 1600 authors in the rough set database3 as well as the growth and maturity of the International Rough Set Society4 . Numerous biographies of Zdzislaw Pawlak have been published5 . This volume of the TRS presents papers that reﬂect the profound inﬂuence of a number of research initiatives by Zdzislaw Pawlak. In particular, this volume introduces a number of new advances in the foundations and applications of artiﬁcial intelligence, engineering, logic, mathematics, and science. These advances have signiﬁcant implications in a number of research areas such as the foundations of rough sets, approximate reasoning, bioinformatics, computational intelligence, cognitive science, data mining, information systems, intelligent systems, machine intelligence, and security. In addition, it is evident from the papers included in this volume that rough set theory and its application form a very active research area worldwide. A total of 41 researchers from 8 countries are represented in this volume, namely, Canada, India, France, Norway, Poland, P.R.

1 2

3 4 5

Prof. Pawlak passed away on 7 April 2006. See, e.g., Pawlak, Z., Skowron, A.: Rudiments of rough sets, Information Sciences 177 (2007) 3-27; Pawlak, Z., Skowron, A.: Rough sets: Some extensions, Information Sciences 177 (2007) 28-40; Pawlak, Z., Skowron, A.: Rough sets and Boolean reasoning, Information Sciences 177 (2007) 41-73. http://rsds.wsiz.rzeszow.pl/rsds.php http://roughsets.home.pl/www/ See, e.g., Peters, J.F. and Skowron, A., Zdzislaw Pawlak: Life and Work. Transactions on Rough Sets V, LNCS 4100 (2006) 1-24. See, also, R. Slowi´ nski, Obituary, Prof. Zdzislaw Pawlak (1926-2006), Fuzzy Sets and Systems 157 (2006) 2419-2422.

VI

Preface

China, Sweden, Russia, Thailand, and the USA. Evidence of the vigor, breadth and depth of research in the theory and applications of rough sets can be found in the articles in this volume. Most of the contributions of this commemorative volume of the TRS are on an invitational basis and every paper has been refereed in the usual way. This special issue of the TRS contains 23 papers and extended abstracts that explore a number of research streams that are either directly or indirectly related to research initiatives by Zdzislaw Pawlak. These research streams are represented by papers on propositional logics (Mohua Banerjee and Md. Aquil Khan), intuitionistic rough sets for database applications (Theresa Beaubouef and Fred Petry), missing attribute value problem (Jerzy W. Grzymala-Busse and Witold J. Grzymala-Busse), Zdzislaw Pawlak’s contributions to the study of vagueness (Mihir Chakraborty), data mining (Alicja Wakulicz-Deja and Grzegorz Ilczuk), approximation of concepts (Anna Gomoli´ nska), intelligent systems (Andrzej Jankowski and Andrzej Skowron), acoustics (Bozena Kostek), rule evaluation (Jiye Li, Puntip Pattaraintakorn, and Nick Cercone), rough sets in China (Qing Liu and Hui Sun), four-valued logic (Jan Maluszy´ nski, Andrzej Szalas and Aida Vit´ oria), crisp and fuzzy information systems (Alicja Mieszkowicz-Rolka and Leszek Rolka), artiﬁcial intelligence and rough sets (Tosiharu Munakata), topology and information systems (Piero Pagliani and Mihir K. Chakraborty), conjugate information systems (Maria Semeniuk-Polkowska), incomplete transactional databases (Grzegorz Protaziuk and Henryk Rybinski), classiﬁers, rule induction and rough sets (Jerzy Stefanowski), approximation spaces (Jaroslaw Stepaniuk), relevant attributes in high-dimensional data (Julio J. Vald´es and Alan J. Barton), knowledge discovery in databases (Anita Wasilewska, Ernestina Menasalvas, Christelle Scharﬀ), information quanta and approximation operators (Marcin Wolski), lattice theory for rough sets (Jouni J¨ arvinen). The editors of this volume extend their hearty thanks to reviewers of papers that have been submitted to the TRS during the past 12 months: Manuel OjedaAciego, Mohua Banerjee, Jan Bazan, Mihir Chakraborty, Anna Gomoli´ nska, Etienne Kerre, Pawan Lingras, Victor Marek, Piero Pagliani, Sheela Ramanna, ´ ezak, Jerzy Stefanowski, Jaroslaw Stepaniuk, Piotr Synak, Piotr Dominik Sl¸ Wasilewski and Yiyu Yao. This issue of the TRS has been made possible thanks to the laudable eﬀorts of a great many generous persons and organizations. The editors and authors of this volume also extend an expression of gratitude to Alfred Hofmann, Ursula Barth, Christine G¨ unther and the LNCS staﬀ at Springer for their support in making this volume of the TRS possible. In addition, the editors extend their thanks to Marcin Szczuka for his consummate skill and care in the compilation of this volume. The editors have been supported by the State Committee for Scientiﬁc Research of the Republic of Poland (KBN),

Preface

VII

research grant No. 3T11C00226, and the Natural Sciences and Engineering Research Council of Canada (NSERC) research grant 185986. December 2006

Ivo D¨ untsch Jerzy W. Grzymala-Busse Ewa Orlowska James F. Peters Lech Polkowski Andrzej Skowron

LNCS Transactions on Rough Sets

This journal subline has as its principal aim the fostering of professional exchanges between scientists and practitioners who are interested in the foundations and applications of rough sets. Topics include foundations and applications of rough sets as well as foundations and applications of hybrid methods combining rough sets with other approaches important for the development of intelligent systems. The journal includes high-quality research articles accepted for publication on the basis of thorough peer reviews. Dissertations and monographs up to 250 pages that include new research results can also be considered as regular papers. Extended and revised versions of selected papers from conferences can also be included in regular or special issues of the journal. Honorary Editor: Editors-in-Chief:

Zdzislaw Pawlak – deceased James F. Peters, Andrzej Skowron

Editorial Board M. Beynon G. Cattaneo M.K. Chakraborty A. Czy˙zewski J.S. Deogun D. Dubois I. D¨ untsch S. Greco J.W. Grzymala-Busse M. Inuiguchi J. Jrvinen D. Kim J. Komorowski C.J. Liau T.Y. Lin E. Menasalvas M. Moshkov T. Murai

M. do C. Nicoletti H.S. Nguyen S.K. Pal L. Polkowski H. Prade S. Ramanna R. Slowi´ nski J. Stefanowski J. Stepaniuk Z. Suraj ´ R. Swiniarski M. Szczuka S. Tsumoto G. Wang Y. Yao N. Zhong W. Ziarko

Table of Contents

Contributed Papers Propositional Logics from Rough Set Theory . . . . . . . . . . . . . . . . . . . . . . . . Mohua Banerjee and Md. Aquil Khan

1

Intuitionistic Rough Sets for Database Applications . . . . . . . . . . . . . . . . . . Theresa Beaubouef and Frederick E. Petry

26

An Experimental Comparison of Three Rough Set Approaches to Missing Attribute Values . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Jerzy W. Grzymala-Busse and Witold J. Grzymala-Busse

31

Pawlak’s Landscaping with Rough Sets . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Mihir K. Chakraborty

51

A Comparison of Pawlak’s and Skowron–Stepaniuk’s Approximation of Concepts . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Anna Gomoli´ nska

64

Data Preparation for Data Mining in Medical Data Sets . . . . . . . . . . . . . . Grzegorz Ilczuk and Alicja Wakulicz-Deja

83

A Wistech Paradigm for Intelligent Systems . . . . . . . . . . . . . . . . . . . . . . . . . Andrzej Jankowski and Andrzej Skowron

94

The Domain of Acoustics Seen from the Rough Sets Perspective . . . . . . . Bozena Kostek

133

Rule Evaluations, Attributes, and Rough Sets: Extension and a Case Study . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Jiye Li, Puntip Pattaraintakorn, and Nick Cercone

152

The Impact of Rough Set Research in China: In Commemoration of Professor Zdzislaw Pawlak . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Qing Liu and Hui Sun

172

A Four-Valued Logic for Rough Set-Like Approximate Reasoning . . . . . . Jan Maluszy´ nski, Andrzej Szalas, and Aida Vit´ oria

176

On Representation and Analysis of Crisp and Fuzzy Information Systems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Alicja Mieszkowicz-Rolka and Leszek Rolka

191

On Partial Covers, Reducts and Decision Rules with Weights . . . . . . . . . . Mikhail Ju. Moshkov, Marcin Piliszczuk, and Beata Zielosko

211

XII

Table of Contents

A Personal View on AI, Rough Set Theory and Professor Pawlak . . . . . . Toshinori Munakata

247

Formal Topology and Information Systems . . . . . . . . . . . . . . . . . . . . . . . . . . Piero Pagliani and Mihir K. Chakraborty

253

On Conjugate Information Systems: A Proposition on How to Learn Concepts in Humane Sciences by Means of Rough Set Theory . . . . . . . . . Maria Semeniuk–Polkowska

298

Discovering Association Rules in Incomplete Transactional Databases . . . Grzegorz Protaziuk and Henryk Rybinski

308

On Combined Classiﬁers, Rule Induction and Rough Sets . . . . . . . . . . . . . Jerzy Stefanowski

329

Approximation Spaces in Multi Relational Knowledge Discovery . . . . . . . Jaroslaw Stepaniuk

351

Finding Relevant Attributes in High Dimensional Data: A Distributed Computing Hybrid Data Mining Strategy . . . . . . . . . . . . . . . . . . . . . . . . . . . Julio J. Vald´es and Alan J. Barton

366

A Model PM for Preprocessing and Data Mining Proper Process . . . . . . . Anita Wasilewska, Ernestina Menasalvas, and Christelle Scharﬀ

397

Monographs Lattice Theory for Rough Sets . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Jouni J¨ arvinen

400

Author Index . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .

499

Propositional Logics from Rough Set Theory Mohua Banerjee and Md. Aquil Khan Department of Mathematics and Statistics, Indian Institute of Technology, Kanpur 208 016, India {mohua,mdaquil}@iitk.ac.in

Abstract. The article focusses on propositional logics with semantics based on rough sets. Many approaches to rough sets (including generalizations) have come to the fore since the inception of the theory, and resulted in diﬀerent “rough logics” as well. The essential idea behind these logics is, quite naturally, to interpret well-formed formulae as rough sets in (generalized) approximation spaces. The syntax, in most cases, consists of modal operators along with the standard Boolean connectives, in order to reﬂect the concepts of lower and upper approximations. Non-Boolean operators make appearances in some cases too. Information systems (“complete” and “incomplete”) have always been the “practical” source for approximation spaces. Characterization theorems have established that a rough set semantics based on these “induced” spaces, is no diﬀerent from the one mentioned above. We also outline some other logics related to rough sets, e.g. logics of information systems – which, in particular, feature expressions corresponding to attributes in their language. These systems address various issues, such as the temporal aspect of information, multiagent systems, rough relations. An attempt is made here to place this gamut of work, spread over the last 20 years, in one platform. We present the various relationships that emerge and indicate questions that surface.

1

Introduction

A “logic of rough sets” would, in the natural sense, represent a formal system, statements in the language of which would be interpreted as rough sets in some approximation space. Thus “models” in the semantics of such a system would be approximation spaces, equipped with a meaning function that assigns rough sets to well-formed formulae (wﬀs) of the language. Rough sets have been deﬁned in more than one way for a Pawlak approximation space (X, R) – [1] lists ﬁve deﬁnitions, all of which are equivalent to each other. One of these is most commonly used: (*) a rough set in (X, R), is the pair (A, A), for each A ⊆ X, where A, A denote the lower and upper approximations of A respectively. Another is a deﬁnition given by Pawlak in [2], and of interest to us in this paper:

The author acknowledges the support of the Department of Computer Science, University of Regina, Canada, during a visit to which the paper was ﬁnalized.

J.F. Peters et al. (Eds.): Transactions on Rough Sets VI, LNCS 4374, pp. 1–25, 2007. c Springer-Verlag Berlin Heidelberg 2007

2

M. Banerjee and Md.A. Khan

(**) A ⊆ X is a rough set in (X, R), provided the boundary of A, BnA = ∅. For generality’s sake, we could remove the restriction in (**) and consider deﬁnable sets (i.e. subsets with empty boundary) as special cases of rough sets. Thus, in the semantics based on approximation spaces, the meaning function deﬁning models, assigns to wﬀs either subsets of the domain, or pairs of subsets in accordance with (*) [3,4,5,6,7,8,9,10]. This is true even for semantics based on generalized approximation spaces, where diﬀerent relations (may be more than one in number, with operations on them) are considered [6,11]. The logics invariably involve modalities to express the concepts of lower and upper approximations – some are simply known normal modal logics, or have non-Boolean connectives (and no modalities) in the language, but there are translations into modal logics. We make a study of this group of systems in Section 2. It may be remarked that the “rough logic” proposed by Pawlak [3] (the ﬁrst system to be called so) makes an appearance here (cf. Section 2.6). The “practical” source of Pawlak approximation spaces are complete / deterministic information systems. These have the form S ≡ (U, A, V al, f ), where U is a set of objects, A a set of attributes, V al a set of values for the attributes, and f a function from U × A to V al. An equivalence relation RS is induced on U (thus giving the approximation space (U, RS )), as x RS y in U , if and only if f (x, a) = f (y, a), for all a ∈ A. The converse also holds: given any approximation space (U, R), one can deﬁne an information system S ≡ (U, A, V al, f ) such that the induced equivalence RS is just the relation R. So, in eﬀect, a semantics based on approximation spaces induced by complete information systems, is identical to the one discussed above. Generalized information systems, termed incomplete/nondeterministic, are those where f is a function from U × A to P(V al), and yields diﬀerent kinds of binary relations (e.g. similarity, inclusion – cf. Section 3.1) apart from equivalences, on U . Thus any information system (complete or incomplete) on a domain U , induces a relational system or a (generalized) approximation space on U , i.e. the (non-empty) set U together with a set of binary relations. This is called a standard structure on U [12,13,14]. For example, for the complete information system (U, A, V al, f ) above, (U, RS ) is a standard structure on U . In Section 3.1, (U, simS , inS ) is a standard structure for the incomplete information system S ≡ (U, A, V al, f ), with similarity and inclusion relations simS , inS . (Diﬀerent sets of relations can give diﬀerent standard structures on the same set U .) The induced relations in the standard structure may be characterized by a set of properties. As we know, equivalences are characterized by the properties of reﬂexivity, symmetry and transitivity. The similarity and inclusion relations considered in Section 3.1 are characterized by the properties (S1), (S2), (S4) − (S6) given there. By a general structure on U [12,13,14], one means any relational system comprising a non-empty set, along with binary relations that satisfy the set of properties characterizing the induced relations in the standard structure. Again, for the complete information system (U, A, V al, f ) above, any Pawlak approximation space (U, R) is a general structure. A general structure for S of

Propositional Logics from Rough Set Theory

3

Section 3.1, would be of the form (U, sim, in), where sim, in are binary relations on U satisfying (S1), (S2), (S4) − (S6). One ﬁnds logics with semantics deﬁned on incomplete information systems, for instance, in [15], or with semantics deﬁned on general structures [16]. However, Vakarelov [12,13,14,17] has established a series of characterization results, enabling an identiﬁcation of semantics based on general and standard structures (as in case of the Pawlak approximation space and complete information system above). In case of [15] too, we demonstrate here that the logic in question is equivalent to a normal modal logic with certain generalized approximation spaces deﬁning models. These systems are discussed in Section 3. In another line, there are “logics of information systems”, which accommodate in their language, expressions corresponding to objects and attributes [18,19,4,20]. Amongst these is a system that addresses the temporal aspect of information (cf. [4]), while [20] presents a logic for multiagent systems. There are also treatises on “rough relations” – a logic has been proposed [21] on the one hand, and on the other, we have the proposal of a logic programming language in “rough datalog” [22]. In Section 4, we brieﬂy sketch these and other approaches, such as rough mereology [23]. It will be seen that, some of the logics [4,16,20] have atomic propositions as (or built from) descriptors, the key feature of decision logic [2]. Decision logic is well-known, and not presented in this article. One should mention that a few of the logics described here, have also been used as a base to express various concepts involving rough sets. For instance, Yao and Lin [6] have deﬁned graded and probabilistic rough sets, using graded and probabilistic modal operators in the language of normal modal systems. Common and distributed knowledge operators have been interpreted in generalized approximation spaces by Wong [24]. In [25], another modal system (inspired by [3]) has been used to propose postulates for rough belief change. A comparative study of the presented logics is made in Section 5. The paper concludes by indicating possible future directions of investigation in Section 6.

2

Logics with Semantics Based on Approximation Spaces

In this section, we take a look at logics with approximation spaces deﬁning models. We ﬁnd six kinds of systems. For a logic L, “α is a theorem of L” shall be indicated by the notation L α. 2.1

Normal Modal Systems

The modal nature of the lower and upper approximations of rough sets was evident from the start. Hence, it is no surprise that normal modal systems were focussed upon, during investigations on logics for rough sets. In particular, in case of Pawlak rough sets, the two approximations considered as operators clearly obey all the S5 laws. The formal connection between the syntax of S5 and its semantics in terms of rough sets is given as follows [26]. According to the Kripke semantics for S5, a wﬀ α is interpreted by a function π as a subset in a non-empty domain U , the subset representing the extension

4

M. Banerjee and Md.A. Khan

of the formula – i.e. the collection of situations/objects/worlds where the wﬀ holds. Moreover, in an S5-model M ≡ (U, R, v) (say), the accessibility relation R is an equivalence on U . Further, if , ♦ denote the necessity and possibility operators respectively then for any wﬀ α, v(α) = v(α) and v(♦α) = v(α). A wﬀ α is true in M, if v(α) = U . Now it can easily be seen that all the S5 theorems involving and ♦ translate into valid properties of lower and upper approximations. Taking a cue from this connection, similar links have been pointed out (e.g. in [6,27]) between “rough sets” on generalized approximation spaces, and diﬀerent normal modal systems. The basic idea is to deﬁne generalized approximation operators corresponding to any binary relation R on the domain U – this has been done by many (e.g. for tolerance relations in [28] and others – cf. [29]). More explicitly, a map r : U → P(U ) is deﬁned as r(x) ≡ {y ∈ U : xRy}. Then the operators apr, apr : P(U ) → P(U ) are given by apr(A) ≡ {x : r(x) ⊆ A}, and apr(A) ≡ {x : r(x) ∩ A = ∅}. The rough set operators then satisfy various properties, depending upon the nature of R. Now let L denote a normal modal language, and M ≡ (U, R, v) be a model for L. v, as before, interprets a wﬀ as a subset in U . Then it is straightforward to observe that for any wﬀ α of L, v(α) = apr(v(α)), and dually, v(♦α) = apr(v(α)). By the above interpretation, the modal logics like KB, KT, K4, S5 etc. could be said to capture the properties of rough sets in generalized approximation spaces based on diﬀerent R (symmetric, reﬂexive, transitive, equivalence etc.). As remarked in the Introduction, this link has been made use of further. Considering graded and probabilistic modal operators on the above systems, graded and probabilistic rough sets have been deﬁned in [6]. Wong [24] has interpreted common and distributed knowledge operators (as deﬁned in logic of knowledge) in generalized approximation spaces with an indexed set of indiscernibility relations (corresponding to the knowledge operator of each agent). 2.2

DAL

[11] considers generalized approximation spaces containing a family of equivalence relations instead of just one. The logic DAL that is deﬁned in [11], has models based on these spaces. Further, the set of equivalence relations is assumed to be closed with respect to the operations of intersection and transitive closure of union of relations. The language of DAL, expectedly, includes a family of modal operators intended to correspond to the indiscernibility relations on the domains of the models. Formally, this is done by having a set R (say) of relational variables apart from the set P of propositional ones. There are binary operations ∩, , and a collection REL of relational expressions is built inductively out of the members of R with these operations. Apart from the classical Boolean connectives, a modal connective [R] is then introduced in the language for each R ∈ REL.

Propositional Logics from Rough Set Theory

5

A DAL-model is a structure U ≡ (U, {ρR }R∈REL , m), where, (i) for any R ∈ REL, ρR is an equivalence relation in the set U ; (ii) ρR∩S is the greatest equivalence relation in U included in both ρR and ρS ; (iii) ρRS is the least equivalence relation including both ρR and ρS ; and (iv) m is the meaning function from P ∪ R to P(U ) ∪ {ρR }R∈REL such that m(p) ⊆ U , for p ∈ P, and m(R) ≡ ρR , for R ∈ REL. For evaluating truth of wﬀs in DAL-models, one deﬁnes a function v that is determined by the meaning function m: v(p) ≡ m(p), for p ∈ P, v([R]α) ≡ {x ∈ U : y ∈ v(α), for all y such that x m(R) y}, the Boolean cases being deﬁned in the standard way. Deﬁnitions of truth and validity then are as usual: α is true in U, provided v(α) = U , and valid if it is true in all DAL-models. DAL has been axiomatized as follows. The connective is the dual of []. A1. All classical tautologies, A2. [R](α → β) → ([R]α → [R]β), A3. [R]α → α, A5. R α → [R] R α, A5. [R S]α → [R]α ∧ [S]α, A6. (([P ]α → [R]α) ∧ ([P ]α → [S]α)) → ([P ]α → [R S]α), A7. [R]α ∨ [S]α → [R ∧ S]α, A8. (([R]α → [P ]α) ∧ ([S]α → [P ]α)) → ([R ∧ S]α → [P ]α). The only rules of inference are Modus Ponens and Necessitation (corresponding to the connective [R] for each R ∈ REL). The axiomatization yields a completeness result with respect to the aforementioned semantics. Theorem 1. For any DAL-wﬀ α, DAL α, if and only if α is valid. 2.3

Pre-rough Logic

Following in the footsteps of Rasiowa, the algebra of rough sets was investigated in [7] in order to arrive at a logic for the theory. An algebraic structure called pre-rough algebra was proposed – this is a quasi Boolean algebra [30] along with a topological operator satisfying all the properties of an interior, and more. A corresponding logic P RL was framed, and observed to be sound and complete with respect to a semantics based on rough sets. The language of P RL has the primitive logical symbols ¬, , . , ♦ are duals of , , while ⇒ is deﬁned as: α ⇒ β ≡ (¬α β) (¬♦α ♦β), for any wﬀs α, β of P RL.

6

M. Banerjee and Md.A. Khan

As in the case of S5, a model for P RL is of the form M ≡ (U, R, v), where the departure from the S5-semantics lies in the deﬁnition of the meaning function v with respect to the connectives of conjunction and implication ⇒. For any α, β in P RL, S, T ⊆ U , v(α β) ≡ v(α) v(β), and v(α ⇒ β) ≡ v((¬α β) (¬♦α ♦β)), where c S T ≡ (S ∩ T ) ∪ (S ∩ T ∩ (S ∩ T ) ) (c denoting complementation). Deﬁnition of truth of a wﬀ α in M remains the same: this is if and only if v(α) = U . It may then be noticed that ⇒ reﬂects rough inclusion: a wﬀ α ⇒ β is true in (U, R, v) provided v(α) is roughly included in v(β). Further, / are operations that reduce to ordinary set intersection / union only when working on deﬁnable sets. α is valid (written |=RS α), if and only if α is true in every P RL-model. Following are the axiom schemes for P RL: 1. α ⇒ α 2a. ¬¬α ⇒ α 2b. α ⇒ ¬¬α 3. α β ⇒ α 4. α β ⇒ β α 5a. α (β γ) ⇒ (α β) (α γ) 5b. (α β) (α γ) ⇒ α (β γ) 6. α ⇒ α 7a. (α β) ⇒ (α) (β) 7b. (α) (β) ⇒ (α β) 8. α ⇒ α 9. ♦α ⇒ α 10a. (α β) ⇒ α β 10b. α β ⇒ (α β) Rules of inference : 1.

α 2. α⇒β α⇒β β⇒γ β α⇒γ modus ponens hypothetical syllogism

3.

α β⇒α

4.

α⇒β ¬β ⇒ ¬α

5.

α⇒β α⇒γ α⇒βγ

6.

α ⇒ β, β ⇒ α γ ⇒ δ, δ ⇒ γ (α ⇒ γ) ⇒ (β ⇒ δ)

7.

α⇒β α ⇒ β

8.

α α

9.

α ⇒ β ♦α ⇒ ♦β α⇒β

One can then prove, for any P RL-wﬀ α,

Propositional Logics from Rough Set Theory

7

Theorem 2. P RL α, if and only if |=RS α. We shall meet this logic and its semantics again in the coming sections. 2.4

3-Valued L ukasiewicz Logic L3

The connection of rough sets with 3-valuedness, also came up in the context of algebraic investigations. For example, in [31,32,33], an equivalence of 3-valued L ukasiewicz (Moisil) algebras with rough set structures was observed. In terms of logic, the way we can set up a formal link between the intensely studied L3 and a rough set semantics – in fact, the semantics just outlined in Section 2.3, is as follows. Let us recall Wajsberg’s axiomatization of L3 (cf. [34]). The logical symbols ¬, → are taken to be primitive. Axiom schemes: 1. 2. 3. 4.

α → (β → α). (α → β) → ((β → γ) → (α → γ)). ((α → ¬α) → α) → α. (¬α → ¬β) → (β → α).

The only rule of inference is Modus Ponens. L3 is known to be sound and complete with respect to the class of 3-valued L ukasiewicz (Moisil) algebras, as well as with respect to the semantics on 3 ≡ {0, 1/2, 1}, with L ukasiewicz negation and implication [34]. Now a logic L1 is said to be embeddable into a logic L2 , provided there is a translation of wﬀs of L1 into L2 , such that L1 α if and only if L2 α for any wﬀ α of L1 . We use the denotation L1 L2 . L1 L2 denotes existence of embeddings both ways. [31] establishes the following. There are translations ◦ from L3 into P RL and ∗ from P RL into L3 given by (¬α)◦ ≡ ¬α◦ , (α → β)◦ ≡ (♦¬α◦ β ◦ ) (♦β ◦ ¬α◦ ); (¬α)∗ ≡ ¬α∗ , (α β)∗ ≡ (α∗ → β ∗ ) → β ∗ , (α β)∗ ≡ ¬(¬α∗ ¬β ∗ ), (♦α)∗ ≡ ¬α∗ → α∗ . (One may notice that for any α, (α◦ )∗ and (α∗ )◦ are logically equivalent to α in the respective systems.) It is then shown that L3 P RL. Thus Theorem 3 (a) L3 α, if and only if |=RS α◦ , for an L3-wﬀ α and (b) L3 α∗ , if and only if |=RS α, for a P RL-wﬀ α.

8

2.5

M. Banerjee and Md.A. Khan

Logic for Regular Double Stone Algebras

Another line of algebraic investigation has resulted in linking rough set structures with the class of regular double Stone algebras [35]. A double Stone algebra (DSA) is a structure (L, , ,∗ ,+ , 0, 1) such that (L, , , 0, 1) is a bounded distributive lattice, y ≤ x∗ if and only if y x = 0, y ≥ x+ if and only if y x = 1 and x∗ x∗∗ = 1, x+ x++ = 0. The operations ∗ ,+ , as evident, are two kinds of complementation on the domain. The DSA is regular if, in addition to the above, for all x ∈ L, x x+ ≤ x x∗ holds. This is equivalent to requiring that x∗ = y ∗ , x+ = y + imply x = y, for all x, y ∈ L. Considering the deﬁnition (*) of rough sets (cf. Introduction), one ﬁnds that the collection RS of rough sets (X, X) over an approximation space (U, R) can be made into a regular DSA. The zero of the structure is the element (∅, ∅), while the unit is (U, U ). The operations , ,∗ ,+ are deﬁned as (X, X) (Y , Y ) ≡ (X ∪ Y , X ∪ Y ), (X, X) (Y , Y ) ≡ (X ∩ Y , X ∩ Y ), c c (X, X)∗ ≡ (X , X ), (X, X)+ ≡ (X c , X c ). For the converse, Comer shows that any regular DSA is isomorphic to a subalgebra of RS for some approximation space (U, R). untsch [8] as follows. Using these facts, a logic LD for rough sets is deﬁned by D¨ The language of LD has two unary connectives ∗ ,+ (for two kinds of negation), apart from the binary connectives ∨, ∧ and constant symbol . We write α∗ , α+ instead of ∗ α,+ α, just to keep parity with the algebraic notation used above. A model of LD is a pair (W, v), where W is a (non-empty) set and v is the meaning function assigning to propositional variables, pairs in P(W ) × P(W ) such that if v(p) = (A, B) then A ⊆ B. v(p) = (A, B) is to express that “p holds at all states of A and does not hold at any state outside B”. For , we have v() ≡ (W, W ). v is extended to the set of all wﬀs recursively: if v(α) = (A, B) and v(β) = (C, D) then v(α ∨ β) ≡ (A ∪ C, B ∪ D), v(α ∧ β) ≡ (A ∩ C, B ∩ D), v(α∗ ) ≡ (B c , B c ), v(α+ ) ≡ (Ac , Ac ). A wﬀ α is true in a model (W, v), provided v(α) = (W, W ). We would now like to make explicit, how v interprets the wﬀs of LD as rough sets over some approximation space. One refers to [8], and [35].

Propositional Logics from Rough Set Theory

9

Consider the range ran(v) of the map v in P(W ) × P(W ). It can be shown that it forms a regular DSA through the operations , ,∗ ,+ : v(α) v(β) ≡ v(α ∨ β), v(α) v(β) ≡ v(α ∧ β), v(α)∗ ≡ v(α∗ ), v(α)+ ≡ v(α+ ). v(∗ ) (or v(+ )) is the zero ((∅, ∅)) of the algebra, while v() = (W, W ) is the unit. In fact, the variety of regular DSA’s is just the one generated by regular DSA’s of the kind ran(v), where v ranges over all meaning functions for all models. Using the correspondence between classes of algebras and logic [36], [8] concludes, amongst other properties of LD , that Theorem 4. LD has a ﬁnitely complete and strongly sound Hilbert style axiom system. Through Comer’s representation result, ran(v) corresponding to any model (W, v) of LD , is isomorphic to a subcollection of RS for some approximation space (U, R). We can now say that v(α) for a wﬀ α, can be identiﬁed with a rough set over some (U, R) in precisely the following manner. Let U consist of all the join irreducible elements of ran(v), i.e. v(α) ∈ U , if and only if v(α) = (∅, ∅), and for all wﬀs β, γ, if v(α) = v(β) v(γ) then either v(α) = v(β) or v(α) = v(γ). An equivalence relation R on U can then be obtained, where R is given by: v(α) R v(β) if and only if v(α∗∗ ) = v(β ∗∗ ), i.e. if and only if B = D, where v(α) = (A, B) and v(β) = (C, D). Now deﬁne f : ran(v) → P(U ) such that for v(α) = (A, B), f (A, B) ≡ {v(β) = (C, D) ∈ U : C ⊆ A, D ⊆ B}. Finally, deﬁne the map g : ran(v) → P(U ) × P(U ) as: g(A, B) ≡ (f (A, A), f (B, B)), where v(α) = (A, B). (Note that (A, A), (B, B) ∈ U , as v(α++ ) = (A, A), and v(α∗∗ ) = (B, B).) It can then be shown that (a) g is injective, and (b) g preserves , ,∗ ,+ . Moreover, if v(α) = (A, B), g(v(α)) = ( f (A, B), f (A, B) ), a rough set in the approximation space (U, R). [8] does not present an explicit proof method for the logic LD – the only comment on the matter is vide Theorem 4. Recently, Dai [9] has presented a sequent calculus for a logic (denoted RDSL) with a semantics based on the regular DSAs formed by collections of rough sets of the kind RS over some approximation space (U, R) (deﬁned earlier in the section). The language of RDSL is the same as that of LD , except that the constant symbol ⊥ (dual for ) is included amongst the

10

M. Banerjee and Md.A. Khan

primitive symbols. Models are of the form (RS, v), where v, the meaning function, is a map from the set of propositional variables to RS. Thus v(p), for a propositional variable p, is a pair (X, X) in the approximation space (U, R). v is extended to the set of all wﬀs in the same way as for models of LD . We note that an RDSL-model (RS, v) may be identiﬁed with the LD -model (U, v). On the other hand, due to Comer’s representation result, given any LD -model (W, v), there is an isomorphism f from ran(v) to a subalgebra (S, say) of RS on some approximation space. One can thus ﬁnd an RDSLmodel (RS, v ) such that ran(v ) is S, i.e. v (p) ≡ f (v(p)), for every propositional variable p. So, in this sense, the classes of models of the two logics are identiﬁable. As in classical sequent calculus, for ﬁnite sequences of wﬀs Γ ≡ (p1 , p2 , . . . pm ) and Δ ≡ (q1 , q2 , . . . qn ) in RDSL, the sequent Γ ⇒ Δ is said to be valid in a model (RS, v) if and only if v(p1 ) . . . v(pm ) ≤ v(q1 ) . . . v(qn ). , are the operations in the regular DSA (RS, , ,∗ ,+ , < ∅, ∅ >, < U, U >). Γ ⇒ Δ is said to be valid (in notation, |=RDSA Γ ⇒ Δ) if and only if Γ ⇒ Δ is valid in every RDSL-model. The standard classical axiom p ⇒ p and rules for the connectives ∧, ∨ and constant symbols , ⊥ are considered to deﬁne derivability (RDSL ). In addition, the axioms and rules for the two negations ∗ ,+ are as follows. 1. p ⇒ p∗∗ . 2. p∗ ⇒ p∗∗∗ . 3. p ⇒ p++ . 4. p+ ⇒ p+++ . (R∗ ) Γ ⇒ Δ (R+ ) Γ ⇒ Δ Δ∗ ⇒ Γ ∗ Δ+ ⇒ Γ + Soundness and completeness are then proved, with respect to the semantics sketched. Theorem 5. RDSL Γ ⇒ Δ, if and only if |=RDSA Γ ⇒ Δ. 2.6

Logic for Rough Truth or of Rough Consequence

In [3], a logic Rl (the ﬁrst in literature to be called “rough logic”) was proposed, along with a very appealing notion of rough truth. The language of Rl consists of the standard Boolean connectives, and models M ≡ (U, R, v) are based on approximation spaces. v assigns subsets of the domain U to wﬀs in the usual manner. Five logical values of “truth”, “falsity”, “rough truth”, “rough falsity” and “rough inconsistency” are considered in this work, with truth and falsity representing the limit of our partial knowledge. As we know, a wﬀ α is true in M, if v(α) = U . α is said to be surely/possibly true on x ∈ U , if x ∈ v(α) (v(α)) respectively. α is roughly true in M, if it is possibly true on every x in U , i.e. v(α) = U , or in other words, v(α) is

Propositional Logics from Rough Set Theory

11

externally indiscernible [37] in (U, R). On the other hand, α is roughly false, when v(α) = ∅ (v(α) is internally indiscernible), and α is roughly inconsistent, if it is both roughly true and false (v(α) is totally indiscernible). Let us consider the modal system S5. Note that models of S5 and Rl are identical. We can then eﬀect a translation of the above concepts into S5. In(U, R, v), a wﬀ α can be termed roughly true if v(α) = v(♦α) = U , roughly false if v(α) = v(α) = ∅, and roughly inconsistent if both hold. In [10], a logic Lr having the same models as above was proposed, with the speciality that the syntax-semantics relationships are explored with rough truth replacing truth and rough validity replacing validity. The notion of consistency is replaced by one of rough consistency too. The consequence relation deﬁning the logic is also non-standard. These ideas were ﬁrst mooted in [5,26], and Lr is a modiﬁed version of the formal system discussed there. Lr has a normal modal language. A model M ≡ (U, R, v) is a rough model of Γ , if and only if for every γ ∈ Γ , v(♦γ) = U , i.e. γ is roughly true in M. α is a rough semantic consequence of Γ (denoted Γ |≈α) if and only if every rough model of Γ is a rough model of α. If Γ is empty, α is said to be roughly valid, written |≈α. There are two rules of inference: R1 . if S5

α R2 . ♦α β ♦β ♦α → ♦β ♦α ∧ ♦β

The consequence relation is deﬁned as follows. Let Γ be any set of wﬀs and α any wﬀ in Lr . α is a rough consequence of Γ (denoted Γ |∼α) if and only if there is a sequence α1 , ..., αn (≡ α) such that each αi (i = 1, ..., n) is either (i) a theorem of S5, or (ii) a member of Γ , or (iii) derived from some of α1 , ..., αi−1 by R1 or R2 . If Γ is empty, α is said to be a rough theorem, written |∼α. A kind of “rough Modus Ponens” is then derivable, in the form: if Γ |∼α, S5 α → β with S5 α ≈ α then β. Here ≈ reﬂects the notion of “rough equality”, α ≈ β ≡ (α ↔ β) ∧ (♦α ↔ ♦β). One also obtains soundness of Lr with respect to the above semantics: if Γ |∼α then Γ |≈α. It is clear that in the face of an incomplete description of a concept p, p and “not” p (in the classical sense) may not always represent conﬂicting situations. To accommodate this possibility, a set Γ of wﬀs is termed roughly consistent if and only if the set ♦Γ ≡ {♦γ : γ ∈ Γ } is S5-consistent. With the help of this notion, one obtains Theorem 6. (Completeness) (a) Γ is roughly consistent if and only if it has a rough model. (b) For any Lr -wﬀ α, if Γ |≈α then Γ |∼α. Thus, Lr appears as another system that is able to address rough sets and related notions. We shall remark on its relationship with other well-known systems in Section 5. It may be mentioned that Lr has been used as the base logic for a proposal of rough belief change in [25].

12

3

M. Banerjee and Md.A. Khan

Logics with Semantics Based on Information Systems

We now present logics, the models of which are deﬁned on approximation spaces induced by information systems. We ﬁnd one pioneering system N IL that has inspired the proposal of many others in the same line. The section also includes a logic by Nakamura, the models of which are directly deﬁned on information systems. 3.1

N IL

Recall that an incomplete information system is of the form S ≡ (U, A, V al, f ), where U is a set of objects, A a set of attributes, V al a set of values for the attributes, and f a function from U × A to P(V al). The logic N IL proposed by Orlowska and Pawlak [16] works on incomplete information systems, in which the function f satisﬁes an additional condition: () f (x, a) = ∅, for all x ∈ U, a ∈ A. One observes that, given S ≡ (U, A, V al, f ), two particular kinds of binary relations on the domain U are induced – these dictate the formulation of N IL. Let x, y ∈ U . Similarity (simS ): x simS y if and only if f (x, a) ∩ f (y, a) = ∅, for all a ∈ A. Inclusion (inS ): x inS y if and only if f (x, a) ⊆ f (y, a), for all a ∈ A. It can be shown that for every incomplete information system S ≡ (U, A, V al, f ) and x, y, z ∈ U , the following hold. (S1) x inS x. (S2) if x inS y and y inS z then x inS z. (S3) if x simS y for some y, then x simS x. (S4) if x simS y then y simS x. (S5) if x simS y, x inS u, y inS v then u simS v. Further, if the condition () is satisﬁed by f then sim satisﬁes (S6) x simS x. Thus a standard structure (cf. Introduction) corresponding to an incompleteinformation system S ≡ (U, A, V al, f ) with condition (), would be (U, simS , inS ). On the other hand, a general structure for S would be of the form (U, sim, in), where sim, in are binary relations on U satisfying (S1), (S2), (S4) − (S6). For brevity, we refer to these as standard and general N IL-structures respectively. N IL could be termed as a modal version of decision logic introduced by Pawlak [2], an association similar to that of rough logic [3] and S5 (cf. Section 2.6). The atomic propositions of N IL are the descriptors of decision logic – of the form (a, v), where a is an “attribute constant”, and v a constant representing “value of attribute”. Apart from the standard Boolean connectives ¬, ∨, the language contains modal connectives , 1 , 2 corresponding to sim, in and the inverse in−1 of in respectively. Wﬀs are built, as usual, out of the atomic propositions (descriptors) and the connectives. Note that there are no operations on the attribute or value constants.

Propositional Logics from Rough Set Theory

13

A N IL-model M ≡ (U, sim, in, m) consists of a general structure (U, sim, in) as above, along with a meaning function m from the set of all descriptors to the set P(U ). m is extended recursively to the set of all N IL-wﬀs in the usual manner. In particular, m(α) ≡ {x ∈ U : y ∈ m(α) for all y such that x sim y}. Similarly one deﬁnes m(1 α), and m(2 α). α is true in the model M, if m(α) = U. The following deductive system for N IL was proposed in [16]. Axiom schemes: A1. All classical tautologies, A2. 2 (α → β) → (2 α → 2 β), A3. 1 (α → β) → (1 α → 1 β), A4. (α → β) → (α → β), A5. α → 1 ¬2 ¬α, A6. α → 2 ¬1 ¬α, A7. 2 α → α, A8. 1 α → α, A9. α → α, A10. 2 α → 2 2 α, A11. 1 α → 1 1 α, A12. α → ¬¬α, A13. α → 2 1 α. Rules of inference:

(R1) α, α → β (R2) α β 2 α (R3)

α 1 α

(R4) α α

It has been proved that Theorem 7. For any N IL-wﬀ α, N IL α if and only if α is true in all N ILmodels. 3.2

Logics by Vakarelov

Vakarelov addresses the issue of completeness of various logics, the models of which are based on standard structures corresponding to some information system. For instance, in the case of N IL, the question would be about a completeness theorem with respect to the class of N IL-models deﬁned on standard N IL-structures (cf. Section 3.1). In [12], such a theorem is proved, via a key

14

M. Banerjee and Md.A. Khan

characterization result. In fact, this result set the ground for a series of similar observations when the binary relations involved are changed. Proposition 1. (Characterization) Let (U, sim, in) be a general N IL-structure. Then there exists an information system S ≡ (U, A, V al, f ) with f satisfying (), such that simS = sim and inS = in. In other words, the classes of N IL-models based on standard and general N ILstructures are identical. Hence one obtains the required completeness theorem. The condition (), viz. f (x, a) = ∅ for all x ∈ U, a ∈ A, is a restrictive one. However, it is observed by Vakarelov that even if this condition is dropped, a characterization result similar to Proposition 1 can be obtained. Instead of reﬂexivity of sim (cf. property (S6), Section 3.1), we now have just the condition of quasireﬂexivity – cf. property (S3): if x sim y for some y, then x sim x. The corresponding logic can be obtained from N IL by replacing the axiom A9 by ¬(p ∧ ¬p) → (α → α). Following this approach, one handles the cases of incomplete information systems inducing diﬀerent binary relations. For example, [14,13,17] consider these relations amongst others, for S ≡ (U, A, V al, f ): Indiscernibility (indS ): x indS y if and only if f (x, a) = f (y, a), for all a ∈ A, w Weak indiscernibility (indw S ): x indS y if and only if f (x, a) = f (y, a), for some a ∈ A, w Weak similarity (simw S ): x simS y if and only if f (x, a) ∩ f (y, a) = ∅, for some a ∈ A. Complementarity (com): x com y if and only if f (x, a) = (V ALa \ f (y, a)), for all a ∈ A, where V ala is the value set for the particular attribute a, and V al ≡ ∪{V ala : a ∈ A}. The characterization result for each has been obtained, the corresponding logical system is deﬁned and the completeness theorem with respect to models on the intended standard structures is proved. 3.3

Logic by Nakamura

[15] discusses a logic with models on incomplete information systems. We recall (cf. Introduction) that given a complete information system S ≡ (U, A, V al, f ), one can deﬁne the equivalence relation RS . The lower approximation of X(⊆ U ) under this relation is denoted as X S , and its upper approximation as X S . Nakamura deﬁnes a completation S0 of an incomplete information system S as a complete information system that can be constructed from S by selecting any one value from f (x, a)(⊆ V al), for each x ∈ U, a ∈ A. If f (x, a) = ∅, one selects a special symbol . The relationship of S0 and S is expressed as S0 ≥ S. Now the “lower” and “upper approximations” X, X of X ⊆ U in an incomplete information system S ≡ (U, A, V al, f ) are deﬁned as follows: (∗)

X ≡ ∩S0 ≥S X S0 , X ≡ ∪S0 ≥S X S0 .

Propositional Logics from Rough Set Theory

15

With this background, a logic IN CRL is proposed, having the standard Boolean connectives, and two modal operators [], (corresponding to “surely” and “possibly” respectively). An IN CRL-model is an incomplete information system S ≡ (U, A, V al, f ) along with a meaning function vS from the set of propositional variables of the language to P(U ). vS is extended as usual for the wﬀs involving Boolean connectives. For wﬀs with modal operators, one makes use of completations S0 of S and the preceding deﬁnitions of lower and upper approximations given in (∗). vS ([ ]α) ≡ ∩S0 ≥S vS (α)S

0

= vS (α),

vS ( α) ≡ ∪S0 ≥S vS (α)S0 = vS (α). Truth and validity of wﬀs are deﬁned again as for most of the previous systems. Nakamura points out relationships of IN CRL with the modal system KT B, in particular that all theorems of KT B are valid wﬀs of IN CRL. We shall take a further look at the two logics in Section 5.

4

Other Approaches

This section outlines a few proposals of logics related to rough sets, the models of which are based on structures that are even more generalized than the ones already presented. As we shall see, these logics have dimensions not accounted for in the systems presented so far. 4.1

Temporal Approach

Orlowska (cf. [4]), deﬁnes a logic LT with models on dynamic information systems, in order to deal with the temporal aspect of information. A set T of moments of time, and a suitable relation R on the set T are considered along with the set U of objects and A of attributes. Formally, a dynamic information system is a tuple S ≡ (U, A, V al, T, R, f ), where V al ≡ ∪{V ala : a ∈ A}, (V ala , as in Section 3.2, being the value set for the particular attribute a) and the information function f : U × T × A → V al satisﬁes the condition that f (x, t, a) ∈ V ALa , for any x ∈ U, t ∈ T, a ∈ A. In the language of LT , atomic statements are descriptors of decision logic, together with an object constant x – so these are triples (x, a, v), and are intended to express: “object x assumes value v for attribute a”. There are modal operators to reﬂect the relations R and R−1 . The truth of all statements of the language is evaluated in a model based on a dynamic information system, with respect to moments of time, i.e. members of the set T . An LT -model is a tuple M ≡ (S, m) where S is a dynamic information system, and m a meaning function which assigns objects, attributes and values from U, A, V al to the respective constants. The satisﬁability of a formula α in a model M at a moment t(∈ T ) of time is deﬁned inductively as follows: M, t |= (x, a, v) if and only if f (m(x), t, m(a)) = m(v).

16

M. Banerjee and Md.A. Khan

For the Boolean cases, we have the usual deﬁnitions. For the modal case, M, t |= [R]α if and only if for all t ∈ T , if (t, t ) ∈ R then M, t |= α. A wﬀ is true in M, provided it is satisﬁed in M at every t ∈ T . LT is complete with respect to this class of models, for the axioms of linear time temporal logic, and an axiom which says that the values of attributes are uniquely assigned to objects. 4.2

Multiagent Systems

[20] describes a logic, that takes into account a (ﬁnite) collection of agents and their knowledge bases. We denote the logic as LMA . The language of LMA has “agent constants” along with two special constants 0,1. Binary operations +,. are provided to build the set T of terms from these constants. Wﬀs of one kind are obtained from terms, and are of the form s ⇒ t, s, t ∈ T , where ⇒ is a binary relational symbol. s ⇒ t is to reﬂect that “the classiﬁcation ability of agent t is at least as good as that of agent s”. Furthermore, there are attribute as well as attribute-value constants. Descriptors formed by these constants constitute atomic propositions, and using connectives ∧, ¬ and modal operators It , t ∈ T (representing “partial knowledge” of each agent), give wﬀs of another kind. LMA -models are not approximation spaces, but what could be called “partition spaces” on information systems. Informally put, a model consists of an information system S ≡ (U, A, V al, f ), and a family of partitions {Et }t∈T on the domain U – each corresponding to the knowledge base of an agent. The family is shown to have a lattice structure, and the ordering involved gives the interpretation of the relational symbol ⇒. Wﬀs built out of descriptors are interpreted in the standard way, in the information system S. The partial knowledge operator It for a term t reﬂects the lower approximation operator with respect to the partition Et on U . An axiomatization of LMA is presented, to give soundness and completeness results. In the context of multiagent systems, it is worth mentioning the approach followed in [38], even though a formal logic based on it has not been deﬁned yet. Property systems (P -systems) are deﬁned as triples of the form (U, A, |=), where U is a set of objects, A a set of properties, and |= a “fulﬁlment” relation between U and A. For each P -system P, a collection P op of interior and closure operators satisfying speciﬁc properties are considered. These operators could be regarded as generalizations of lower and upper approximations. Now given a family {Pk }k∈K of P -systems (each for an agent, say) over some index set K and over the same set U of objects, one obtains a multiagent pre-topological approximation space as a structure (U, {Pkop }k∈K ). It is to be seen if such a generalized structure could form the basis of a semantics of some formal logical framework. 4.3

Rough Relations

Discussion about relations on approximation spaces, started from [39]. We ﬁnd two directions of work on this topic.

Propositional Logics from Rough Set Theory

17

Logic of Rough Relations: [40] considers another generalization of the notion of an approximation space – taking systems of the form AS ≡ (U, I, v), where U is a non-empty set of objects, I : U → P(U ) an uncertainty function, and v : P(U ) × P(U ) → [0, 1] is a rough inclusion function satisfying the following conditions: v(X, X) = 1 for any X ⊆ U , v(X, Y ) = 1 implies v(Z, Y ) ≥ v(Z, X) for any X, Y, Z ⊆ U, v(∅, X) = 1 for any X ⊆ U . For any subset X of U , we then have the lower and upper approximations: L(AS, X) ≡ {x ∈ U : v(I(x), X) = 1}, U (AS, X) ≡ {x ∈ U : v(I(x), X) > 0}. A ‘rough set’ in AS is the pair (L(AS, X), U (AS, X)). The above is motivated from the fact that any Pawlak approximation space (U, R) is an instance of a generalized space as just deﬁned. Indeed, we consider the function I that assigns to every object its equivalence class under R, and the inclusion function v as: v(S, R) ≡

card(S∩R) card(S)

1

if S = ∅ if S = ∅

For an approximation space AS ≡ (U, I, v) with U = U1 × U2 and v as in the special case above, [21] discusses relations R ⊆ U1 × U2 . The lower and upper approximation of R in AS are taken, and a rough relation is just a rough set in AS. A decidable multimodal logic is proposed – for reasoning about properties of rough relations. The modal operators correspond to a set of relations on the domain of the above generalized approximation spaces, as well as the lower and upper approximations of these relations. An axiomatization for the logic is given, and completeness is proved with respect to a Kripke-style semantics. Rough Datalog: Just as decision tables [2] are (complete) information systems with special attributes, viz. the decision attributes, [22] considers a decision system (U, A ∪ {d}) – but with a diﬀerence. Each attribute a in A is a partial map from U to a value set Va , and d, the decision attribute, is a partial map from U to {0, 1}. It is possible that for some x ∈ U , all attribute values (including the value of d) are undeﬁned. A ‘rough set’ X is taken to be a pair (X + , X − ), where X + is the set of elements of U that may belong to X, while X − contains those elements of U that may not belong to X. d indicates the information about membership of an object of U in X. Formally, let A ≡ {a1 , ..., an }, A(x) ≡ (a1 (x), ..., an (x)) for each x ∈ U , and A−1 (t) ≡ {x ∈ U : A(x) = t}, for t ∈ Va1 × ... × Van . (Note that for some x ∈ U , A(x) could be undeﬁned). Then X + ≡ {x ∈ U : A is deﬁned for x, and d(x ) = 1, for some x ∈ A−1 (A(x))}, and X − ≡ {x ∈ U : A is deﬁned for x, and d(x ) = 0, for some x ∈ A−1 (A(x))}.

18

M. Banerjee and Md.A. Khan

This deﬁnition implies that X + and X − may not be disjoint, allowing for the presence of conﬂicting (contradictory) decisions in the decision table. On the other hand, X + and X − may not cover U either, allowing for the possibility that there is no available information about membership in X. With these deﬁnitions, ‘rough relations’ are considered in [22]. Standard relational data base techniques, such as relational algebraic operations (e.g. union, complement, Cartesian product, projection) on crisp relations, are extended to the case of rough relations. A declarative language for deﬁning and querying these relations is introduced - pointing to a link of rough sets (as just deﬁned) with logic programming. 4.4

Logics with Attribute Expressions

As we have seen, LT and LMA (cf. Sections 4.1 and 4.2 respectively) have attribute expressions in the language that are interpreted in information systems. N IL (cf. Section 3.1), also has attribute constants in the language. But unlike the models of LT and LMA , the standard or general N IL-structures deﬁning N IL-models do not accommodate attributes, and the wﬀs (which are built using the attribute constants) point to collections of objects of the domain. A class of logics with attribute expressions are also deﬁned in [18,19]. Models are based on structures of the form (U, A, {ind(P )}P ⊆A ), where the “indiscernibility” relation ind(P ) for each subset P of the attribute set A, has to satisfy certain conditions. For the models of one of the logics, for example, the following conditions are stipulated for ind(P ): (U 1) ind(P ) is an equivalence relation on U , (U 2) ind(P ∪ Q) = ind(P ) ∩ ind(Q), (U 3) if P ⊆ Q then ind(Q) ⊆ ind(P ), and (U 4) ind(∅) = U × U . Other logics may be obtained by changing some of (U 1) − (U 4). The language of the logics has a set of variables each representing a set of attributes, as well as constants to represent all one element sets of attributes. Further, the language can express the result of (set-theoretic) operations on sets of attributes. The logics are multimodal – there is a modal operator to reﬂect the indiscernibility relation for each set of attributes as above. A usual Kripke-style semantics is given, and a number of valid wﬀs presented. However, as remarked in [19], we do not know of a complete axiomatization for such logics. 4.5

Rough Mereology

This is an approach inspired by the theory of mereology due to Le´sniewski (1916). Le´sniewski propounds a theory of sets that has containment as the primitive relation, rather than membership. Drawing from this classical theory, rough mereology has been proposed [23], providing a useful notion of rough containment, of “being a part, in a degree”. Formally, this can be deﬁned as a real binary function μ on the domain with values in [0,1], satisfying certain conditions (abstracted from the properties of

Propositional Logics from Rough Set Theory

19

classical containment). A given information system (U, A, V al, f ), a partition of A into, say A1 , ..., An and a set of weights {w1 , ..., wn }, generate μ(x, y), can n x, y ∈ U . It is assumed that wi ∈ [0, 1], i = 1, ..., n, and i=1 wi = 1. A pre-rough inclusion μo is ﬁrst deﬁned: μo (x, y) ≡ ni=1 wi .(|indi (x, y)|/|Ai |), where indi (x, y) ≡ {a ∈ Ai : f (x, a) = f (y, a)}. μo can then be extended to rough inclusion μ over P(U ) by using t-norms and t-conorms. Rough inclusion can be used, for instance, in specifying approximate decision rules. It may be remarked that predicate logics corresponding to rough inclusions have been proposed recently in [41].

5

Comparative Study

We now discuss some relationships between the logics presented in Sections 2 and 3. 5.1

Embeddings

Let us recall the notion of an embedding of logics – cf. Section 2.4. We consider the logics P RL, L3, LD , RDSL presented in Sections 2.3, 2.4 and 2.5 respectively, and point out interrelationships, as well as relations with other known logics. (1) L3 P RL: This has already been seen in Section 2.4. untsch and Pagliani, (2) L3 LD : As summarized in [1] and observed by D¨ regular double Stone algebras and 3-valued L ukasiewicz algebras are equivalent to each other via suitable transformations. Passing on to the respective logics, we would thus ﬁnd embeddings both ways, between LD and L3. (3) LD RDSL: We can deﬁne, in RDSL, that a wﬀ α is a theorem (valid), if and only if the sequent ⇒ α is derivable (valid). Using the formal argument made in Section 2.5 to show that the classes of models of the logics LD and RDSL are identiﬁable and Theorems 4, 5, one gets the result with the identity embedding. (4) L3 LSN : LSN denotes constructive logic with strong negation [30]. We note that semi-simple Nelson algebras are the algebraic counterparts for ukasiewicz LSN . The equivalence of semi-simple Nelson algebras and 3-valued L algebras through suitable translations has also been observed e.g. by Pagliani. Hence the stated embedding. (5) P RL S5: One observes [31] a translation of wﬀs of P RL into S5 that assigns the operations of negation ¬ and necessity in P RL those same operations of S5. Further, is translated in terms of the conjunction ∧ and disjunction ∨ of S5 as:

20

M. Banerjee and Md.A. Khan

(α β) ≡ (α ∧ β ) ∨ (α ∧ M β ∧ ¬M (α ∧ β )). Then it can be shown that P RL α if and only if α , for any wﬀ α of P RL. (6) S5 Lr : The logic Lr for rough truth is able to capture, as the class of its theorems, exactly the “♦-image” of the class of S5-theorems, i.e. S5 ♦α if and only if |∼α [5,10]. Note that the languages of Lr and S5 are the same. We translate α in S5 to α∗ ≡ Lα. Then α if and only if |∼α∗ . For the other direction, we consider the translation α◦ ≡ M α. (7) J Lr : In 1948, Ja´skowski proposed a “discussive” logic – he wanted a formalism to represent reasoning during a discourse. Each thesis, a discussive assertion of the system, is supposed either to reﬂect the opinion of a participant in the discourse, or to hold for a certain “admissible” meaning of the terms used in it. Formally, any thesis α is actually interpreted as “it is possible that α”, and the modal operator ♦ is used for the expression. The logic J (cf. [42]) is such a system. The J-consequence, deﬁned over S5, is such that: J α if and only if S5 ♦α. Because of the relationship between Lr and S5 noted in (6) above, we have J Lr with the identity embedding. In the whole process, one has obtained an alternative formulation of the paraconsistent logic J (proposed in a diﬀerent context altogether), and established a link between Pawlak’s and Ja´skowski’s ideas. 5.2

KT B and Nakamura’s Logic IN CRL

We refer to Section 3.3, and present a connection between IN CRL, and the normal modal system KT B. KT B, as we know, is sound and complete with respect to the class of reﬂexive and symmetric Kripke frames. Let S ≡ (U, A, V al, f ) be an incomplete information system, and let us consider the relation on U deﬁned as follows: x y if and only if there exists a completation S0 of S such that x RS0 y. Clearly is reﬂexive and symmetric, but not transitive. From the deﬁnitions of vS ([ ]α) and vS ( α), we see that x ∈ vS ([ ]α) if and only if, for all y ∈ U such that x y, y ∈ vS (α), and x ∈ vS ( α) if and only if, there exists y ∈ U such that x y and y ∈ vS (α). So all provable wﬀs of the modal logic KT B are valid in IN CRL. What about the converse – are all valid wﬀs of IN CRL provable in KT B? [15] makes a cryptic comment about this, we establish the converse here. KT B provides an axiomatization for IN CRL: We show that if α is not provable in KT B then it is not valid in IN CRL. It suﬃces then, to construct an incomplete information system S ≡ (U, A, {V ala }a∈A , f ) for any given KT Bframe (W, R), such that is identical with R. Let g be a function from R (⊆ W × W ) to some set C of constants, satisfying the following conditions: (i) g(x, y) = g(y, x), (ii) g(x, y) = g(t, z) implies that either x = t and y = z, or x = z and y = t.

Propositional Logics from Rough Set Theory

21

(g essentially assigns, upto symmetry, a unique constant from C to every pair in R.) Now consider U ≡ W, A ≡ {a}, where a is a new symbol. Further, deﬁne f (x, a) ≡ {g(x, y) : y ∈ U and (x, y) ∈ R}, so that V ala ⊆ C. We claim that xRy if and only if x y. Suppose xRy. Then g(x, y) ∈ f (x, a)∩ f (y, a) and hence x y. Conversely, if x y, there exists d ∈ f (x, a) ∩ f (y, a). Now d ∈ f (x, a) implies that d = g(x, z), for some z ∈ U such that (x, z) ∈ R, and d ∈ f (y, a) implies that d = g(y, t), for some t ∈ U such that (y, t) ∈ R. From the property of g, it follows that either x = y or x = t, whence by reﬂexivity and symmetry of R, we get xRy. The proof above, in fact, yields a characterization theorem, viz. given any reﬂexive, symmetric frame (W, R), there exists an incomplete information system S ≡ (U, A, {V ala }a∈A , f ) satisfying the condition () (cf. Section 3.1) such that R = = simS . 5.3

Normal Modal Systems and Vakarelov’s Logics

Vakarelov has proved the characterization theorem for incomplete information systems with respect to diﬀerent sets of relations [12,14,13,17]. As we have remarked in the Introduction, a special case would be obtained with respect to the indiscernibility relation on the Pawlak approximation space. One ﬁnds that if we restrict the logics presented in [14,13,17] to take a modal operator corresponding only to the indiscernibility relation, the resulting system would be just the modal logic S5. As noted at the end of Section 5.2, if an incomplete information system satisﬁes the condition (), then the similarity relation simS is the same as the relation . So it follows that if we restrict the logic N IL to take only the modality in the language then the corresponding logic will be just IN CRL, or, in other words, KT B. 5.4

DAL Again

Observing Vakarelov’s strain of work, it may be tempting to look for a kind of characterization result in the case of DAL (cf. Section 2.2) as well. Consider a general DAL-structure U ≡ (U, {Ri }i∈I ), where the family {Ri }i∈I of equivalence relations is closed under intersection and transitive closure of union. Can one ﬁnd an incomplete information system S ≡ (U, A, V al, f ) such that the standard structure for S is just U? Let us assume that the standard structure is obtained “naturally” from S, viz. that the equivalence relations in it are the ones induced by the subsets of A. As it turns out, this is a hard question. However, we can ﬁnd an information system, such that the standard structure obtained from it in the above manner cannot be a general DAL-structure. Suppose for some incomplete information system S ≡ (U, A, V al, f ), R and P are the equivalence relations induced by subsets R , P of A respectively – we denote this as ind(R ) = R and ind(P ) = P . For the equivalence relation

22

M. Banerjee and Md.A. Khan

R ∩ P , R ∪ P ⊆ A is such that ind(R ∪ P ) = R ∩ P . But in the case of R P , there may not be any Q ⊆ A such that ind(Q) = R P. Consider the following example [11]. Example 1. U ≡ {o1, o2, o3, o4, o5, o6, o7}, where each oi consists of circles and squares. Let A ≡ {number of circles (), number of squares ()}. The information function is given by the following table: ————————————————————– ————————————————————– o1 1 1 o2 1 2 o3 2 1 o4 2 2 o5 3 3 o6 3 4 o7 3 4 ————————————————————— Equivalence classes of indiscernibility relations ind() and ind() are: ind() : {o1, o2}, {o3, o4}, {o5, o6, o7}, ind() : {o1, o3}, {o2, o4}, {o5}, {o6, o7}. The transitive closure of these relations gives the following equivalence classes: ind() ind() : {o1, o2, o3, o4}, {o5, o6, o7}. Clearly there is no Q ⊆ A such that ind(Q) = ind() ind().

6

Summary and Questions

We have tried to present the various proposals of logics with semantics based on rough sets, including some generalizations. Two main approaches emerge, discussed in Sections 2 and 3. One of these considers logics, the models of which are approximation spaces, while the other considers approximation spaces, but those induced by information systems. However, it is found through characterization results, that both lines of study converge, in that the two semantics for a particular system are identical. This actually reﬂects on the apt description of the properties of the relations deﬁning the approximation spaces. The only exception is the logic DAL of the ﬁrst category. As remarked in Section 5.4, given a general DAL-structure U ≡ (U, {Ri }i∈I ), it does not seem easy to construct an information system “naturally” to obtain U back as its standard structure. In case of the logics with attributes as expressions (cf. Section 4.4), one encounters a problem even earlier. The models here are based on structures of the form (U, A, {ind(P )}P ⊆A ), and there does not appear easily a corresponding “general” structure of the kind U ≡ (U, {Ri }i∈I ), with appropriate closure conditions on {Ri }i∈I . These logics have not been axiomatized, though the language can express a lot about attributes – that few of the other systems are able to do.

Propositional Logics from Rough Set Theory

23

An interesting picture is obtained from the logics of Section 2, leaving out DAL and other systems with models based on generalized spaces. Most of the logics are embeddable into each other (cf. Section 5). We have LD L3 P RL S5 Lr J. (1) In one sense then, the embeddings in (1) establish that no ‘new’ logic surfaces with the kind of rough set semantics deﬁned. But in another sense, well-known systems have been imparted a rough set interpretation. It should be noted that though the embeddings are deﬁned with respect to theoremhood, the relationships would hold in some cases (e.g. L3 − P RL and Lr − J) if derivability of wﬀs from non-empty premise sets is considered [31,10]. One could attempt to settle the question for the rest. (1) indicates another interesting future line of work, viz. an investigation for logics and interrelations, that may result on replacing S5 by other non-modal systems (as in [6]). All the systems presented other than LT (cf. Section 4.1), deal with static information. The semantics of LT essentially gives rise to a family of approximation spaces on the same domain, the indiscernibility relations changing with moments of time. One could further enquire about the behaviour of rough sets in such a dynamic information system. As remarked in Section 4.2, another open direction relates to a study of logics that may be obtained from the generalized approach in [38]. Overall, one may say that it has been a remarkable journey in the exploration of logics, beginning with a deceptively simple proposal of “rough sets”. We have seen the introduction of novel concepts – e.g. of “rough truth”, “rough modus ponens”, “rough consistency”, “rough mereology”. The journey has, by no means, ended. Pawlak’s theory has just opened up the horizon before us, to reveal a number of yet unexplored directions in the study of “rough logics”.

References 1. Banerjee, M., Chakraborty, M.K.: Algebras from rough sets. In Pal, S.K., Polkowski, L., Skowron, A., eds.: Rough-neuro Computing: Techniques for Computing with Words. Springer Verlag, Berlin (2004) 157–184 2. Pawlak, Z.: Rough Sets. Theoretical Aspects of Reasoning about Data. Kluwer Academic Publishers, Dordrecht (1991) 3. Pawlak, Z.: Rough logic. Bull. Polish Acad. Sc. (Tech. Sc.) 35 (1987) 253–258 4. Orlowska, E.: Kripke semantics for knowledge representation logics. Studia Logica XLIX (1990) 255–272 5. Chakraborty, M.K., Banerjee, M.: Rough consequence. Bull. Polish Acad. Sc.(Math.) 41(4) (1993) 299–304 6. Yao, Y., Lin, T.Y.: Generalization of rough sets using modal logics. Intelligent Automation and Soft Computing 2 (1996) 103–120 7. Banerjee, M., Chakraborty, M.K.: Rough sets through algebraic logic. Fundamenta Informaticae 28(3,4) (1996) 211–221 8. D¨ untsch, I.: A logic for rough sets. Theoretical Computer Science 179 (1997) 427–436

24

M. Banerjee and Md.A. Khan

9. Dai, J.H.: Logic for rough sets with rough double Stone algebraic semantics. In Slezak, D., Wang, G., Szczuka, M.S., D¨ untsch, I., Yao, Y., eds.: Proc. RSFSDMGrC(1), Canada, LNCS 3641, Springer Verlag (2005, 141-148) 10. Banerjee, M.: Logic for rough truth. Fundamenta Informaticae 71(2-3) (2006) 139–151 11. Farinas Del Cerro, L., Orlowska, E.: DAL – a logic for data analysis. Theoretical Computer Science 36 (1997) 251–264 12. Vakarelov, D.: Abstract characterization of some knowledge representation systems and the logic N IL of nondeterministic information. In Jorrand, P., Sgurev, V., eds.: Artiﬁcial Intelligence II. North–Holland (1987) 255–260 13. Vakarelov, D.: A modal logic for similarity relations in Pawlak knowledge representation systems. Fundamenta Informaticae 15 (1991) 61–79 14. Vakarelov, D.: Modal logics for knowledge representation systems. Theoretical Computer Science 90 (1991) 433–456 15. Nakamura, A.: A rough logic based on incomplete information and its application. Int. J. Approximate Reasoning 15 (1996) 367–378 16. Orlowska, E., Pawlak, Z.: Representation of nondeterministic information. Theoretical Computer Science 29 (1984) 27–39 17. Vakarelov, D., Balbiani, P.: A modal logic for indiscernibilty and complementarity in information systems. Fundamenta Informaticae 50 (2002) 243–263 18. Orlowska, E.: Logic of nondeterministic information. Studia Logica 1 (1985) 91–100 19. Orlowska, E.: Logic of indiscernibility relations. In Goos, G., Hartmanis, J., eds.: Proc. Symposium on Computation Theory, Zabr´ ow, 1984, LNCS 208, Springer Verlag (1985, 177–186) 20. Rauszer, C.M.: Rough logic for multiagent systems. In Masuch, M., Polos, L., eds.: Knowledge Representation and Reasoning under Uncertainty, LNAI 808. SpringerVerlag (1994) 161–181 21. Stepaniuk, J.: Rough relations and logics. In Polkowski, L., Skowron, A., eds.: Rough Sets in Knowledge Discovery 1: Methodology and Applications. PhysicaVerlag (1998) 248–260 22. Maluszy´ nski, Vit´ oria, A.: Toward rough datalog: embedding rough sets in prolog. In Pal, S.K., Polkowski, L., Skowron, A., eds.: Rough-neuro Computing: Techniques for Computing with Words. Springer Verlag, Berlin (2004) 297–332 23. Polkowski, L., Skowron, A.: Rough mereology: a new paradigm for approximate reasoning. Int. J. Approximate Reasoning 15(4) (1997) 333–365 24. Wong, S.K.M.: A rough set model for reasoning about knowledge. In Orlowska, E., ed.: Incomplete Information: Rough Set Analysis. Studies in Fuzziness and Soft Computing, vol. 13. Physica-Verlag (1998) 276–285 25. Banerjee, M.: Rough belief change. Transactions of Rough Sets V LNCS 4100 (2006) 25–38 26. Banerjee, M., Chakraborty, M.K.: Rough consequence and rough algebra. In Ziarko, W.P., ed.: Rough Sets, Fuzzy Sets and Knowledge Discovery, Proc. Int. Workshop on Rough Sets and Knowledge Discovery (RSKD’93). Workshops in Computing, London, Springer Verlag (1994, 196–207) 27. Yao, Y.: Constructive and algebraic methods of the theory of rough sets. Information Sciences 109 (1998) 21–47 28. Pomykala, J.: Approximation, similarity and rough construction. preprint CT–93– 07, ILLC Prepublication Series, University of Amsterdam (1993)

Propositional Logics from Rough Set Theory

25

29. Komorowski, J., Pawlak, Z., Polkowski, L., Skowron, A.: Rough sets: a tutorial. In Pal, S.K., Skowron, A., eds.: Rough Fuzzy Hybridization: A New Trend in Decision-Making. Springer Verlag, Singapore (1999) 3–98 30. Rasiowa, H.: An Algebraic Approach to Non-classical Logics. North Holland, Amsterdam (1974) 31. Banerjee, M.: Rough sets and 3-valued L ukasiewicz logic. Fundamenta Informaticae 32 (1997) 213–220 32. Pagliani, P.: Rough set theory and logic-algebraic structures. In Orlowska, E., ed.: Incomplete Information: Rough Set Analysis. Studies in Fuzziness and Soft Computing, vol. 13. Physica-Verlag (1998) 109–190 33. Iturrioz, L.: Rough sets and three-valued structures. In Orlowska, E., ed.: Logic at Work: Essays Dedicated to the Memory of Helena Rasiowa. Studies in Fuzziness and Soft Computing, vol. 24. Physica-Verlag (1999) 596–603 34. Boicescu, V., Filipoiu, A., Georgescu, G., Rudeano, S.: L ukasiewicz-Moisil Algebras. North Holland, Amsterdam (1991) 35. Comer, S.: Perfect extensions of regular double Stone algebras. Algebra Universalis 34 (1995) 96–109 36. Andr´eka, H., N´emeti, I., Sain, I.: Abstract model theoretic approach to algebraic logic. CCSOM working paper, Department of Statistics and Methodology, University of Amsterdam (1992) 37. Pawlak, Z.: Rough sets. Int. J. Comp. Inf. Sci. 11(5) (1982) 341–356 38. Pagliani, P., Chakraborty, M.K.: Information quanta and approximation spaces I: non-classical approximation operators. In: Proc. 2005 IEEE Conf. on Granular Computing, IEEE Press (2005, 605–610) 39. Pawlak, Z.: Rough relations. ICS PAS Reports 435 (1981) 40. Skowron, A., Stepaniuk, J.: Tolerance approximation spaces. Fundamenta Informaticae 27 (1996) 245–253 41. Polkowski, L.: Rough mereological reasoning in rough set theory: recent results and problems. In Wang, G., Peters, J.F., Skowron, A., Yao, Y., eds.: Proc. Rough Sets and Knowledge Technology (RSKT 2006), China, 2006, LNAI 4062, Springer Verlag (2006, 79–92) 42. da Costa, N.C.A., Doria, F.A.: On Ja´skowski’s discussive logics. Studia Logica 54 (1995) 33–60

Intuitionistic Rough Sets for Database Applications Theresa Beaubouef1 and Frederick E. Petry2 1

2

1

Southeastern Louisiana University Dept. of Computer Science & Ind. Technology Hammond, LA 70402, USA [email protected] Center for Intelligent and Knowledge-Based Systems Tulane University New Orleans, LA 70118, USA [email protected]

Introduction

We introduce the intuitionistic rough set and intuitionistic rough relational and object oriented database models. The intuitionistic rough set database models draw beneﬁts from both the rough set and intuitionistic techniques, providing greater management of uncertainty for databases applications in a less than certain world. We provide the foundation for the integration of intuitionistic rough sets into modeling of uncertainty in databases. This builds upon some of our previous research [2,3] with integrating fuzzy and rough set techniques for uncertainty management in databases.

2

Intuitionistic Rough Sets

An intuitionistic set [1] (intuitionistic fuzzy set) is a generalization of the traditional fuzzy set. Let set X be ﬁxed. An intuitionistic set A is deﬁned by the following: A = {x, μA (x), νA (x) : x ∈ X} where μA (x) → [0, 1], and νA (x) → [0, 1]. The degree of membership of element x ∈ X to the set A is denoted by μA (x), and the degree of nonmembership of element x ∈ X to the set A is denoted by νA (x). A is a subset of X. For all x ∈ X, 0 ≤ μA (x)+νA (x) ≤ 1. A hesitation margin, πA (x) = 1−(μA (x)+νA (x)), expresses a degree of uncertainty about whether x belongs to X or not, or uncertainty about the membership degree. This hesitancy may cater toward membership or nonmembership. We next deﬁne the intuitionistic rough set, which incorporates the beneﬁcial properties of both rough set [5] and intuitionistic set techniques. Intuitionistic rough sets are generalizations of fuzzy rough sets that give more information about the uncertain, or boundary region. They follow the deﬁnitions for partitioning of the universe into equivalence classes as in rough sets, but instead J.F. Peters et al. (Eds.): Transactions on Rough Sets VI, LNCS 4374, pp. 26–30, 2007. c Springer-Verlag Berlin Heidelberg 2007

Intuitionistic Rough Sets for Database Applications

27

of having a simple boundary region, there are basically two boundaries formed from the membership and nonmembership functions. Let U be a universe, Y a rough set in U , deﬁned on a partitioning of U into equivalence classes. Deﬁnition 1. An intuitionistic rough set Y in U is Y, μY (x), νY (x), where μY (x) is a membership function which associates a grade of membership from the interval [0,1] with every element (equivalence class) of U , and νY (x) associates a degree of non membership from the interval [0,1] with every element (equivalence class) of U , where 0 ≤ μY (x) + νY (x) ≤ 1, where x denotes the equivalence class containing x. A hesitation margin is πY (x) = 1 − (μY (x) + νY (x)). Consider the following special cases μ, ν for some element of Y : 1, 0 denotes total membership. This correspond to elements found in RY . 0, 1 denotes elements that do not belong to Y . Same as U − RY . 0.5, 0.5 corresponds to traditional rough set boundary region. p, 1 − p corresponds to fuzzy rough set in that there is a single boundary. In this case we assume that any degree of membership has a corresponding complementary degree of non membership. p, 0 corresponds to fuzzy rough set. 0, q This case can not be modeled by fuzzy rough sets. It denotes things that are not a member of RY or RY . It falls somewhere in the region U − RY . p, q Intuitionistic set general case , has membership and nonmembership. Let Y denote the complement of Y . Then the intuitionistic set having μY (x), μY (x) is the same as fuzzy rough set. The last two cases above, 0, q and p, q, cannot be represented by fuzzy sets, rough sets, or fuzzy rough sets. These are the situations which show that intuitionistic rough sets provide greater uncertainty management than the others alone. Note, however, that with the intuitionistic set we do not lose the information about uncertainty provided by other set theories, since from the ﬁrst few cases we see that they are special cases of the intuitionistic rough set. Although there are several various way of combining rough and fuzzy sets, we focus on those fuzzy rough sets as deﬁned in [2,3] and used for fuzzy rough databases, since our intuitionistic rough relational database model follows from this. The intuitionistic rough relational database model will have an advantage over the rough and fuzzy rough database models in that the non membership uncertainty of intuitionistic set theory will also play a role, providing even greater uncertainty management than the original models.

3

Intuitionistic Rough Relational Database Model

The intuitionistic rough relational database, as in the ordinary relational database, represents data as a collection of relations containing tuples. Because a relation is considered a set having the tuples as its members, the tuples are

28

T. Beaubouef and F.E. Petry

unordered. In addition, there can be no duplicate tuples in a relation. A tuple ti takes the form (di1 , di2 , . . . , dim , diμ , diν ), where dij is a domain value of a particular domain set Dj and diμ ∈ Dμ , where Dμ is the interval [0,1], the domain for intuitionistic membership values, and Dv is the interval [0,1], the domain for intuitionistic nonmembership values. In the ordinary relational database, dij ∈ Dj . In the intuitionistic rough relational database, except for the intuitionistic membership and nonmembership values, however, dij ∈ Dj , and although dij is not restricted to be a singleton, dij = ∅. Let P (Di ) denote any non-null member of the powerset of Di . Deﬁnition 2. A intuitionistic rough relation R is a subset of the set cross product P (D1 ) × P (D2 ) × . . . × P (Dm ) × Dμ × Dnu . An intuitionistic rough tuple t is any member of R. If ti is some arbitrary tuple, then ti = (di1 , di2 , . . . , dim , diμ , diν ) where dij ∈ Dj and diμ ∈ Dμ , diν ∈ Dν . Let [dxy ] denote the equivalence class to which dxy belongs. When dxy is a set of values, the equivalence class is formed by taking the union of equivalence classes of members of the set; if dxy = {c1 , c2 , ..., cn }, then [dxy ] = [c1] × [c2] × . . .× [cn ]. Deﬁnition 3. Tuples ti = (di1 , di2 , . . . , din , diμ , diν ) and tk = (dk1 , dk2 , . . . , dkn , dkμ , dkν ) are redundant if [dij ] = [dkj ] for all j = 1, . . . , n. In [3], we deﬁned several operators for the rough relational algebra, and in [2] demonstrated the expressive power of the fuzzy rough versions of these operators in the fuzzy rough relational database model. In an extension of this work we do the same for the rough intuitionistic database.

4

Intuitionistic Rough Object-Oriented Database (IROODB) Model

We next develop the intuitionistic rough object-oriented database model. We follow the formal framework and type deﬁnitions for generalized object-oriented databases proposed by [4] and extended for rough sets in [3]. We extend this framework, however, to allow for intuitionistic rough set indiscernibility and approximation regions for the representation of uncertainty as we have previously done for relational databases [2,3]. The intuitionistic rough object database scheme is formally deﬁned by the following type system and constraints. type ], where T can be a literal type Tliteral , The type system, T S = [T, P, fimpl which can be a base type, a collection literal type, or a structured literal type. It also contains Tobject , which speciﬁes object types, Tref erence , the set of speciﬁcations for reference types, and a void type. In the type system, each domain domts ∈ Dts , the set of domains. This domain set, along with a set of operators Ots and a set of axioms Ats , capture the semantics of the type speciﬁcation. The type system is then deﬁned based on these type speciﬁcations, the set of all programs P , and the implementation function mapping each type speciﬁcation

Intuitionistic Rough Sets for Database Applications

29

for a domain onto a subset of ρ(P ) – the powerset of P that contains all the implementations for the type system: type : T → ρ(P ) giving ts → {p1 , p2 , . . . pn }. fimpl

We are particularly interested in object types, and specify a class t of object types as Class id(id1 : s1 ; . . . ; idn : sn ) or Class id : id1 , . . . , idn (id1 : s1 ; . . . ; idn : sn ) where id, an identiﬁer, names an object type, {idi : 1 ≤ i ≤ m} is a ﬁnite set of identiﬁers denoting parent types of t, and {idi : si : 1 ≤ i ≤ n} is the ﬁnite set of characteristics speciﬁed for object type t within its syntax. This set includes all the attributes, relationships and method signatures for the object type. The identiﬁer for a characteristic is idi and the speciﬁcation is si for each of the idi : si . See [4] for details of how rough set concepts are integrated in this OO model, and how changing the granularity of the partitioning aﬀects query results. In that paper the OO model is extended for fuzzy and rough set uncertainty. If we extend the rough OODB further to allow for intuitionistic types, the type speciﬁcations T can be generalized to a set Tˇ as in [4], so that the deﬁnitions of the domains are generalized to intuitionistic sets. For every ts ∈ T , having domain ts being domts , the type system ts ∈ T is generalized to ts ∈ Tˇ , where domain of ts is denoted by domts and is deﬁned as the set ρ(domts ) of intuitionistic sets on domts , and Ots is generalized to Ots , which contains the generalized version of the operators. type The generalized type system then is a triple GT S = [Tˇ, P, f impl ], where Tˇ type

is the generalized type system, P is the set of all programs, and f impl maps each ts ∈ Tˇ onto that subset of P that contains the implementation for ts. An type instance of this GTS is a generalized type t = [ts, f impl (ts)], ts ∈ Tˇ . A generalized object belonging to this class is deﬁned by o = [oid, N, t, type f impl (ts), ν], where ν draws values from the generalized domain that allows an object to contain intuitionistic membership and nonmembership values as part of the state of the object. Both intuitionistic and rough set uncertainty management can be used in this generalized OODB model. We extended a formal framework of object-oriented databases to allow for modeling of various types of imprecision, vagueness, and uncertainty that typically occur in spatial data. The model is based on a formal type system and speciﬁed constraints, thus preserving integrity of the database, while at the same time allowing an OODB to be generalized in such a way as to include both intuitionistic and rough set uncertainty, both well-developed methods of uncertainty management. Incorporation of intuitionistic and rough set uncertainty into the OODB model is essential for representing imprecision and uncertainty in spatial data entities and in their interrelationships.

5

Conclusion

We introduced the intuitionistic rough set, then discussed how the intuitionistic rough set generalizes each of traditional rough, fuzzy, fuzzy-rough, and

30

T. Beaubouef and F.E. Petry

intuitionistic sets. The intuitionistic rough relational database model was introduced. This model allows for both rough and intuitionistic modeling of uncertainty. Because real world applications involve uncertainty, this model can more accurately represent data and relationships than traditional relational databases. We have also introduced our model for intuitionistic object-oriented databases and shown the signiﬁcance of both rough sets and intuitionistic sets for uncertainty management.

References 1. Atanassov, K.: Intuitionistic Fuzzy Sets. Fuzzy Sets and Systems 20 (1986) 87–96 2. Beaubouef T., Petry F.: Fuzzy Rough Set Techniques for Uncertainty Processing in a Relational Database. International Journal of Intelligent Systems 15(5) (2000) 389–424 3. Beaubouef, T., Petry, F., Buckles, B.: Extension of the Relational Database and its Algebra with Rough Set Techniques. Computational Intelligence 11(2) (1995) 233–245 4. De Tr´e, G., De Caluwe, R. A Generalized Object-Oriented Database Model with Generalized Constraints. In Proc. of NAFIPS’99, New York (1999) 381–386. 5. Pawlak, Z.:Rough Sets. International Journal of Man-Machine Studies 21 (1984) 127–134

An Experimental Comparison of Three Rough Set Approaches to Missing Attribute Values Jerzy W. Grzymala-Busse1,2 and Witold J. Grzymala-Busse3 1

Department of Electrical Engineering and Computer Science University of Kansas, Lawrence, KS 66045, USA 2 Institute of Computer Science, Polish Academy of Sciences, 01–237 Warsaw, Poland 3 Touchnet Information Systems, Inc., Lenexa, KS 66219, USA

Abstract. In this paper we present results of experiments conducted to compare three types of missing attribute values: lost values, ”do not care” conditions and attribute-concept values. For our experiments we selected six well known data sets. For every data set we created 30 new data sets replacing speciﬁed values by three diﬀerent types of missing attribute values, starting from 10%, ending with 100%, with increment of 10%. For all concepts of every data set concept lower and upper approximations were computed. Error rates were evaluated using ten-fold cross validation. Overall, interpreting missing attribute values as lost provides the best result for most incomplete data sets. Keywords: missing attribute values, incomplete data sets, concept approximations, LERS data mining system, MLEM2 algorithm.

1

Introduction

Real-life data are frequently incomplete, i.e., values for some attributes are missing. Appropriate handling of missing attribute values is one of the most important tasks of data mining. In this paper we assume that missing attribute values have three diﬀerent interpretations. The ﬁrst possibility is that missing attribute values are lost. Such values are interpreted as originally speciﬁed, but currently unavailable since these values were incidentally erased, forgotten to be recorded, etc. A rough set approach to incomplete data sets in which all attribute values were lost was presented for the ﬁrst time in [12], where two algorithms for rule induction, modiﬁed to handle lost attribute values, were introduced. The next possibility are ”do not care” conditions. Such missing attribute values were irrelevant during collection of data. Simply, an expert decided that the attribute value was irrelevant for a classiﬁcation or diagnosis of the case. For example, a data set describing ﬂu patients may contain, among other attributes, an attribute Color of hair. Though some scrupulous patients may ﬁll in this value, other patients may assume that this attribute is irrelevant for the ﬂu diagnosis J.F. Peters et al. (Eds.): Transactions on Rough Sets VI, LNCS 4374, pp. 31–50, 2007. c Springer-Verlag Berlin Heidelberg 2007

32

J.W. Grzymala-Busse and W.J. Grzymala-Busse

and leave it unspeciﬁed. If we suspect that this attribute does matter, the best interpretation for missing attribute values is replacing them by all possible existing attribute values. A rough set approach to incomplete data sets in which all attribute values were ”do not care” conditions was presented for the ﬁrst time in [4], where a method for rule induction was introduced in which each missing attribute value was replaced by all values from the domain of the attribute. The third possibility is a missing attribute value interpreted as an attributeconcept value. It is a similar case to a ”do not care” condition, however, it is restricted to a speciﬁc concept. A concept (class) is a set of all cases classiﬁed (or diagnosed) the same way. Using this interpretation, we will replace a missing attribute value by all values of the same attribute typical for the concept to which the case belongs. Let us consider a patient, sick with ﬂu, from the ﬂu data set, with a missing attribute value for Color of hair. Other patients, sick with ﬂu, ﬁlled in values brown and grey for this attribute. On the other hand, healthy patients characterized the color of their hair as blond and brown. Using attribute-concept value interpretation, this missing attribute value is replaced by brown and grey. If we would use ”do not care” condition interpretation, the same missing attribute value should be replaced by blond, brown, and grey. This approach was introduced in [10]. In general, incomplete decision tables are described by characteristic relations, in a similar way as complete decision tables are described by indiscernibility relations [7,8,9]. In rough set theory, one of the basic notions is the idea of lower and upper approximations. For complete decision tables, once the indiscernibility relation is ﬁxed and the concept (a set of cases) is given, the lower and upper approximations are unique. For incomplete decision tables, for a given characteristic relation and concept, there are three important and diﬀerent possibilities to deﬁne lower and upper approximations, called singleton, subset, and concept approximations [7]. Singleton lower and upper approximations were studied in [14,15,19,21,22]. Note that similar deﬁnitions of lower and upper approximations, though not for incomplete decision tables, were studied in [16,24,25]. Note that some other rough-set approaches to missing attribute values were presented in [4,11,13,23] as well.

2

Blocks of Attribute-Value Pairs—Complete Data

We assume that the input data sets are presented in the form of a decision table. An example of a decision table is shown in Table 1. Rows of the decision table represent cases, while columns are labeled by variables. The set of all cases will be denoted by U . In Table 1, U = {1, 2, ..., 6}. Independent variables are called attributes and a dependent variable is called a decision and is denoted by d. The set of all attributes will be denoted by A. In Table 1, A = {Temperature, Headache, Cough}. Any decision table deﬁnes a function ρ that maps the direct product of U and A into the set of all values. For example, in

An Experimental Comparison of Three Rough Set Approaches

33

Table 1. A complete decision table Attributes

Decision

Case

Temperature

Headache

Cough

Flu

1 2 3 4 5 6

high very high high high normal normal

yes yes no yes yes no

yes no no yes no yes

yes yes no yes no no

Table 1, ρ(1, T emperature) = high. A decision table with completely speciﬁed function ρ will be called completely speciﬁed, or, for the sake of simplicity, complete. In practice, input data for data mining are frequently aﬀected by missing attribute values. In other words, the corresponding function ρ is incompletely speciﬁed (partial). A decision table with an incompletely speciﬁed function ρ will be called incomplete. Function ρ describing Table 1 is completely speciﬁed. An important tool to analyze complete decision tables is a block of the attribute-value pair. Let a be an attribute, i.e., a ∈ A and let v be a value of a for some case. For complete decision tables if t = (a, v) is an attribute-value pair then a block of t, denoted [t], is a set of all cases from U that for attribute a have value v. Rough set theory [17], [18] is based on the idea of an indiscernibility relation, deﬁned for complete decision tables. Let B be a nonempty subset of the set A of all attributes. The indiscernibility relation IN D(B) is a relation on U deﬁned for x, y ∈ U as follows (x , y) ∈ IND(B ) if and only if ρ(x , a) = ρ(y, a) for all a ∈ B . The indiscernibility relation IN D(B) is an equivalence relation. Equivalence classes of IN D(B) are called elementary sets of B and are denoted by [x]B . For example, for Table 1, elementary sets of IN D(A) are {1, 4}, {2}, {3}, {5}, {6}. Additionally, IND(B) = {(1, 1), (1, 4), (2, 2), (3, 3), (4, 1), (4, 4), (5, 5), (6, 6)}. The indiscernibility relation IN D(B) may be computed using the idea of blocks of attribute-value pairs. Let a be an attribute, i.e., a ∈ A and let v be a value of a for some case. For complete decision tables if t = (a, v) is an attribute-value pair then a block of t, denoted [t], is a set of all cases from U that for attribute a have value v. For Table 1, [(Temperature, high)] = {1, 3, 4}, [(Temperature, very high)] = {2}, [(Temperature, normal)] = {5, 6}, [(Headache, yes)] = {1, 2, 4, 5}, [(Headache, no)] = {3, 6},

34

J.W. Grzymala-Busse and W.J. Grzymala-Busse

[(Cough, yes)] = {1, 4, 6}, [(Cough, no)] = {2, 3, 5}. The indiscernibility relation IN D(B) is known when all elementary blocks of IND(B) are known. Such elementary blocks of B are intersections of the corresponding attribute-value pairs, i.e., for any case x ∈ U , [x]B = ∩{[(a, v)] | a ∈ B, ρ(x, a) = v}. We will illustrate the idea how to compute elementary sets of B for Table 1 and B = A. [1]A [2]A [3]A [5]A [6]A

= [4]A = {1, 3, 4} ∩ {1, 2, 4, 5} ∩ {1, 4, 6} = {1, 4}, = {2} ∩ {1, 2, 4, 5} ∩ {2, 3, 5} = {2}, = {1, 3, 4} ∩ {3, 6} ∩ {2, 3, 5} = {3}, = {5, 6} ∩ {1, 2, 4, 5} ∩ {2, 3, 5} = {5}, = {5, 6} ∩ {3, 6} ∩ {1, 4, 6} = {6},

For completely speciﬁed decision tables lower and upper approximations are deﬁned using the indiscernibility relation. Any ﬁnite union of elementary sets, associated with B, will be called a B-deﬁnable set. Let X be any subset of the set U of all cases. The set X is called a concept and is usually deﬁned as the set of all cases deﬁned by a speciﬁc value of the decision. In general, X is not a B-deﬁnable set. However, set X may be approximated by two B-deﬁnable sets, the ﬁrst one is called a B-lower approximation of X, denoted by BX and deﬁned as follows ∪{[x]B | x ∈ U, [x]B ⊆ X}, The second set is called a B-upper approximation of X, denoted by BX and deﬁned as follows ∪{[x]B | x ∈ U, [x]B ∩ X = ∅). Data set presented in Table 1 is consistent (the lower approximation is equal to the upper approximation for every concept), hence the certain rule set and the possible rule set are identical. Rules in the LERS format (every rule is equipped with three numbers, the total number of attribute-value pairs on the left-hand side of the rule, the total number of examples correctly classiﬁed by the rule during training, and the total number of training cases matching the left-hand side of the rule) [6] are: 2, 2, 2 (Temperature, high) & (Headache, yes) -> (Flu, yes) 1, 1, 1 (Temperature, very high) -> (Flu, yes) 1, 2, 2 (Temperature, normal) -> (Flu, no) 1, 2, 2 (Headache, no) -> (Flu, no)

An Experimental Comparison of Three Rough Set Approaches

35

Note that the above rules were induced by the MLEM2 (Modiﬁed Learning from Examples Module, version 2) option of the LERS (Learning from Examples based on Rough Sets) data mining system [2,5,6].

3

Blocks of Attribute-Value Pairs—Incomplete Data

For the rest of the paper we will assume that all decision values are speciﬁed, i.e., they are not missing. Additionally, we will assume that lost values will be denoted by ”?”, ”do not care” conditions by ”*”, and attribute-concept values by ”−”. Additionally, we will assume that for each case at least one attribute value is speciﬁed. Table 2 is Table 1 with eight attribute values missing. All of these missing attribute values are lost. Table 2. An incomplete decision table (all missing attribute values are lost values) Attributes

Decision

Case

Temperature

Headache

Cough

Flu

1 2 3 4 5 6

high ? ? high ? normal

? yes no ? yes no

yes ? ? yes no ?

yes yes no yes no no

For incomplete decision tables, a block of an attribute-value pair must be modiﬁed in the following way: – If for an attribute a there exists a case x such that ρ(x, a) = ?, i.e., the corresponding value is lost, then the case x should not be included in any blocks [(a, v)] for all values v of attribute a, – If for an attribute a there exists a case x such that the corresponding value is a ”do not care” condition, i.e., ρ(x, a) = ∗, then the case x should be included in blocks [(a, v)] for all speciﬁed values v of attribute a. – If for an attribute a there exists a case x such that the corresponding value is an attribute-concept value, i.e., ρ(x, a) = −, then the corresponding case x should be included in blocks [(a, v)] for all speciﬁed values v ∈ V (x, a) of attribute a, where V (x , a) = {ρ(y, a) | ρ(y, a) is speciﬁed , y ∈ U, ρ(y, d) = ρ(x, d)}.

36

J.W. Grzymala-Busse and W.J. Grzymala-Busse

Thus, for Table 2, [(Temperature, high)] = {1, 4}, [(Temperature, normal)] = {6}, [(Headache, yes)] = {2, 5}, [(Headache, no)] = {3, 6}, [(Cough, yes)] = {1, 4}, [(Cough, no)] = {5}. For incomplete data sets the idea of the elementary block is extended to a characteristic set. For a case x ∈ U the characteristic set KB (x) is deﬁned as the intersection of the sets K(x, a), for all a ∈ B, where the set K(x, a) is deﬁned in the following way: – If ρ(x, a) is speciﬁed, then K(x, a) is the block [(a, ρ(x, a)] of attribute a and its value ρ(x, a), – If ρ(x, a) = ? or ρ(x, a) = ∗ then the set K(x, a) = U , – If ρ(x, a) = −, then the corresponding set K(x, a) is equal to the union of all blocks of attribute-value pairs (a, v), where v ∈ V (x, a) if V (x, a) is nonempty. If V (x, a) is empty, K(x, a) = U . Thus, for Table 2 KA (1) = {1, 4} ∩ U ∩ {1, 4} = {1, 4}, KA (2) = U ∩ {2, 5} ∩ U = {2, 5}, KA (3) = U ∩ {3, 6} ∩ U = {3, 6}, KA (4) = {1, 4} ∩ U ∩ {1, 4} = {1, 4}, KA (5) = U ∩ {2, 5} ∩ {5} = {5}, KA (6) = {6} ∩ {3, 6} ∩ U = {6}, Characteristic set KB (x) may be interpreted as the set of cases that are indistinguishable from x using all attributes from B and using a given interpretation of missing attribute values. Thus, KA (x) is the set of all cases that cannot be distinguished from x using all attributes. In [24] KA (x) was called a successor neighborhood of x, see also [16,19,24,25]. Obviously, when a data set is complete, for given B ⊆ A, all characteristic sets KB (x) are identical with elementary blocks [x]B . The characteristic relation R(B) is a relation on U deﬁned for x, y ∈ U as follows (x , y) ∈ R(B ) if and only if y ∈ KB (x ). The characteristic relation R(B) is reﬂexive but—in general—does not need to be symmetric or transitive. Also, the characteristic relation R(B) is known if we know characteristic sets KB (x) for all x ∈ U . In our example, R(A) = {(1, 1), (1, 4), (2, 2), (2, 5), (3, 3), (3, 6), (4, 1), (4, 5), (5, 5), (6, 6)}. The most convenient way to deﬁne the characteristic relation is through the characteristic sets. For decision tables, in which all missing attribute values are lost, a special characteristic relation was deﬁned in [21], see also, e.g., [20,22]. For incompletely speciﬁed decision tables lower and upper approximations may be deﬁned in a few diﬀerent ways. First, the deﬁnition of deﬁnability should

An Experimental Comparison of Three Rough Set Approaches

37

Fig. 1. Bankruptcy data—certain rule sets

be modiﬁed. A union of some intersections of attribute-value pair blocks, in any such intersection all attributes should be diﬀerent and attributes are members of B, will be called B-locally deﬁnable sets. A union of characteristic sets KB (x), where x ∈ X ⊆ U will be called a B-globally deﬁnable set. Any set X that is B -globally deﬁnable is B -locally deﬁnable, the converse is not true. In this paper we quote three diﬀerent deﬁnitions of lower and upper approximations [7,8,9]. Let X be a concept, let B be a subset of the set A of all attributes, and let R(B) be the characteristic relation of the incomplete decision table with characteristic sets K(x), where x ∈ U . Our ﬁrst deﬁnition uses a similar idea as in the previous articles on incompletely speciﬁed decision tables [14,15,20,21,22], i.e., lower and upper approximations are sets of singletons from the universe U satisfying some properties. Thus, lower and upper approximations are deﬁned by constructing both sets from singletons. We will call these approximations singleton. Namely, a singleton B-lower approximation of X is deﬁned as follows: BX = {x ∈ U | KB (x) ⊆ X}. A singleton B-upper approximation of X is BX = {x ∈ U | KB (x) ∩ X = ∅}. In our example of the decision table presented in Table 2 let us say that B = A. Then the singleton A-lower and A-upper approximations of the two concepts: {1, 2, 4} and {3, 5, 6} are:

38

J.W. Grzymala-Busse and W.J. Grzymala-Busse

Fig. 2. Bankruptcy data—possible rule sets

A{1, 2, 4} = {1, 4}, A{3, 5, 6} = {3, 5, 6}, A{1, 2, 4} = {1, 2, 4}, A{3, 5, 6} = {2, 3, 5, 6}. Note that the set A{1, 2, 4} is not even A-locally deﬁnable. Hence, as it was previously argued in [7,8,9], singleton approximations should not be used for rule induction. Obviously, if a set is not B-locally deﬁnable then it cannot be expressed by rule sets using attributes from B. We may deﬁne lower and upper approximations for incomplete decision tables by using characteristic sets instead of elementary sets. There are two ways to do this. Using the ﬁrst way, a subset B-lower approximation of X is deﬁned as follows: BX = ∪{KB (x) | x ∈ U, KB (x) ⊆ X}. A subset B-upper approximation of X is BX = ∪{KB (x) | x ∈ U, KB (x) ∩ X = ∅}.

An Experimental Comparison of Three Rough Set Approaches

39

Fig. 3. Breast cancer (Slovenia) data—certain rule sets

Since any characteristic relation R(B) is reﬂexive, for any concept X, singleton B-lower and B-upper approximations of X are subsets of the subset B-lower and B-upper approximations of X, respectively [9]. For the same decision table, presented in Table 2, the subset A-lower and A-upper approximations are A{1, 2, 4} = {1, 4}, A{3, 5, 6} = {3, 5, 6}, A{1, 2, 4} = {1, 2, 4, 5}, A{3, 5, 6} = {2, 3, 5, 6}. The second possibility is to modify the subset deﬁnition of lower and upper approximation by replacing the universe U from the subset deﬁnition by a concept X. A concept B-lower approximation of the concept X is deﬁned as follows: BX = ∪{KB (x) | x ∈ X, KB (x) ⊆ X}. Obviously, the subset B-lower approximation of X is the same set as the concept B-lower approximation of X [7]. A concept B-upper approximation of the concept X is deﬁned as follows:

40

J.W. Grzymala-Busse and W.J. Grzymala-Busse

Fig. 4. Breast cancer (Slovenia) data—possible rule sets

BX = ∪{KB (x) | x ∈ X, KB (x) ∩ X = ∅} = ∪{KB (x) | x ∈ X}. The concept B-upper approximation of X is a subset of the subset B-upper approximation of X [7]. For the decision table presented in Table 2, the concept A-lower and A-upper approximations are A{1, 2, 4} = {1, 4}, A{3, 5, 6} = {3, 5, 6}, A{1, 2, 4} = {1, 2, 4, 5}, A{3, 5, 6} = {2, 3, 5, 6}. Note that for complete decision tables, all three deﬁnitions of lower approximations, singleton, subset and concept, coalesce to the same deﬁnition. Also, for complete decision tables, all three deﬁnitions of upper approximations coalesce to the same deﬁnition. This is not true for incomplete decision tables, as our example shows. For Table 2, certain rules [3], induced from the concept lower approximations are

An Experimental Comparison of Three Rough Set Approaches

41

Fig. 5. Breast cancer (Wisconsin) data—certain rule sets

1, 2, 2 (Temperature, high) -> (Flu, yes) 1, 2, 2 (Headache, no) -> (Flu, no) 1, 1, 1 (Cough, no) -> (Flu, no) and possible rules [3], induced from the concept upper approximations, are 1, 2, 2 (Temperature, high) -> (Flu, yes) 1, 2, 2 (Headache, yes) -> (Flu, yes) 1, 2, 2 (Headache, no) -> (Flu, no) 1, 1, 1 (Cough, no) -> (Flu, no) Table 3 shows a modiﬁcation of Table 2, where all lost values are replaced by ”do not care” conditions. For decision tables where all missing attribute values are ”do not care” conditions a special characteristic relation was deﬁned in [14], see also, e.g., [15]. Blocks of attribute-value pairs are [(Temperature, high)] = {1, 2, 3, 4, 5}, [(Temperature, normal)] = {2, 3, 5, 6},

42

J.W. Grzymala-Busse and W.J. Grzymala-Busse

Table 3. An incomplete decision table (all missing attribute values are lost values) Attributes

Decision

Case

Temperature

Headache

Cough

Flu

1 2 3 4 5 6

high * * high * normal

* yes no * yes no

yes * * yes no *

yes yes no yes no no

Fig. 6. Breast cancer (Wisconsin) data—possible rule sets

[(Headache, yes)] = {1, 2, 4, 5}, [(Headache, no)] = {1, 3, 4, 6}, [(Cough, yes)] = {1, 2, 3, 4, 6}, [(Cough, no)] = {2, 3, 5, 6}. Characteristic sets are KA (1) = {1, 2, 3, 4, 5} ∩ U ∩ {1, 2, 3, 4, 6} = {1, 2, 3, 4}, KA (2) = U ∩ {1, 2, 4, 5} ∩ U = {1, 2, 4, 5}, KA (3) = U ∩ {1, 3, 4, 6} ∩ U = {1, 3, 4, 6},

An Experimental Comparison of Three Rough Set Approaches

43

KA (4) = {1, 2, 3, 4, 5} ∩ U ∩ {1, 2, 3, 4, 6} = {1, 2, 3, 4}, KA (5) = U ∩ {1, 2, 4, 5} ∩ {2, 3, 5, 6} = {2, 5}, KA (6) = {2, 3, 5, 6} ∩ {1, 3, 4, 6} ∩ U = {3, 6}, For the decision table presented in Table 3, the concept A-lower and A-upper approximations are A{1, 2, 4} = ∅, A{3, 5, 6} = {3, 6}, A{1, 2, 4} = {1, 2, 3, 4, 5}, A{3, 5, 6} = U. In our example, the concept A-lower approximation of {1, 2, 4} is the empty set. With large percentage of missing attribute values interpreted as ”do not care” conditions, empty lower approximations cause large increases of error rates during ten-fold cross validation. For Table 3, the only certain rule, induced from the concept lower approximation, is

Fig. 7. Image segmentation data—certain rule sets

44

J.W. Grzymala-Busse and W.J. Grzymala-Busse

Fig. 8. Image segmentation data—possible rule sets

2, 2, 2 (Temperature, normal) & (Headache, no) -> (Flu, no) and possible rules, induced from the concept upper approximations, are 1, 3, 5 (Temperature, high) -> (Flu, yes) 1, 2, 5 (Temperature, high) -> (Flu, no) 1, 3, 4 (Temperature, normal) -> (Flu, no) Table 4 is another modiﬁcation of Table 2, where all lost values are replaced by attribute-concept values. Blocks of attribute-value pairs are [(Temperature, high)] = {1, 2, 4}, [(Temperature, normal)] = {3, 5, 6}, [(Headache, yes)] = {1, 2, 4, 5}, [(Headache, no)] = {3, 6}, [(Cough, yes)] = {1, 2, 4}, [(Cough, no)] = {3, 5, 6}. Characteristic sets are KA (1) = {1, 2, 4} ∩ {1, 2, 4, 5} ∩ {1, 2, 4} = {1, 2, 4}, KA (2) = {1, 2, 4} ∩ {1, 2, 4, 5} ∩ {1, 2, 4} = {1, 2, 4}, KA (3) = {3, 5, 6} ∩ {3, 6} ∩ {3, 5, 6} = {3, 6},

An Experimental Comparison of Three Rough Set Approaches

45

Table 4. An incomplete decision table (all missing attribute values are lost values) Attributes

Decision

Case

Temperature

Headache

Cough

Flu

1 2 3 4 5 6

high – – high – normal

– yes no – yes no

yes – – yes no –

yes yes no yes no no

Fig. 9. Iris data—certain rule sets

KA (4) = {1, 2, 4} ∩ {1, 2, 4, 5} ∩ {1, 2, 4} = {1, 2, 4}, KA (5) = {3, 5, 6} ∩ {1, 2, 4, 5} ∩ {3, 5, 6} = {5}, KA (6) = {3, 5, 6} ∩ {3, 6} ∩ {3, 5, 6} = {3, 6}, For the decision table presented in Table 4, the concept A-lower and A-upper approximations are A{1, 2, 4} = {1, 2, 4}, A{3, 5, 6} = {3, 5, 6},

46

J.W. Grzymala-Busse and W.J. Grzymala-Busse

Fig. 10. Iris data—possible rule sets

A{1, 2, 4} = {1, 2, 4}, A{3, 5, 6} = {3, 5, 6}. For Table 4, certain rules, induced from the concept lower approximations, are identical with possible rules, induced from concept upper approximations 1, 3, 3 (Temperature, high) -> (Flu, yes) 1, 3, 3 (Temperature, normal) -> (Flu, no)

4

Experiments

For our experiments six typical data sets were used, see Table 5. These data sets were complete (all attribute values were completely speciﬁed), with the exception of breast cancer (Slovenia) data set, which originally contained 11 cases (out of 286) with missing attribute values. These 11 cases were removed. In two data sets: bankruptcy and iris all attributes were numerical. These data sets were processed as numerical (i.e., discretization was done during rule induction by MLEM2). The image segmentation data set was converted into symbolic using a discretization method based on agglomerative cluster analysis (this method was described, e.g., in [1]).

An Experimental Comparison of Three Rough Set Approaches

47

Table 5. Data sets used for experiments Data set cases Bankruptcy Breast cancer (Slovenia) Breast cancer (Wisconsin) Image segmentation Iris Lymphography

66 277 625 210 150 148

Number of attributes

concepts

5 9 9 19 4 18

2 2 9 7 3 4

Fig. 11. Lymphography data—certain rule sets

To each data set we conducted a series of three experiments, adding incrementally (with 10% increment) missing attribute values of three diﬀerent types. Thus, we started each series of experiments with no missing attribute values, then we added 10% of missing attribute values of given type, then we added additional 10% of missing attribute values of the same type, etc., until reaching a level of 100% missing attribute values. For each data set and a speciﬁc type of missing attribute values ten additional data sets were created. Furthermore, for each data set with some percentage of missing attribute values, experiments were conducted separately for certain and possible rule sets, using concept lower and upper approximations, respectively. Ten-fold cross

48

J.W. Grzymala-Busse and W.J. Grzymala-Busse

Fig. 12. Lymphography data—possible rule sets

validation was used to compute an error rate. Rule sets were induced by the MLEM2 option of the LERS data mining system [2,5,6]. Results of our experiments are presented in Figures 1–12. In all 12 ﬁgures, lost values, ”do not care” conditions, and attribute-concept values denote percentage of error rate for experiments with missing attribute values interpreted as lost values, ”do not care” conditions, and attribute-concept values, respectively.

5

Conclusions

During all series of experiments the error rate was aﬀected by large variance. Moreover, for some data sets (e.g., breast cancer (Wisconsin)), adding a small amount of missing attribute values resulted in a decreased error rate. Most likely, in these data sets, attributes aﬀected by missing attribute values were not important. In eﬀect, the induced rule sets were more general and better. It is clear that inducing certain rule sets while using a ”do not care” condition approach to missing attribute values was the worst approach. This was caused by the fact that lower approximations of concepts, with large number of missing attribute values, were empty. Another surprising conclusion is that for some data sets (breast cancer (Slovenia) and breast cancer (Wisconsin)) adding a large number of missing attribute values does not aﬀect the error rate seriously—the error rate was almost the same for larger and larger number of missing attribute values.

An Experimental Comparison of Three Rough Set Approaches

49

Overall, it seems that the interpretation of missing attribute values as lost is the best approach among our three types of missing attribute value interpretations. Taking into account a large variance, the diﬀerence between error rates for certain and possible rule sets is negligible.

References 1. Chmielewski, M.R. and Grzymala-Busse, J.W.: Global discretization of continuous attributes as preprocessing for machine learning. Int. Journal of Approximate Reasoning 15 (1996) 319–331. 2. Chan, C.C. and Grzymala-Busse, J.W.: On the attribute redundancy and the learning programs ID3, PRISM, and LEM2. Department of Computer Science, University of Kansas, TR-91-14, December 1991, 20 pp. 3. Grzymala-Busse, J.W.: Knowledge acquisition under uncertainty—A rough set approach. Journal of Intelligent & Robotic Systems 1 (1988), 3–16. 4. Grzymala-Busse, J.W.: On the unknown attribute values in learning from examples. Proc. of the ISMIS-91, 6th International Symposium on Methodologies for Intelligent Systems, Charlotte, North Carolina, October 16–19, 1991. Lecture Notes in Artiﬁcial Intelligence, vol. 542, Springer-Verlag, Berlin, Heidelberg, New York (1991) 368–377. 5. Grzymala-Busse, J.W.: LERS—A system for learning from examples based on rough sets. In Intelligent Decision Support. Handbook of Applications and Advances of the Rough Sets Theory, ed. by R. Slowinski, Kluwer Academic Publishers, Dordrecht, Boston, London (1992) 3–18. 6. Grzymala-Busse., J.W.: MLEM2: A new algorithm for rule induction from imperfect data. Proceedings of the 9th International Conference on Information Processing and Management of Uncertainty in Knowledge-Based Systems, IPMU 2002, July 1–5, Annecy, France, 243–250. 7. Grzymala-Busse, J.W.: Rough set strategies to data with missing attribute values. Workshop Notes, Foundations and New Directions of Data Mining, the 3-rd International Conference on Data Mining, Melbourne, FL, USA, November 19–22, 2003, 56–63. 8. Grzymala-Busse, J.W.: Data with missing attribute values: Generalization of idiscernibility relation and rule induction. Transactions on Rough Sets, Lecture Notes in Computer Science Journal Subline, Springer-Verlag, vol. 1 (2004) 78–95. 9. Grzymala-Busse, J.W.: Characteristic relations for incomplete data: A generalization of the indiscernibility relation. Proc. of the RSCTC’2004, the Fourth International Conference on Rough Sets and Current Trends in Computing, Uppsala, Sweden, June 1–5, 2004. Lecture Notes in Artiﬁcial Intelligence 3066, SpringerVerlag 2004, 244–253. 10. Grzymala-Busse, J.W.: Three approaches to missing attribute valuesA rough set perspective. Proceedings of the Workshop on Foundation of Data Mining, associated with the Fourth IEEE International Conference on Data Mining, Brighton, UK, November 1–4, 2004, 55–62. 11. Grzymala-Busse, J.W. and Hu, M.: A comparison of several approaches to missing attribute values in data mining. Proceedings of the Second International Conference on Rough Sets and Current Trends in Computing RSCTC’2000, Banﬀ, Canada, October 16–19, 2000, 340–347.

50

J.W. Grzymala-Busse and W.J. Grzymala-Busse

12. Grzymala-Busse, J.W. and Wang A.Y.: Modiﬁed algorithms LEM1 and LEM2 for rule induction from data with missing attribute values. Proc. of the Fifth International Workshop on Rough Sets and Soft Computing (RSSC’97) at the Third Joint Conference on Information Sciences (JCIS’97), Research Triangle Park, NC, March 2–5, 1997, 69–72. 13. Hong, T.P., Tseng L.H. and Chien, B.C.: Learning coverage rules from incomplete data based on rough sets. Proc. of the IEEE International Conference on Systems, Man and Cybernetics, Hague, the Netherlands, October 10–13, 2004, 3226–3231. 14. Kryszkiewicz, M.: Rough set approach to incomplete information systems. Proc. of the Second Annual Joint Conference on Information Sciences, Wrightsville Beach, NC, September 28–October 1, 1995, 194–197. 15. Kryszkiewicz, M.: Rules in incomplete information systems. Information Sciences 113 (1999) 271–292. and knowledge base systems. Fourth International Symposium on Methodologies of Intelligent Systems (Poster Sessions), Charlotte, North Carolina, October 12–14, 1989, 75–86. Tucson, Arizona, December 4–8, 1989, 286–293. 16. Lin, T.Y.: Topological and fuzzy rough sets. In Intelligent Decision Support. Handbook of Applications and Advances of the Rough Sets Theory, ed. by R. Slowinski, Kluwer Academic Publishers, Dordrecht, Boston, London (1992) 287–304. 17. Pawlak, Z.: Rough Sets. International Journal of Computer and Information Sciences 11 (1982) 341–356. 18. Pawlak, Z.: Rough Sets. Theoretical Aspects of Reasoning about Data. Kluwer Academic Publishers, Dordrecht, Boston, London (1991). 19. Slowinski, R. and Vanderpooten, D.: A generalized deﬁnition of rough approximations based on similarity. IEEE Transactions on Knowledge and Data Engineering 12 (2000) 331–336. 20. Stefanowski, J.: Algorithms of Decision Rule Induction in Data Mining. Poznan University of Technology Press, Poznan, Poland (2001). 21. Stefanowski, J. and Tsoukias, A.: On the extension of rough sets under incomplete information. Proc. of the 7th International Workshop on New Directions in Rough Sets, Data Mining, and Granular-Soft Computing, RSFDGrC’1999, Ube, Yamaguchi, Japan, November 8–10, 1999, 73–81. 22. Stefanowski, J. and Tsoukias, A.: Incomplete information tables and rough classiﬁcation. Computational Intelligence 17 (2001) 545–566. 23. Wang, G.: Extension of rough set under incomplete information systems. Proc. of the IEEE International Conference on Fuzzy Systems (FUZZ IEEE’2002), vol. 2, Honolulu, HI, May 12–17, 2002, 1098–1103. 24. Yao, Y.Y.: Relational interpretations of neighborhood operators and rough set approximation operators. Information Sciences 111 (1998) 239–259. 25. Yao, Y.Y. and Lin, T.Y.: Generalization of rough sets using modal logics. Intelligent Automation and Soft Computing 2 (1996) 103–119.

Pawlak’s Landscaping with Rough Sets Mihir K. Chakraborty Department of Pure Mathematics, University of Calcutta 35 Ballygunge Circular Road, Kolkata 700019, India [email protected]

Abstract. This paper reviews, rather non-technically, Pawlak’s approach to vagueness through rough sets and looks for a foundation of rough sets in an early work of Obtulowicz. An extension of Obtulowicz’s proposal is suggested that in turn, hints at a uniﬁed approach to rough sets and fuzzy sets.

1

Introduction

The concluding decades of the past century have added several outstanding, signiﬁcant and elegant contributions to human knowledge of which Rough Set Theory is one. Zdzislaw Pawlak, a Professor of Computer Science from Poland, ﬁrst proposed this theory in 1982 through his publication entitled ‘Rough Sets’ [20]. Surprisingly, this is again a contribution to humanity from one belonging to the ﬁeld of computer science – during the same period, the same community gifted several other elegant creations, like Fuzzy Set Theory by Lotﬁ Zadeh in 1965. It is also interesting to note that both the theories address basically the same issue, viz. ‘vagueness’ and this fact is not merely a coincidence. ‘Vagueness’ had been an outstanding issue. Great minds of the antiquity, both of the East and the West delved into the notion exhibited in various forms. (Theseuses’ ship [43,17], the Sorites [43,13], or the tetra-lemma (Catuskoti) [35,32], for example). Following Enlightenment, with the rise of modern rationality, embodied in the methods of physical sciences, more speciﬁcally physics, ‘vagueness’ had been gradually pushed aside to the fringes like the indigenous population in Australia or in America and other places. Use of imprecise terms were not only marginalized, but virtually banished from all serious discourses as expressed by the rationalist, humanist Bertrand Russell in the lines (in, Our Knowledge of External World as a Field of Scientiﬁc Method in Philosophy) (cf. Hao Wang [42]): “The study of logic becomes the central study in philosophy: it gives the method of research in philosophy, just as mathematics gives the method in physics;”

I would like to express my sincere thanks to Smita Sirker, Dept. of Philosophy, Jadavpur University, for kindly reading the ﬁrst draft and making valuable comments, particularly on the main philosophy of this article viz. the relationship between the existence of an object in a concept and indiscernibility.

J.F. Peters et al. (Eds.): Transactions on Rough Sets VI, LNCS 4374, pp. 51–63, 2007. c Springer-Verlag Berlin Heidelberg 2007

52

M.K. Chakraborty

and further from his article ‘Vagueness’ (Australian Journal of Philosophy)[37]: “Logical words, like the rest, when used by human beings, share the vagueness of all other words. There is however, less vagueness about logical words than about the words of daily life.” However, in spite of the great tide of modern rationality all over the world, a sense of understanding that there exists an essential role of vagueness in human knowledge system as well as life, was not totally wiped out from Eastern thoughts. Ironically, the most advanced technology – computer science, the most recent gift of modernity, has ushered the study of vagueness spectacularly from the disrespectful margin straightaway to centre-stage. That ‘vagueness’ in general is diﬀerent from ‘probability’ has somewhat been accepted nowadays after the long, ﬁerce debates that took place during the years immediately following the advent of fuzzy set theory in 1965. So Pawlak did not have to ﬁght that battle. Yet he had to utter this warning which is an excellent distinctive criterion viz. “Vagueness is the property of sets... whereas uncertainty is the property of an element”[28]. Uncertainty leads to probabilistic studies. It is often said of course, that vagueness is uncertainty too but not of probabilistic kind. However, right from the beginning Pawlak wanted to point at the distinction between Rough Set theory and Fuzzy Set theory. In the introduction to his short communication [23] he declares “we compare this concept with that of the fuzzy set and we show that these two concepts are diﬀerent.” Diﬀerent in what sense? Early Pawlak (during the 80s) was ﬁrm in his belief that Rough Set is properly addressing vagueness since it talks about ‘boundaries’ of a set and the property ‘rough’ is ascribed to a set. On the other hand, although the qualiﬁer ‘fuzzy’ has been ascribed to sets too, in reality the theory deals with degree of membership of an object in a ‘set’ and hence is dealing with some kind of uncertainty of belongingness of objects. So according to the above quoted norm, fuzzy set theory is not addressing vagueness proper. However, in later Pawlak, perhaps a change in opinion is observed as reﬂected in the following categorical remark “Both fuzzy and rough set theory represent two diﬀerent approaches to vagueness. Fuzzy set theory addresses gradualness of knowledge, expressed by the fuzzy membership - whereas rough set theory addresses granularity of knowledge expressed by indiscernibility relation” [22]. We shall discuss the role of indiscernibility to some length in the foundations of fuzzy set theory as well as rough set theory and thus in vagueness. But it needs to be mentioned that the relationship between the two theories had quite naturally been a favourite topic of study in those turbulent decades. For example, Wygralak in the 1985 BUSEFAL Conference presented a paper [44] in which he established that basic operations on rough sets (i.e. union and intersection) can be expressed as some special operations as their membership functions. Pawlak already talked about the distinction and the irreducibility of rough sets to fuzzy sets. This point needs a little clariﬁcation since rough sets approximations are not distributive with respect to all set theoretic operations. Pawlak in [23] checked for the natural candidate for representation of rough sets by the 3-valued membership function

Pawlak’s Landscaping with Rough Sets

53

⎧ ⎨ 1 if x belongs to the lower approximation X(x) = .5 if x belongs to the boundary region ⎩ 0 otherwise. With such functions representing rough sets X and Y , the membership function of X ∪ Y can not be represented by max(X(x), Y (x)) as is done in fuzzy set theory, since it has to coincide with the function ⎧ ⎨ 1 if x belongs to the lower approximation of X ∪ Y X ∪ Y (x) = .5 if x belongs to the boundary region ⎩ 0 otherwise, but which fails. Similarly, X ∩ Y can not be represented by the function min(X(x), Y (x)). However if the membership function is modiﬁed as below the desired result is obtained. Wygralak proposed to deﬁne and of two rough sets X, Y , by min(1, X(x) + Y (x)) if X(x) = Y (x) = .5 and [x] ⊆ X ∪ Y X Y (x) = max(X(x), Y (x)) otherwise. and XY (x) =

max(0, X(x) + Y (x) − 1) if X(x) = Y (x) = .5 and [x] ∩ X ∩ Y = ∅ min(X(x), Y (x)) otherwise.

X Y is roughly equal to X ∪ Y and X Y is roughly equal to X ∩ Y . In spite of this claim of Wygralak, one may doubt about the acceptability or otherwise of such functions as operators for conjunction and disjunction because they might miss some important properties - but it requires detailed investigations to make a ﬁnal comment. In Moscow Conference, 1988, Dubois and Prade argue that “fuzzy sets and rough sets aim to diﬀerent purposes and that it is more natural to try to combine the two models of uncertainty (vagueness for fuzzy sets and coarseness for rough sets) in order to get a more accurate account of imperfect information.” They proposed interesting mathematical constructs known as rough-fuzzy sets and fuzzy-rough sets [9]. In the opinion of the present author, there is an essential indiscernibility underlying in all kinds of vagueness - indiscernibility giving rise to both granularity and gradualness. But indiscernibles may be of various types though one can probe into some essential features of this elusive notion. These investigations had also been a favourite topic in the 80s and 90s [1,34,41,12,14,10].

2

Indiscernibilities

One major diﬀerence in the approaches to indiscernibility lies in assuming it as a ‘yes’/‘no’ type crisp concept or a graded concept. In the ﬁrst approach, objects

54

M.K. Chakraborty

x and y are either indiscernible or not - this approach is adopted in Rough Set theory by Pawlak. Several generalizations, of course, have taken place. The general feature of these is that starting from some knowledge or data, diﬀerent components (or clusters) are computed and rough sets are construed out of these components. We shall, however, restrict this study to Pawlak-rough sets only. It would be interesting to investigate if the present ‘uniﬁed’ approach can be extended to generalized rough sets as well. The second approach, as pursued in Fuzzy Set theory, presumes that x and y may be indiscernible to a degree - that is, it is a graded notion. Another very important diﬀerence rests in assuming or not assuming or weakening the transitivity property of indiscernibility. In the ﬁrst case we get standard equivalence relation on the universe of discourse, as is the base of Pawlak-rough sets. The second case generates tolerance relation (reﬂexive-symmetric) as base – rough set theory on this base is also pursued. In the third case, we have fuzzy transitivity, viz. Ind(x, y) & Ind(y, z) ≤ Ind(x, z), where Ind(x, y) represents the indiscernibility degree between x and y, and & is an algebraic operation (a t-norm, perhaps) on a suitable truth set. This graded relation without being reduced to tolerance, relaxes the notion of hard transitivity and elegantly, takes care of the gradualness aspect by using an interactive conjunction operator as follows. Let x1 , x2 , x3 , x4 , ... be a sequence of objects such that Ind(xi , xi+1 ) = .5, for all i. Now let us take the product (×) as the operator for &. Since Ind(x1 , x2 ) & Ind(x2 , x3 ) ≤ Ind(x1 , x3 ), we get Ind(x1 , x3 ) ≥ .5 × .5 = .25. If the least value .25 is taken then Ind(x1 , x4 ) may be taken as Ind(x1 , x3 ) & Ind(x3 , x4 ) = .25 × .5 = .125. Thus indiscernibility degree gradually diminishes. It means that the indiscernibility between x1 and x4 is less than that between x1 and x3 , and this is further less than the indiscernibility between x1 and x2 – a feature quite intuitively acceptable. Symmetry is naturally expected of indiscernibility. In the fuzzy case it means that Ind(x, y) = Ind(y, x). We shall discuss about reﬂexivity property later. Before that let us examine the relationship between indiscernibility relation and a concept. A concept induces an indiscernibility relation in a universe of discourse. For example, the concept A gives rise to IndA given by 1 if x, y ∈ A or x, y ∈ Ac IndA (x, y) = 0 otherwise. Concepts A, B, and C similarly give rise to the relation 1 if x ∈ X if and only if y ∈ X, X = A, B, C IndA,B,C (x, y) = 0 otherwise. An instance of this latter case is depicted in the following diagram with ﬁfteen objects, x1 to x15 .

Pawlak’s Landscaping with Rough Sets

x1

x2

x14

x4

x3 x15

x5

x13 x9 x7 x12

55

x10

x6

x8

x11

We could also show IndA,B,C by constructing the following approximation space or information system. x1 x2 x3 x4 x5 x6 x7 x8 x9 x10 x11 x12 x13 x14 x15

A Y Y Y N Y N N N Y Y N N Y Y N

B N Y Y Y Y Y N N N Y N N N N Y

C N N N N Y Y Y Y Y Y N N N N N

Y stands for ‘yes, belongs to’ and N for ‘no, does not belong to’. So IndA,B,C (x, y) = 1, if and only if the rows corresponding to x and y are identical. Thus a set of concepts generates a partition of the universe and hence equivalently gives an approximation space. Can we retrieve A (in the ﬁrst example) or A, B, C (in the second example) from the indiscernibility relations IndA or IndA,B,C ? The two values 0 and 1 of the indiscernibility relation are not suﬃcient for this purpose. Only from the

56

M.K. Chakraborty

information about indiscernibility one cannot separate A from its complement Ac since IndA (x, x) = 1 whether x ∈ A or x ∈ Ac . So, let us take a third value 2 and stipulate. 2 if x ∈ A Ind∗A (x, x) = 1 if x ∈ Ac and Ind∗A (x, y)

=

1 if x = y, x, y ∈ A or x, y ∈ Ac 0 if x = y, x ∈ A and y ∈ Ac , or x ∈ Ac and y ∈ A.

It is obvious that IndA (x, y) = 1 if and only if Ind∗A (x, y) ≥ 1 and IndA (x, y) = 0 if and only if Ind∗A (x, y) = 0. The two functions IndA and Ind∗A coincide on (x, y), x = y, but while the ﬁrst one cannot make a distinction between A and its complement, Ind∗A can do this. Similarly, in order to retrieve A ∪ B ∪ C in the second case, we need the deﬁnition 2 if x ∈ A ∪ B ∪ C Ind∗A,B,C (x, x) = 1 if x ∈ A ∪ B ∪ C and

⎧ ⎨ 1 if x = y and the rows of x and y in the information Ind∗A,B,C (x, y) = system coincide ⎩ 0 otherwise. One point to be observed here is that although while Ind∗X (x, y) ≥ 1 implies that x and y are indiscernible in terms of the basic concepts, Ind∗X (x, x), i.e. the indiscernibility of x with itself is not of the same category for all x. Also to be noted that this representation may be applied to any subset (not only A∪B ∪C) of U obtained by unions of intersections of A, B and C, in other words unions of the indiscernibility classes or blocks in U determined by them. The procedure, however, is to deﬁne Ind∗A,B,C (x, x), so that only speciﬁed unions of blocks are obtained in the backward process. We shall summarize and axiomatize such properties of Ind∗X later. At this stage we only raise an issue. Could we not think that degree of belongingness of an object to a concept is determined by and is the same as the degree of indiscernibility of the object with itself relative to the same concept? Looking from this angle, indiscernibility becomes more a primitive notion and plays a key role in the process of learning as well as of categorization. A category is created depending on the similarity of its members. Based on the degree of indiscernibility of an object, its belongingness to a set/category/class is decided. x belongs to A to the extent 2 = Ind∗A (x, x), means x is within A, x belongs to A to the extent 1 = Ind∗A (x, x) means x is in Ac . That we are interested in A and not in its complement is also represented in the assignment of a lower degree of indiscernibility to elements of the complement with themselves. The idea shall

Pawlak’s Landscaping with Rough Sets

57

play a key role in what follows and is a basis for enquiries into vagueness. Even in the case of crisp concepts, we have noticed, while retrieving A from Ind∗A the elements of A have grade 2, and elements of Ac have grade 1, and a third value 0 is also needed to construct the partition. If, however, the concepts are vague, that is, admit borderline cases, it is necessary to introduce a fourth value and the indiscernibility generated may be deﬁned as above with certain additional conditions. These ideas were published long back in Obtulowicz’s paper [18] based on a ﬁrm category-theoretic basis proposed by Higgs [12], but probably escaped attention of the researchers in this ﬁeld. We re-present from his work the main deﬁnitions and representation theorems below with little notational changes to ﬁt in the present context. Let U be a universe and L(4) be the ordered set {0 ≤ 1 ≤ 2 ≤ 3} which is a complete distributive lattice (or complete Heyting algebra). Let Ind∗ : U × U → L(4) be an indiscernibility relation that satisﬁes conditions. H1 : Ind∗ (x, y) = Ind∗ (y, x) (Symmetry) H2 : Ind∗ (x, y) ∧ Ind∗ (y, z) ≤ Ind∗ (x, z) (Transitivity) and the following roughness conditions R1 R2 R3 R4

: : : :

1 ≤ Ind∗ (x, x) for x ∈ U if 2 ≤ Ind∗ (x, y), then x = y if Ind∗ (x, y) = 1, then Ind∗ (x, x) = Ind∗ (y, y) if Ind∗ (x, x) = 2, then there exists y such that Ind∗ (x, y) = 1

The signiﬁcance of roughness conditions shall be clear from Proposition 2 below. The following two propositions establish that (U, Ind∗ ) is a representation of any Pawlak-rough set in U . Proposition 1. Let (U, Ind∗ ) be given. Then the relation R deﬁned by xRy if and only if Ind∗ (x, y) ≥ 1 is an equivalence relation, and the pair (I, B) deﬁned by I = {x : Ind∗ (x, x) = 3}, B = {x : Ind∗ (x, x) = 2} constitute the interior and boundary of a rough set in (U, R). Proposition 2. Let (U, R) be an approximation space in which (I, B), the interior and boundary pair determines a rough set. Consider the mapping Ind∗ : U × U → L(4) given by ⎧ ⎨ 3 if x ∈ I Ind∗ (x, x) = 2 if x ∈ B ⎩ 1 if x ∈ U \ (I ∪ B) and ∗

Ind (x, y) =

1 if x = y, xRy holds 0 if x = y, xRy does not hold.

Then Ind∗ satisﬁes the conditions H1 , H2 , R1 , R2 , R3 , R4 . The following important feature is also observed in the two constructions by Propositions 1 and 2.

58

M.K. Chakraborty P rop1

P rop2

P rop1

(U, Ind∗ ) −→ (U, R, I, B) −→ (U, Ind∗ ) −→ (U, R, I, B). Thus an indiscernibility satisfying the roughness conditions gives rise to a rough set in an approximation space and vice versa. One can see that the above representation does not amount to the 3-valued semantics where an object falls under a concept with three grades viz. 3 (when it is in the deﬁnite region), 2 (when it is in the boundary), and 1 (outside the boundary). We have noticed diﬃculties with such representation (by Pawlak or Wygralak) in the beginning of this article. The membership function of Obtulowicz also says that Ind∗ (x, x) & Ind∗ (x, y) ≤ Ind∗ (y, y), i.e. the degree of belongingness of x to a concept & the degree of indiscernibility of x with y ≤ the degree of belongingness of y to the concept. This criterion is the so-called ‘saturatedness’ condition that has been elaborately discussed in [5]. In Leibniz’s terms, this is the version of the doctrine of Identity of Indiscernibles, viz. if an object x has a property P and an object y is indiscernible from x, then y has the property P . In the fuzzy context, a more general condition is taken viz. α(x) & Ind(x, y) ≤ α(y), where α(x) denotes the degree of belongingness of x to the fuzzy set α. The only addition here is the conceptual indulgence to the assumption that belongingness degree of x to a fuzzy set is the same as Ind∗ (x, x). It is also signiﬁcant to notice that the fuzzy set theoretic operators ‘max’ and ‘min’ are now applicable to obtain the union and intersection. A summary of what has been said so far is the following: – The underlying indiscernibility relation for any vague concept in U is a relation Ind∗ satisfying the conditions H1 , H2 , R1 , R2 , R3 , R4 . – Such a relation, which is a particular kind of fuzzy equivalence relation determines uniquely a rough set in the approximation space (U, R) where R is virtually the underlying indiscernibility and conditions R1 , R2 , R3 , R4 determine the interior and boundary of the rough set. – Conversely, any rough set in (U, R) given by the interior and boundary can be generated by an indiscernibility relations satisfying H1 , H2 , R1 , R2 , R3 , R4 . – This representation is one-to-one. – In the special case when the boundary region is empty, the condition R4 is dropped. We need a three-element complete Heyting algebra and Propositions 1 and 2 may be written accordingly. Earlier examples with the concept A, and concepts A, B, C are instances of such representation. One interesting extension of Obtulowicz’s representation of rough sets suggests itself, but has never been taken up. Formally, one can add more categories other than 0, 1, 2 and 3. For instance, let us take one more viz. 4. All the roughness conditions remain the same except that the condition R4 shall now be read as R4 : for each x, Ind(x, x) = 2 or 3 implies that there exists y such that Ind(x, y) = 1.

Pawlak’s Landscaping with Rough Sets

59

The rough set is now extended to (U, R, I, B1 , B2 ), where there are two layers in the boundary, elements of I have grade 4, while those of B1 and B2 have grades 3 and 2 respectively. The representation theorems now take the following shapes. Proposition 3. Let Ind∗ : U × U → L(5)(≡ {0 ≤ 1 ≤ 2 ≤ 3 ≤ 4} be an indiscernibility relation satisfying conditions H1 , H2 , R1 , R2 , R3 , R4 . Then R deﬁned by xRy if and only if Ind∗ (x, y) ≥ 1 is an equivalence relation, and the triple (I, B1 , B2 ) deﬁned by I = {x : Ind∗ (x, x) = 4}, B1 = {x : Ind∗ (x, x) = 3}, B2 = {x : Ind∗ (x, x) = 2} constitute the interior and the two layers of boundaries of a rough set in (U, R). Proposition 4. Let (U, R) be an approximation space in which (I, B1 , B2 ) determines a rough set. Then the mapping Ind∗ : U × U → L(5) given by ⎧ 4 if x ∈ I ⎪ ⎪ ⎨ 3 if x ∈ B1 ∗ Ind (x, x) = 2 if x ∈ B2 ⎪ ⎪ ⎩ 1 if x ∈ U \ (I ∪ B1 ∪ B2 ) and ∗

Ind (x, y) =

1 if x = y, xRy holds 0 if x = y, xRy does not hold.

Then Ind∗ satisﬁes the conditions H1 , H2 , R1 , R2 , R3 , R4 . Layers of the boundary may be enhanced arbitrarily, but ﬁnitely. The interpretation of these layers is zones of objects of gradually weaker possibilities of falling under the concept that render gradualness along with granularity. One can also see the possibility of interpreting membership values under rough membership function [22] as values of the lattice (after normalization, of course). Elements of a block with lower rough membership may be placed in the weaker layer of the boundary. Element of beyond-possible zone should be given the value 1 instead of 0 which should be retained to denote the discernibility of x, y, x = y and to determine the partition of the Universe. In ﬁnite situations there should not be any diﬃculty. The deﬁnitions of union intersection and complementation may be suitably deﬁned by using max, min and reversal of grades. There shall be a departure in this approach from that of Pawlak. ‘Roughness’ should no longer be considered as an adjective to an ordinary set in an approximation space but as a pair (I, B) of unions of blocks of the space, I being called the interior and B the boundary. This approach is equivalent to that taken by us in [2]. It is also in conformity with the philosophy of rough sets viz. “we ‘see’ elements of the universe through available informations” and hence “some elements may be ‘seen’ identical” and “this is to mean that if we see a set through informations, only, the approximations (lower and upper) can be observed.” And further “a vague property determines not only a single set of elements falling under the property but a family of sets which can be identiﬁed with this property up to indiscernibility.” All the above lines within quotation marks are from Pawlak’s writings glued together. So the present approach, though a departure, draws the support from his own feelings too.

60

3

M.K. Chakraborty

Conclusions

Like fuzzy sets, rough sets also have wide usages for example, in artiﬁcial intelligence, cognitive sciences, knowledge discovery from database, machine learning, expert system, inductive reasoning and pattern recognition. As in the case of any successful theory, it has to look back at one point of time. “The rough set theory has reached such a state that some kind of summary of its theoretical foundation is a must” [28] - this realization of Pawlak is quite justiﬁable. As it appears from the preceding discussions, the approach of Obtulowicz may serve as quite a reasonable foundation. This approach also suggests a kind of uniﬁcation of fuzzy sets and rough sets. Starting with some sort of indiscernibility in the Universe which is at the base of any vague concept and which arises out of data (concrete or subjective), objects of the Universe are categorized. In such a categorization, some tokens (not necessarily numbers) with varying algebraic structures play a role. A mathematical entity emerges representing the vague concept. In this representation, the following philosophy is adopted: “The degree of existence of an object in a concept is the degree to which the object is indiscernible with itself relative to the underlying the concept.” The mathematical entity is sometimes a fuzzy set and sometimes a rough set. Divergence occurs because of the nature of the indiscernibility (crisp or fuzzy) and the choice of categories (how many?) and their structures. Incorporation of layers in the boundary as proposed in the paper for the ﬁrst time brings rough sets closer to fuzzy sets. We think that this approach could help in erasing Pawlak’s persistent feeling of a sort of ‘supremacy’ of classical set theory over fuzzy set theory or rough set theory. This feeling is expressed in statements like “fuzzy set involves more advanced mathematical concepts real numbers and functions - whereas in classical set theory the notion of set is used as a fundamental notion of whole mathematics and is used to derive any other mathematical concept e.g. numbers and functions. Consequently, fuzzy set theory cannot replace classical set theory, because, in fact, the theory is needed to deﬁne fuzzy sets.” [21] Again, “in a manner similar to fuzzy set theory, rough set theory is not an alternative to classical set theory but it is embedded in it.”[22] Pawlak’s concern about the foundational problems of classical set theory and interest in the alternatives like, multisets, multi-fuzzy sets, Blizard sets, general sets, Mereology (Lesniewski), Alternative set theory (Vopenka), Penumbral set theory (Apostoli and Kanda) are well known [21,25]. In the proposal of Obtulowicz, what would, in fact, be needed at the beginning is a Universe, a collection of tokens with some structures including order (in particular the numbers) and the notion of indiscernibility which needs only an understanding of ‘pair’ and ‘correspondence’. These may constitute a nice, intuitively acceptable beginning. If one casts an oblique eye, we can take refuge to categorical foundation (Higgs [12], Banerjee and Chakraborty [3], and others). At this point we would like to draw the attention of readers to a paper of Goguen published in 1974. He claims “Ideally, we would like a foundation for fuzzy sets which justiﬁes the intuitive identiﬁcation of fuzzy sets with (inexact) concepts, and in which the familiar set operations are uniquely and inevitably determined. These

Pawlak’s Landscaping with Rough Sets

61

desires are clearly though not explicitly expressed in Zadeh (1965), and they are all satisﬁed by the system given in this paper.” He used as paradigm Lawvere’s “relatively unknown axiomatization of sets in the language of category theory”. In one of our papers [3] Higgs’ and Goguen’s categories are discussed and compared. But category-theoretic approach is still not popular among practitioners. If toiled and fuzzy as well as rough sets are found soundly based on categorical grounds, the gains appear to be enormous - reinstatement of ‘vagueness’ within the discourses of mathematics, logic and thus by sciences. A recent work by Skowron [39] once again brings to focus the importance of the study of vagueness and the role of rough set theoretic methods in such studies. He introduces two operations viz. inductive extension and granulation of approximation spaces and emphasizes on “important consequences of the paper for research on approximation of vague concepts and reasoning about them in the framework of adoptive learning.” He thinks that “this (adoptive learning) requires developing a new approach to vague concepts going beyond the traditional rough or fuzzy approaches.” This paper extends the notion of approximation space by incorporating rough inclusion or graded inclusion. It would be an interesting project to investigate if Obtulowicz’s proposal may be extended to this generalized context also. We engage not only into crisp (two-valued) talks about vagueness, but into multi-valued talks too (theory of graded consequence [6,7]) or we also talk ‘roughly’ (theory of rough consequence [5]). The underlying motivation to deﬁne graded consequence or rough consequence is to allow room for vagueness in the metalogical concepts like consistency, consequence, tautologihood, completeness, etc. This latter notion viz. rough consequence has its origin in Pawlak’s insightful work on rough truth [24] where he states “the rough (approximate) truth and falsity represent our partial knowledge about the world and with the increase of our knowledge the roughly true (or false) formulas tend to be more true (or false) and approach the truth and falsity closer and closer.” One is bound to recall Zadeh when he claims that the notion of truth itself is fuzzy. It would not be out of place to mention that the ﬁrst recorded works on rough logics are by Orlowska and Pawlak [19] and Rasiowa and Skowron [36]. With the advent of graded and rough consequences, the scenario of mathematics should change, in that there may be some mathematical predicates, truth of sentences relative to which may be partial and derivations involving which may not be of full strength. Professor Pawlak was an artist. His favourite example of vagueness was a ‘beautiful’ painting [22]. He was fond of landscaping. His work on rough sets is also a beautiful landscape-installation - to which we oﬀer this humble bouquet of ours that might develop roots striving to become an integral part of this scenario.

References 1. M. Banerjee. A Categorial Approach to the Algebra and Logic of the Indiscernible. Ph.D Thesis, University of Calcutta, 1993. 2. M. Banerjee and M.K. Chakraborty. Rough sets through algebraic logic. Fundamenta Informaticae, 28(3,4): 211–221, 1996.

62

M.K. Chakraborty

3. M. Banerjee and M.K. Chakraborty. Foundations of vagueness: a category-theoretic approach. Electronic Notes in Theoretical Computer Science, 82(4), 2003. 4. M.K. Chakraborty and M. Banerjee. Rough consequence. Bulletin of the Polish Academy of Sciences (Mathematics), 41(4):299–304, 1993. 5. M.K. Chakraborty and E. Orlowska. Substitutivity principles in some theories of uncertainty. Fundamenta Informaticae, 32:107–120, 1997. 6. M.K. Chakraborty and S. Basu. Graded consequence and some metalogical notions generalized. Fundamenta Informaticae, 32:299–311, 1997. 7. M.K. Chakraborty and S. Basu. Approximate reasoning methods in vagueness: graded and rough consequences. ICS Research Report, 29, Warsaw University of Technology, 1995. 8. S. Demri and E. Orlowska. (eds.) Incomplete Information: Structure, Inference, Complexity. Monographs in Theoretical Computer Science, Springer-Verlag, Heidelberg, 2002. 9. D. Dubois and H. Prade. Rough fuzzy sets and fuzzy rough sets. In Proc. International Conference on Fuzzy Sets in Informatics, Moscow. 1988, 20–23. 10. M. Eytan. Fuzzy sets: a topos-logical point of view. Fuzzy Sets and Systems, 5:47–67, 1981. 11. J. Goguen. Concept representation in natural and artiﬁcial languages: axioms, extensions and applications for fuzzy sets. International Journal for Man-Machine Studies, 6:513–561, 1975. 12. D. Higgs. A categorical approach to Boolean-valued set theory. Preprint, 1973. 13. D. Hyde. From heaps and gaps to heaps of gluts. Mind, 106:440–460, 1997. 14. J. Jacas. On the generators of T-indistinguishability operator. Stochastica, XIII: 49–63, 1988. 15. R. Keefe. Theories of Vagueness, Cambridge Studies in Philosophy, Cambridge, UK, 2000. 16. R. Keefe and P. Smith. (eds.) Vagueness: A Reader, MIT Press, Massachusetts, MA, 1997. 17. F. Keikeben. http://members.aol.com/kiekeben/theseus.html, 2000. 18. A. Obtulowicz. Rough sets and Heyting algebra valued sets. Bulletin of the Polish Academy of Sciences (Mathematics), 13(9-10):667–671, 1987. 19. E. Orlowska and Z. Pawlak. Representation of non-deterministic information. Theoretical Computer Science, 29:27–39, 1984. 20. Z. Pawlak. Rough sets. International Journal of Computer and Information Sciences, 11:341–356, 1982. 21. Z. Pawlak. Some issues on rough sets. Transactions of Rough sets I, 1–58, 1998. 22. Z. Pawlak. A treatise on rough sets. Transactions on Rough sets IV, 1–17, 2005. 23. Z. Pawlak. Rough sets and fuzzy sets. Fuzzy Sets and Systems 17:99–102, 1985. 24. Z. Pawlak. Rough logic. Bulletin of the Polish Academy of Sciences(Technical Sciences), 35(5-6):253–258, 1987. 25. Z. Pawlak. Hard and soft sets. ICS Research Report, 10/94, Warsaw University of Technology, 1994. 26. Z. Pawlak. Vagueness – a rough set view. LNCS 1261, Springer, 1997, 106–117. 27. Z. Pawlak. Vagueness and uncertainty: a rough set perspective. Computational Intelligence: An International Journal, 11:217–232, 1995. 28. Z. Pawlak. Rough sets, present state and further prospects. ICS Research Report, 15/19, Warsaw University of Technology, 1995. 29. Z. Pawlak and A. Skowron. Rudiments of rough sets. Information Sciences, to appear.

Pawlak’s Landscaping with Rough Sets

63

30. Z. Pawlak and A. Skowron. Rough sets: some extensions. Information Sciences, to appear. 31. Z. Pawlak and A. Skowron. Rough sets and Boolean reasoning, Information Sciences, to appear. 32. G. Priest and R. Routley. First historical introduction: a preliminary history of paraconsistent and dialethic approaches. In Priest, Routley and Normann, editors, Paraconsistent Logic, Essays on the Inconsistent. Philosophia Verlag, MunchenHamden-Wien, 1989, 1–75. 33. L. Polkowski. Rough Sets: Mathematical Foundations. Advances in Soft Computing, Physica Verlag, Hiedelberg, 2002. 34. A. Pultr, Fuzziness and fuzzy equality. In H.J. Skala, S. Termini and E. Trillas, editors, Aspects of Vagueness. D. Reidel Publishing Co., Dordrecht, Holland, 1984, 119–135. 35. P.T. Raju. The principle of four-coloured negation in Indian philosophy. Review of Metaphysics 7:694-713, 1953. 36. H. Rasiowa and A. Skowron. Rough concepts logic in computation theory. In A. Skowron, editor, LNCS 208, Springer, 1985, 288–297. 37. B. Russell. Vagueness. Australian Journal of Philosophy, 1:84–92, 1923. 38. A. Skowron. The relationship between the rough set theory and evidence theory. Bulletin of the Polish Academy of Sciences (Technical Sciences), 37(1-2):87–90, 1989. 39. A. Skowron. Rough sets and vague concepts. Fundamenta Informaticae, 64: 417–431, 2005. 40. A. Skowron and J.W. Grzymala-Busse. From rough set theory to evidence theory. In Yager, Fedrizzi and Kacprzyk, editors, Advances in the Dempster-Shafer Theory of Evidence. John Wiley & Sons, New York, 1994, 193–236. 41. E. Trillas and L. Valverde. An inquiry into indistinguishability operators. In H.J. Skala, S. Termini and E. Trillas, editors, Aspects of Vagueness. D. Reidel Publishing Co., Dordrecht, Holland, 1984, 231–256. 42. H. Wang. Beyond Analytic Philosophy. MIT Press, Cambridge, 1986. 43. D. Wiggins. Sameness and Substance. Oxford Blackwell, 1980, 92–94. 44. M. Wygralak. Some remarks on rough and fuzzy sets. BUSEFAL 21, 1985, 43–49. 45. L.A. Zadeh. Fuzzy Sets. Information and Control, 8:338–353, 1965.

A Comparison of Pawlak’s and Skowron–Stepaniuk’s Approximation of Concepts Anna Gomoli´ nska University of Bialystok, Department of Mathematics, Akademicka 2, 15267 Bialystok, Poland [email protected]

Abstract. In this article, we compare mappings of Pawlak’s lower and upper approximations of concepts with those proposed by Skowron and Stepaniuk. It is known that both approaches coincide for the standard rough inclusion, so we consider the case of an arbitrary rough inclusion function. Even if the approximation space investigated is based on an arbitrary non-empty binary relation, the lower approximation mappings are equal in both approaches. Nevertheless, the upper approximation mappings are diﬀerent in general.

In view of many generalizations and extensions of rough set theory some kind of uniﬁcation of the basic theory seems to be badly needed. (Z. Pawlak [1], p. 10)

1

Introduction

Nowadays, the Pawlak rough approximation of concepts [2,3,4,5] has become a classical research topic. Lower and upper rough approximations have been investigated by many researchers in the rough set community, to mention [6,7,8] by way of example. Skowron and Stepaniuk’s proposal regarding rough approximation of concepts [9,10] is well-known, yet less popular among researchers as a subject of study. A possible reason may be the fact that both approaches coincide for the standard rough inclusion, whereas this very function is the most known among rough inclusions. The aim of this paper is to study and to compare both Pawlak’s and Skowron– Stepaniuk’s approaches to approximation of concepts in the rough-set framework. To this end, we relax the usual assumptions about the approximation space considered. We start with an approximation space, understood as a structure M = (U, , κ), where U (the universe) is a non-empty set of objects, is a non-empty binary relation on U , and κ is a mapping on the set of pairs of sets of objects called a rough inclusion function. Step by step, we consider spaces

The research was supported by the grant 3T11C00226 from the Ministry of Science of the Republic of Poland.

J.F. Peters et al. (Eds.): Transactions on Rough Sets VI, LNCS 4374, pp. 64–82, 2007. c Springer-Verlag Berlin Heidelberg 2007

A Comparison of Pawlak’s and Skowron–Stepaniuk’s Approximation

65

based on serial relations, reﬂexive relations, transitive relations, symmetric (and in particular, tolerance) relations, and equivalence relations. By deﬁnition, κ is assumed to satisfy two postulates only which are in accordance with the axioms of rough mereology. However, we also investigate cases, where κ fulﬁlls additional conditions. In the paper, we examine two pairs of approximation mappings in line with Pawlak’s approach and two pairs of approximation mappings in line with Skowron–Stepaniuk’s proposal. Each pair consists of a lower approximation mapping and an upper approximation mapping. Both in the Pawlak case as well as in the Skowron–Stepaniuk case, approximation mappings constituting one pair are viewed as basic, whereas mappings forming the remaining pair are “deﬁnable” versions of the basic mappings. As regarding the results, the basic lower approximation mappings are equal in both approaches, and similarly for their “deﬁnable” variants. Basic upper approximation mappings (and similarly for their “deﬁnable” counterparts) are diﬀerent in general and may be compared only under special conditions put on κ. Therefore, we mainly try to compare these mappings indirectly via their properties. Apart from that, we aim at a uniform presentation of facts about lower and upper approximation mappings. Some of the facts are new, others are only recalled. We try to answer such questions as: What are the results of application of a given approximation mapping to the empty set and to the whole universe? What are the relationships among various forms of approximation? For instance, how is the lower approximation related to the upper one? Moreover, how are approximations of a concept related to the concept itself? Are the approximation mappings under investigation monotone, and if it is the case, what are the consequences? Last but not least, what a mapping may be obtained by various compositions of the approximation mappings? As we shall see, the mappings investigated can lack some essential properties attributed to an approximation mapping if the relation , underlying a given approximation space M , is not reﬂexive. For example, a lower approximation of a concept may not be included in that concept. However, slightly abusing the terminology, we shall use the names ‘lower approximation’ and ‘upper approximation’ for the sake of uniformity. Basic terminology and notation is introduced in Sect. 2. In Sect. 3, we present the notion of a rough approximation space and the mappings of Pawlak’s as well as Skowron–Stepaniuk’s lower and upper rough approximation of concepts. In Sect. 4, properties of these mappings are studied in the case of an approximation space based on an arbitrary non-empty binary relation . In Sect. 5, we examine the approximation mappings for special cases of approximation spaces, where and/or its converse relation −1 are serial and where is, in turn, a reﬂexive relation, a transitive relation, a symmetric relation, and an equivalence relation. The results are summarized brieﬂy in the last section.

2

Preliminaries

Let X, Y be any sets. Throughout the paper, the power set of X, the cardinality of X, the Cartesian product X × X, the identity mapping on X, and the set

66

A. Gomoli´ nska

of all mappings f : X → Y will be denoted by ℘X, #X, X 2 , idX , and Y X , respectively. Consider (partially) ordered sets (X, ≤) and (Y, ). A mapping f : X → Y is referred to as monotone, written f ∈ MON, if for any x, y ∈ X, x ≤ y implies f x f y. The operation of composition of relations will be denoted by ◦. In the case of mappings, the composition of f : X → Y with g : Y → Z is a mapping g ◦ f : X → Z such that for any x ∈ X, (g ◦ f )x = g(f x). For any sets X, Y , deﬁne a relation on (℘Y )℘X and operations , on ((℘Y )℘X )2 such that for any mappings f, g : ℘X → ℘Y and any Z ⊆ X, def

f g ⇔ ∀Z ⊆ X.f Z ⊆ gZ, def

def

(f g)Z = f Z ∩ gZ & (f g)Z = f Z ∪ gZ.

(1)

By assumption, ◦ will take the precedence of the operations just deﬁned, whereas the logical connectives of conjunction and disjunction will take the precedence of implication and double implication. Proposition 1. For any mappings f, g, h : ℘X → ℘Y , we have: (a) f f (b) f g & g f ⇒ f = g (c) f g & g h ⇒ f h (d) f g ⇒ f ◦ h g ◦ h (e) h ∈ MON & f g ⇒ h ◦ f h ◦ g (f ) f g h ⇔ f h & g h (g) f g h ⇔ f g & f h The proof is easy and, hence, omitted. Let us only note that is a partial ordering on (℘Y )℘X in virtue of (a)–(c).

3

Rough Approximation Spaces

The notion of a rough approximation space was obtained by Prof. Pawlak in the early 80’s of the 20th century as one of the results of investigations on approximation of vague concepts in information systems [2,3,4,5]. This basic notion was next reﬁned and generalized in several directions (see, e.g., [11,12,13,14,15]), yet we shall only focus upon the extension proposed by Skowron and Stepaniuk in [9,10], and elaborated in a series of research articles [16,17,18]. In [19,20]1 , Polkowski and Skowron introduced and characterized axiomatically the formal notion of a rough inclusion. Although this notion is unnecessary when discussing Pawlak’s classical approach, it is fundamental for Skowron– Stepaniuk’s one. Consider a non-empty set U of entities called objects. In com1

See also more recent papers.

A Comparison of Pawlak’s and Skowron–Stepaniuk’s Approximation

67

pliance with rough mereology, by a rough inclusion function (RIF for short) upon U we understand any mapping κ : (℘U )2 → [0, 1] satisfying rif 1 and rif 2 below: def

rif 1 (κ) ⇔ ∀X, Y.(κ(X, Y ) = 1 ⇔ X ⊆ Y ) def

rif 2 (κ) ⇔ ∀X, Y, Z.(Y ⊆ Z ⇒ κ(X, Y ) ≤ κ(X, Z)) RIFs are intended as mappings measuring the degrees of inclusion of sets of objects in sets of objects. Apart from the above postulates, one may consider other conditions, for instance, def

rif 3 (κ) ⇔ ∀X = ∅.κ(X, ∅) = 0, def

rif 4 (κ) ⇔ ∀X = ∅.∀Y.(κ(X, Y ) = 0 ⇒ X ∩ Y = ∅), def

rif 4∗ (κ) ⇔ ∀X = ∅.∀Y.(X ∩ Y = ∅ ⇒ κ(X, Y ) = 0), def

rif 5 (κ) ⇔ ∀X = ∅.∀Y.(κ(X, Y ) = 0 ⇔ X ∩ Y = ∅). One can easily see that rif 4∗ (κ) implies rif 3 (κ), whereas rif 5 (κ) if and only if rif 4 (κ) and rif 4∗ (κ). The most famous RIF is the standard one, deﬁned for the ukasiewicz [21] and ﬁnite universe and denoted by κ£ here, which goes back to L is based on the frequency count. κ£ is given by #(X∩Y ) if X = ∅ £ #X κ (X, Y ) = (2) 1 otherwise and fulﬁlls not only rif 1 , rif 2 but also rif 5 and some other conditions. By a rough approximation space we mean a triple M = (U, , κ), where U — the universe of M — is a non-empty set of objects as earlier, is a non-empty binary relation on U , and κ is a RIF upon U . Objects will be denoted by u with sub/superscripts if needed. Sets of objects of U are viewed as concepts of M . With every object u, there are associated two basic concepts: the image and the co-image of {u}, → {u} and ← {u}, deﬁned along the standard lines and called elementary granules of information 2 drawn to u. It is worth recalling that → {u} = −1← {u} and ← {u} = −1→ {u}. Let us note that induces mappings Γ , Γ∗ : U → ℘U , called uncertainty mappings in line with Skowron–Stepaniuk’s approach, such that for every object u ∈ U, (3) Γ u = ← {u} & Γ∗ u = → {u}. Thus, elementary granules of information are simply values of Γ, Γ ∗ . Clearly, Γ∗ = Γ−1 , and u ∈ Γ∗ u if and only if (u, u ) ∈ , i.e., if and only if u ∈ Γ u . Moreover, Γ∗ = Γ if is symmetric. On the other hand, every mapping Γ : U → ℘U induces a relation Γ on U such that for any objects u, u ∈ U , (u, u ) ∈ Γ ⇔ u ∈ Γ u . 2

(4)

The term ‘information granule’ was proposed by Zadeh [22] to denote a clump of objects drawn together on the basis of indiscernibility, similarity or functionality.

68

A. Gomoli´ nska

Since ΓΓ = Γ and Γ = , structures (U, , κ) and (U, Γ , κ) are interdeﬁnable, and similarly for (U, Γ, κ) and (U, Γ , κ). In the classical Pawlak approach, a rough approximation space is a pair (U, ), where U is a ﬁnite non-empty set and is an equivalence relation understood as a relation of indiscernibility of objects. Then, elementary granules of information are equivalence classes, i.e. sets of objects indiscernible from one another. Clearly, a natural augmentation of (U, ) with a RIF κ, results in an approximation space (U, , κ) in line with our approach. Keeping with the recent state-of-art, one can say that Skowron–Stepaniuk’s approximation spaces, introduced in [9,10], are of the form N = (U, Γ$ , κ$ ), where Γ$ is an uncertainty mapping such that for every object u, u ∈ Γ$ u, κ$ is a RIF, and $ is a list of tuning parameters to obtain a satisfactory quality of approximation of concepts. For the latter, such spaces are called parameterized approximation spaces as well. Henceforth, the parameters $ will be dropped for simplicity. One can easily see that N is based on a reﬂexive relation3 , e.g. Γ$ . Due to our earlier observations on interdeﬁnability of approximation spaces and the corresponding structures based on uncertainty mappings, and slightly abusing the original terminology, we shall think of Skowron–Stepaniuk’s approximation spaces as structures of the form (U, , κ), where is a reﬂexive relation on U . In the sequel, a concept X is referred to as -deﬁnable (resp., −1 -deﬁnable) if it is a set-theoretical union of elementary granules of the form Γ u (resp., Γ∗ u). Henceforth, references to will be omitted whenever possible. For instance, we shall write Γ and Γ ∗ instead of Γ and Γ∗ , respectively. Where is symmetric, both forms of deﬁnability coincide, so we may simply speak of deﬁnable or undeﬁnable concepts. The main idea underlying the Pawlak rough approximation of concepts is that even if a concept is not deﬁnable in a given space, it can be approximated from the inside and the outside by deﬁnable concepts. In this way, the Pawlak lower and upper rough approximation mappings, low∪ , upp∪ ∈ (℘U )℘U , respectively, are obtained such that for any concept X, def def low∪ X = {Γ u | Γ u ⊆ X} & upp∪ X = {Γ u | Γ u ∩ X = ∅}. (5) The lower approximation of X, low∪ X, is the largest -deﬁnable concept included in X, whereas the upper approximation of X, upp∪ X, is the least deﬁnable concept containing X provided that is serial. The diﬀerence bnd∪ X = upp∪ X − low∪ X def

(6)

is called the boundary region of X. When this region is empty, X is referred to as exact ; otherwise it is rough. In Pawlak’s approximation spaces, it turns out that a concept is exact if and only if it is deﬁnable. Apart from low∪ , upp∪ , we shall also refer to the mappings low, upp ∈ (℘U )℘U given below as the Pawlak lower and upper rough approximation mappings, respectively: 3

Primarily, parameterized approximation spaces were based on reﬂexive and symmetric (i.e., tolerance) relations.

A Comparison of Pawlak’s and Skowron–Stepaniuk’s Approximation def

def

lowX = {u | Γ u ⊆ X} & uppX = {u | Γ u ∩ X = ∅} ∪

69

(7)

∪

As a matter of fact, low = low and upp = upp for the Pawlak approximation spaces. In a general case, however, low, upp diﬀer from their -deﬁnable versions low∪ , upp∪ , respectively. The lower approximation of X, lowX, may be viewed as the set consisting of all objects u which surely belong to X since their elementary granules Γ u are included in X. On the other hand, the upper approximation of X, uppX, may be perceived as the set consisting of all objects u which possibly belong to X since their elementary granules Γ u overlap with X. The Skowron–Stepaniuk lower and upper rough approximation mappings, lowS, uppS ∈ (℘U )℘U , respectively, are deﬁned by the following conditions, for any concept X, def

def

lowS X = {u | κ(Γ u, X) = 1} & uppS X = {u | κ(Γ u, X) > 0}.

(8)

That is, the lower approximation of X, lowS X, consists of all objects u that their elementary granules Γ u are included in X to the highest degree 1. On the other hand, the upper approximation of X, uppS X, is the set of all objects u that their elementary granules Γ u are included in X to some positive degree. The boundary region of X is deﬁned as the set def

bndS X = {u | 0 < κ(Γ u, X) < 1}.

(9)

Mappings lowS∪ , uppS∪ , being -deﬁnable versions of lowS , uppS , are also referred to as the Skowron–Stepaniuk lower and upper rough approximation mappings, respectively. They are given by the following equalities: def lowS∪ X = {Γ u | κ(Γ u, X) = 1} def {Γ u | κ(Γ u, X) > 0} (10) uppS∪ X = Obviously, we can repeat the construction of approximation mappings for −1 what can be useful if is not symmetric. As a result, mappings low∗ , upp∗ , lowS∗ , and uppS∗ may be derived (as well as their −1 -deﬁnable versions which will not be presented here), where for any concept X, low∗ X = {u | Γ ∗ u ⊆ X} & upp∗ X = {u | Γ ∗ u ∩ X = ∅}, def

def

lowS∗ X = {u | κ(Γ ∗ u, X) = 1} & uppS∗ X = {u | κ(Γ ∗ u, X) > 0}. (11) def

def

The mapping upp∗ isparticularly important for our purposes. It turns out that for any concept X, {Γ u | u ∈ X} = upp∗ X. Mappings low, upp, lowS , and uppS will be viewed as basic. Thus, for every basic mapping f , f ∪ = upp∗ ◦ f.

(12)

Example 1. For the sake of illustration of the approximation mappings, consider a set U = {3, . . . , 10} consisting of 8 objects denoted by 3, . . . , 10, and a binary

70

A. Gomoli´ nska

relation on U generating the uncertainty mappings4 Γ, Γ ∗ shown in Table 1. is reﬂexive, so it is a similarity relation (but even not a tolerance relation). Let κ be any RIF as earlier and κ1 , κ2 be such RIFs that for any concepts X, Y where X = ∅, κ1 (X, Y ) = 0 ⇔ X ∩ upp∗ Y = ∅, κ2 (X, Y ) = 0 ⇔ uppX ∩ upp∗ Y = ∅.

(13)

Both κ1 and κ2 satisfy rif 4 since κ2 (X, Y ) = 0 implies κ1 (X, Y ) = 0, and the latter implies X ∩ Y = ∅. Indeed, X ⊆ uppX and Y ⊆ upp∗ Y due to reﬂexivity of . It is easy to see that κ1 (X, Y ) > 0 if and only if there exist u ∈ X and u ∈ Y such that (u, u ) ∈ . Furthermore, κ2 (X, Y ) > 0 if and only if there exist u ∈ X and u ∈ Y such that (u, u ) ∈ ◦ . Let uppS1 and uppS2 denote the Skowron–Stepaniuk upper approximation mappings based on κ1 and κ2 , respectively, i.e., for any concept X and i = 1, 2, uppSi X = {u | κi (Γ u, X) > 0}.

(14)

That is, u ∈ uppS1 X if and only if Γ u ∩ upp∗ X = ∅, and u ∈ uppS2 X if and only if (upp ◦ Γ )u ∩ upp∗ X = ∅. Values of upp ◦ Γ are given in Table 1. One can show that (15) upp uppS1 uppS2 , yet the converse inclusions may not hold in general. To see this, consider X = {3, 4}. Note that lowX = {4}, low∗ X = ∅, uppX = {3, 4, 5}, upp∗ X = {3, 4, 6}, uppS1 X = {3, 4, 5, 6}, and uppS2 X = {3, 4, 5, 6, 10}. Table 1. Values of Γ , Γ ∗ , and upp ◦ Γ u 3 4 5 6 7 8 9 10

4

Γu {3,4,6} {3,4} {3,4,5,6} {6,10} {7,8,9} {7,8} {8,9} {9,10}

Γ ∗u {3,4,5} {3,4,5} {5} {3,5,6} {7,8} {7,8,9} {7,9,10} {6,10}

upp(Γ u) {3,4,5,6} {3,4,5} {3,4,5,6} {3,5,6,10} {7,8,9,10} {7,8,9} {7,8,9,10} {6,7,9,10}

Properties of Approximation Mappings

We ﬁrst investigate properties of the basic mappings low, upp, lowS , and uppS , where is an arbitrary non-empty relation on U . Henceforth, f will denote low, upp or uppS . 4

We drop references to for simplicity.

A Comparison of Pawlak’s and Skowron–Stepaniuk’s Approximation

71

Proposition 2. Let X, Y be any concepts. Then, we have: (a) low = lowS uppS (b) rif 4 (κ) ⇒ upp uppS (c) upp∅ = ∅ & lowU = uppS U = U (d) {u | Γ u = ∅} ⊆ lowX = U − upp(U − X) (e) X ⊆ Y ⇒ f X ⊆ f Y (f ) f (X ∩ Y ) ⊆ f X ∩ f Y ⊆ f X ∪ f Y ⊆ f (X ∪ Y ) Proof. We prove (b), (d), and (e) only. To this end, consider any concepts X, Y and any object u. For (b) assume that rif 4 (κ) holds. Consider u ∈ uppX. By the deﬁnition of upp, (b1) Γ u ∩ X = ∅. Hence, Γ u = ∅. As a consequence, κ(Γ u, X) > 0 by the assumption and (b1). By the deﬁnition of uppS , u ∈ uppS X. Thus, uppX ⊆ uppS X. Immediately, upp uppS by the deﬁnition of . In case (d), ﬁrst suppose that Γ u = ∅. Hence, κ(Γ u, X) = 1 in virtue of rif 1 (κ), i.e., u ∈ lowX by the deﬁnition of low. In the sequel, u ∈ lowX if and only if (by the deﬁnition of low) Γ u ⊆ X if and only if Γ u ∩ (U − X) = ∅ if and only if (by the deﬁnition of upp) u ∈ upp(U − X) if and only if u ∈ U − upp(U − X). For (e) assume (e1) X ⊆ Y . First, let f = low and suppose that u ∈ lowX. By the deﬁnition of low, Γ u ⊆ X. Hence, Γ u ⊆ Y by (e1). Again by the deﬁnition of low, u ∈ lowY . Thus, lowX ⊆ lowY . Next, let f = upp and u ∈ uppX. By the deﬁnition of upp, Γ u ∩ X = ∅. Hence, Γ u ∩ Y = ∅ by (e1). By the deﬁnition of upp, u ∈ uppY . Thus, uppX ⊆ uppY . Finally, where f = uppS , assume that u ∈ uppS X. By the deﬁnition of uppS , (e2) κ(Γ u, X) > 0. In virtue of (e1) and rif 2 (κ), κ(Γ u, X) ≤ κ(Γ u, Y ). Hence, κ(Γ u, Y ) > 0 by (e2). By the deﬁnition of uppS , u ∈ uppS Y . Thus, uppS X ⊆ uppS Y .

Let us comment upon the properties. In virtue of property (a), Pawlak’s and Skowron–Stepaniuk’s approaches coincide as regarding the lower approximation. In the latter approach, the lower approximation of a concept is always included in the upper one. The Pawlak upper approximation is, in general, incomparable with the Skowron–Stepaniuk upper approximation unless some additional assumptions like rif 4 (κ) are made. Due to (b), the Pawlak upper approximation of a concept is included in the Skowron–Stepaniuk one if rif 4 (κ) is assumed. As we shall see later on, rif 4∗ (κ) will guarantee that the Skowron–Stepaniuk upper approximation of a concept is included in the Pawlak one provided that −1 is serial (see Proposition 6g). By (c), the Pawlak upper approximation of the empty set is empty as well. On the other hand, both the lower approximation of the universe as well as the Skowron–Stepaniuk upper approximation of the universe are equal to the whole universe. According to (d), the Pawlak lower and upper approximations are dual to each other. Moreover, all objects u with empty elementary granules Γ u belong to the lower approximation of any concept. By (e), the lower and upper approximation mappings are monotone. Property (f), being

72

A. Gomoli´ nska

a direct consequence of (e), may be strengthened for the Pawlak approximation mappings as follows: low(X ∩ Y ) = lowX ∩ lowY & upp(X ∪ Y ) = uppX ∪ uppY

(16)

We show the 1st property. “⊆” holds by Proposition 2f. For “⊇” assume that u ∈ lowX ∩ lowY . Hence, u ∈ lowX and u ∈ lowY . By the deﬁnition of low, Γ u ⊆ X and Γ u ⊆ Y . Thus, Γ u ⊆ X ∩ Y . By the deﬁnition of low, u ∈ low(X ∩ Y ). Below, we collect several facts about compositions of the approximation mappings examined. Proposition 3. In cases (d)–(f ), assume rif 4 (κ). The following may be obtained: (a) low ◦ low uppS ◦ low uppS ◦ uppS (b) low ◦ low low ◦ uppS uppS ◦ uppS (c) upp ◦ low upp ◦ uppS & low ◦ upp uppS ◦ upp (d) upp ◦ low uppS ◦ low & low ◦ upp low ◦ uppS (e) upp ◦ upp uppS ◦ upp uppS ◦ uppS (f ) upp ◦ upp upp ◦ uppS uppS ◦ uppS Proof. We prove (a), (e) only. In case (a), low ◦ low uppS ◦ low by Proposition 2a and Proposition 1d. Next, uppS ◦ low uppS ◦ uppS by Proposition 2a, monotonicity of uppS , and Proposition 1e. In case (e), assume rif 4 (κ). By Proposition 2b, upp uppS . Hence, upp◦upp uppS ◦upp by Proposition 1d, whereas uppS ◦upp uppS ◦uppS by monotonicity of uppS and Proposition 1e.

Clearly, properties analogous to Proposition 2, Proposition 3, and (16) hold for the ∗-versions of the basic mappings5 . Now, we can formulate several properties of the -deﬁnable versions of low, lowS , upp, and uppS . Proposition 4. For any concepts X, Y , we can prove that: (a) low∪ = lowS∪ uppS∪ upp∪ id℘U (b) rif 4 (κ) ⇒ upp∪ uppS∪ (c) upp∪ ∅ = ∅ (d) X ⊆ Y ⇒ f ∪ X ⊆ f ∪ Y (e) f ∪ (X ∩ Y ) ⊆ f ∪ X ∩ f ∪ Y ⊆ f ∪ X ∪ f ∪ Y ⊆ f ∪ (X ∪ Y ) Proof. We prove (a) only. To this end, let X be any concept and u be any object. First, low∪ = lowS∪ directly by (12) and Proposition 2a. In virtue of 5

To safe space, we do not formulate them explicitly. When referring to them, we shall attach ∗ to the name of a property as a superscript. For instance, we shall refer to ∀X, Y.(X ⊆ Y ⇒ f ∗ X ⊆ f ∗ Y ), being the counterpart of Proposition 2e, as Proposition 2e∗ .

A Comparison of Pawlak’s and Skowron–Stepaniuk’s Approximation

73

Proposition 1g, it remains to show that (a1) low∪ uppS∪ , (a2) low∪ upp∪ , and (a3) low∪ id℘U . (a1) holds by Proposition 2a and Proposition 2e∗ for f = upp∗ . For (a2) it suﬃces to prove low∪ X ⊆ upp∪ X. To this end, assume u ∈ low∪ X. By (12), u ∈ upp∗ (lowX), i.e., there is u such that (a4) u ∈ Γ u and u ∈ lowX. Hence, Γ u ⊆ X by the deﬁnition of low. Note that Γ u ∩ X = ∅ since Γ u = ∅ in virtue of (a4). Hence, u ∈ uppX by the deﬁnition of upp. As a consequence, u ∈ upp∗ (uppX) = upp∪ X due to (a4) and (12). For (a3) we prove that low∪ X ⊆ X. To this end, let u ∈ low∪ X. By arguments as earlier,

there is u such that u ∈ Γ u ⊆ X. Immediately, u ∈ X. In the case of the -deﬁnable versions of approximation mappings, the Pawlak lower approximation and the Skowron–Stepaniuk lower approximation are equal. The novelty is that the lower approximation of a concept is included not only in the Skowron–Stepaniuk upper approximation of that concept but also in the concept itself and in the Pawlak upper approximation of that concept. Obviously, (b) is a counterpart of Proposition 2b, whereas (c) corresponds to the 1st part of Proposition 2c. The remaining two properties are counterparts of Proposition 2e and Proposition 2f, respectively. For example, (d) says that the mappings of lower and upper approximations are monotone. Additionally, we can derive the counterpart of (16) for the Pawlak upper approximation: upp∪ (X ∪ Y ) = upp∪ X ∪ upp∪ Y.

(17)

Several observations upon compositions of the -deﬁnable versions of the basic approximation mappings are presented below. Proposition 5. In cases (e)–(g), assume rif 4 (κ). The following can be derived: (a) low∪ ◦ low∪ = low∪ upp∪ ◦ low∪ low∪ ◦ upp∪ = upp∪ upp∪ ◦ upp∪ (b) low∪ ◦ low∪ uppS∪ ◦ low∪ low∪ ◦ uppS∪ = uppS∪ uppS∪ ◦ uppS∪ (c) (upp∪ ◦ low∪ ) uppS∪ upp∪ ◦ uppS∪ (d) (uppS∪ ◦ low∪ ) upp∪ uppS∪ ◦ upp∪ (e) upp∪ upp∪ ◦ uppS∪ & upp∪ ◦ low∪ uppS∪ ◦ low∪ (f ) upp∪ ◦ upp∪ uppS∪ ◦ upp∪ uppS∪ ◦ uppS∪ (g) upp∪ ◦ upp∪ upp∪ ◦ uppS∪ uppS∪ ◦ uppS∪ Proof. We prove (a), (b) only. In case (a), ﬁrst note that low∪ ◦ low∪ low∪ , low∪ ◦ low∪ upp∪ ◦ low∪ , low∪ ◦ upp∪ upp∪ ◦ upp∪ , and low∪ ◦ upp∪ upp∪ by Proposition 4a and Proposition 1d. Subsequently, upp∪ ◦ low∪ upp∪ holds by Proposition 4a, monotonicity of upp∪ , and Proposition 1e. Now, we prove that low∪ low∪ ◦ low∪ . To this end, it suﬃces to show low low ◦ low∪ . Then, by monotonicity of upp∗ and Proposition 1e, we obtain upp∗ ◦ low upp∗ ◦ (low ◦ low∪ ) = (upp∗ ◦ low) ◦ low∪ which ﬁnally results in low∪ low∪ ◦low∪ by (12). Thus, consider a concept X and an object u such that u ∈ lowX. Hence, (a1) Γ u ⊆ X by the deﬁnition of low. Then, for every u ∈ Γ u, u ∈ low∪ X by (a1) and the deﬁnition of low∪ . In other words, Γ u ⊆ low∪ X.

74

A. Gomoli´ nska

By the deﬁnition of low, u ∈ low(low∪ X) = (low ◦ low∪ )X. In this way, we have proved that lowX ⊆ (low ◦ low∪ )X. Hence, immediately, low low ◦ low∪ by the deﬁnition of . Finally, we show that upp∪ low∪ ◦ upp∪ . As in the preceding case, it suﬃces to prove upp low ◦ upp∪ and, then, to apply Proposition 2e∗ , Proposition 1e, and (12). Thus, consider a concept X and an object u such that u ∈ uppX. Immediately, (a2) Γ u ∩ X = ∅ by the deﬁnition of upp. As earlier, for every u ∈ Γ u, u ∈ upp∪ X by (a2) and the deﬁnition of upp∪ . As a consequence, Γ u ⊆ upp∪ X. Hence, u ∈ low(upp∪ X) = (low ◦ upp∪ )X by the deﬁnition of low. We have shown that uppX ⊆ (low ◦ upp∪ )X. Finally, upp low ◦ upp∪ by the deﬁnition of . In case (b), note that low∪ ◦ low∪ uppS∪ ◦ low∪ , low∪ ◦ uppS∪ uppS∪ ◦ uppS∪ , and low∪ ◦uppS∪ uppS∪ follow from Proposition 4a and Proposition 1d. Moreover, uppS∪ ◦ low∪ uppS∪ by Proposition 4a, monotonicity of uppS∪ , and Proposition 1e. It remains to show that uppS∪ low∪ ◦uppS∪ . It suﬃces to prove uppS low◦uppS∪ and, then, to apply Proposition 2e∗ , Proposition 1e, and (12). To this end, consider a concept X and an object u such that u ∈ uppS X. Hence, (b1) κ(Γ u, X) > 0 by the deﬁnition of uppS . Then, for every u ∈ Γ u, u ∈ uppS∪ X by (b1) and the deﬁnition of uppS∪ . Thus, Γ u ⊆ uppS∪ X. Hence, u ∈ low(uppS∪ X) = (low ◦ uppS∪ )X by the deﬁnition of low. That is, we have proved uppS X ⊆ (low ◦ uppS∪ )X. Immediately, uppS low ◦ uppS∪ by the deﬁnition of .

In comparison to Proposition 3, more and stronger relationships may be noted than in the basic case. It is due to Proposition 4a and to the fact that if an object u belongs to the -deﬁnable lower or upper approximation of a concept (in either of the senses considered), then its elementary granule Γ ∗ u is non-empty. By way of example, the composition of the lower approximation mapping with itself equals to the lower approximation mapping, whereas the compositions of the upper approximation mappings with the lower approximation mapping equal to the former mappings in virtue of (a), (b). Such results cannot be obtained for the basic approximation mappings without extra assumptions about .

5

Properties of Approximation Mappings II

In this section, we present and discuss properties of approximation mappings for special cases of approximation spaces. In detail, we consider approximation spaces which, in turn, are based on serial relations, reﬂexive relations, transitive relations, symmetric relations (and, in particular, tolerance relations), and — last but not least — equivalence relations6 . 6

A technical remark can be handy here. Except for a few cases, in a given subsection, we only present these properties which can be derived for the kind of approximation space investigated, yet were not obtained under weaker assumptions. For instance, when discussing approximation spaces based on reﬂexive relations, we do not recall the properties obtained for spaces based on serial relations and, the more, for arbitrary spaces.

A Comparison of Pawlak’s and Skowron–Stepaniuk’s Approximation

5.1

75

The Case of Serial Relations

In this section, we discuss two independent cases, viz., the case where −1 is serial and the case where is serial. Let SER(U ) denote the set of all serial relations on U . First, let −1 ∈ SER(U ). That is, for every u ∈ U , there exists u ∈ U such that (u , u) ∈ (in other words, ← {u} = ∅ or Γ u = ∅). Proposition 6. Let X be any concept. Then, we have: (a) uppU = U & low∅ = low∪ ∅ = ∅ (b) rif 3 (κ) ⇒ uppS ∅ = uppS∪ ∅ = ∅ (c) low upp (d) low ◦ low upp ◦ low upp ◦ upp (e) low ◦ low low ◦ upp upp ◦ upp (f ) uppS ◦ low uppS ◦ upp & low ◦ uppS upp ◦ uppS (g) rif 4∗ (κ) ⇒ uppS upp & uppS∪ upp∪ (h) rif 5 (κ) ⇒ uppS = upp & uppS∪ = upp∪ Proof. We prove (a), (b) only. Consider any object u. For (a) note that Γ u ∩ U = ∅ by seriality of −1 . Hence, u ∈ uppU by the deﬁnition of upp. Thus, U ⊆ uppU and, ﬁnally, uppU = U . Moreover, it can never be Γ u ⊆ ∅. Immediately, low∅ = ∅ by the deﬁnition of low. Hence, low∪ ∅ = (upp∗ ◦ low)∅ = upp∗ (low∅) = upp∗ ∅ = ∅ by (12) and Proposition 2c∗ . For (b) assume that κ satisﬁes rif 3 . By seriality of −1 , Γ u = ∅. Then, (b1) κ(Γ u, ∅) = 0 in virtue of rif 3 (κ). Hence, (b2) uppS ∅ = ∅ by the deﬁnition of uppS . Next, uppS∪ ∅ = (upp∗ ◦ uppS )∅ = upp∗ (uppS ∅) = upp∗ ∅ = ∅ by (b2), (12), and Proposition 2c∗ .

Let us brieﬂy comment upon the results. Seriality of −1 guarantees that elementary granules of the form Γ u are non-empty. Thanks to that, the Pawlak upper approximation of the universe is the universe itself and the lower approximations of the empty set are empty by (a). Moreover, in virtue of (b), if rif 3 is satisﬁed by κ, then the Skowron–Stepaniuk upper approximations of the empty set are empty. By (c), the lower approximation of a concept is included in the Pawlak upper approximation of that concept. The next three properties, being consequences of (c), augment Proposition 3 by new facts on compositions of approximation mappings. If rif 4∗ is satisﬁed by κ, then the Skowron–Stepaniuk upper approximations of a concept are included in the corresponding Pawlak upper approximations of that concept by (g). Moreover, if rif 5 (κ) holds, then the Pawlak upper approximations and the Skowron–Stepaniuk upper approximations coincide due to (h). We now consider the case, where ∈ SER(U ). Then, for every u ∈ U , there exists u ∈ U such that (u, u ) ∈ (in other words, → {u} = ∅ or Γ ∗ u = ∅). First, observe that properties analogous to Proposition 6 can be obtained for the ∗-versions of the basic approximation mappings. In particular, upp∗ U = U . Furthermore, we can prove the following properties.

76

A. Gomoli´ nska

Proposition 7. Let X be any concept. Then, it holds: (a) uppS∪ U = U & (−1 ∈ SER(U ) ⇒ upp∪ U = U ) (b) upp∪ X ∪ low∪ (U − X) = U (c) id℘U upp∪ & (rif 4 (κ) ⇒ id℘U uppS∪ ) (d) uppS∪ uppS∪ ◦ upp∪ Proof. We show (a), (b) only. To this end, consider any concept X. For the 1st part of (a) note that uppS∪ U = (upp∗ ◦ uppS )U = upp∗ (uppS U ) = upp∗ U = U by (12), seriality of , Proposition 2c, and Proposition 6a∗ . Now, assume additionally that −1 is serial. Hence, upp∪ U = (upp∗ ◦ upp)U = upp∗ (uppU ) = upp∗ U = U by (12), Proposition 6a, seriality of , and Proposition 6a∗ . In case (b), upp∪ X ∪ low∪ (U − X) = (upp∗ ◦ upp)X ∪ (upp∗ ◦ low)(U − X) = upp∗ (uppX) ∪ upp∗ (low(U − X)) = upp∗ (uppX ∪ low(U − X)) = upp∗ U = U

by (12), (16∗ ), Proposition 2d, seriality of , and Proposition 6a∗ . Most of the properties above strongly depend on Proposition 6a∗ (a counterpart of Proposition 6a) which is a consequence of seriality of . By (a), the Skowron– Stepaniuk upper approximation of the universe is the whole universe, whereas the Pawlak upper approximation of the universe is the universe if both , −1 are serial. According to (b), every object belongs to the Pawlak upper approximation of a concept and/or to the lower approximation of the complement of that concept. (c) states that every concept is included in its Pawlak upper approximation, and similarly for the Skowron–Stepaniuk upper approximation if κ satisﬁes rif 4 . Finally, the Skowron–Stepaniuk upper approximation of a concept is included in the Skowron–Stepaniuk upper approximation of the Pawlak upper approximation of that concept due to (d). 5.2

The Case of Reﬂexive Relations

Assume that is reﬂexive. Then, for every u ∈ U , (u, u) ∈ . Immediately, (u, u) ∈ −1 , so −1 is reﬂexive as well. Thus, u ∈ Γ u ∩ Γ ∗ u. Clearly, every reﬂexive relation is serial as well. In the context of approximation spaces, reﬂexive relations are referred to as similarity relations. The set of all reﬂexive relations on U will be denoted by RF(U ). Proposition 8. We can prove that: (a) low id℘U upp & (rif 4 (κ) ⇒ id℘U uppS ) (b) f f ∪ (c) low ◦ low low upp ◦ low upp upp ◦ upp (d) low low ◦ upp upp (e) uppS ◦ low low ◦ uppS uppS uppS ◦ upp upp ◦ uppS

A Comparison of Pawlak’s and Skowron–Stepaniuk’s Approximation

77

Proof. We prove (e) only. First, uppS ◦ low uppS , uppS uppS ◦ upp by (a), monotonicity of uppS , and Proposition 1e. Next, low ◦ uppS uppS , uppS upp ◦ uppS by (a) and Proposition 1d.

If is reﬂexive, then Proposition 6c may be strengthened to the property (a) above. In detail, the lower approximation of a concept is included in that concept and, on the other hand, every concept is included in its Pawlak upper approximation. The same holds for the Skowron–Stepaniuk approximation if rif 4 (κ) is satisﬁed. By (b), the lower (resp., upper) approximation of a concept is included in the -deﬁnable version of the lower (upper) approximation of that concept. Moreover, the list of properties of compositions of approximation mappings can be extended with several new dependencies given by (c)–(e). 5.3

The Case of Transitive Relations

Now, suppose that is transitive, i.e., for every u, u , u ∈ U , (u, u ) ∈ and (u , u ) ∈ imply (u, u ) ∈ . Either both , −1 are transitive or both of them are not transitive. We denote the set of all transitive relations on U by TR(U ). Proposition 9. The following dependencies can be proved: (a) low∪ low (b) low low ◦ low & upp ◦ upp upp (c) ∈ RF(U ) ⇒ low ◦ low = low = low∪ & upp ◦ upp = upp Proof. We prove (a), (b) only. Consider any concept X and any object u. In case (a), we show that low∪ X ⊆ lowX which results in low∪ low by the deﬁnition of . To this end, assume that u ∈ low∪ X. By the deﬁnition of low∪ , there is u such that u ∈ Γ u , i.e., (a1) (u, u ) ∈ , and (a2) u ∈ lowX. Next, (a3) Γ u ⊆ X by (a2) and the deﬁnition of low. Consider any u ∈ Γ u (i.e., (u , u) ∈ ). Hence, (u , u ) ∈ in virtue of (a1) and transitivity of . In other words, u ∈ Γ u . Hence, immediately, u ∈ X by (a3). Thus, Γ u ⊆ X. Hence, u ∈ lowX by the deﬁnition of low. For the 1st part of (b) it suﬃces to prove that lowX ⊆ (low ◦ low)X, i.e., lowX ⊆ low(lowX), and to apply the deﬁnition of . To this end, assume that u ∈ lowX. By the deﬁnition of low, (b1) Γ u ⊆ X. We need to prove that u ∈ low(lowX), i.e., Γ u ⊆ lowX in virtue of the deﬁnition of low. Thus, consider any u ∈ Γ u (i.e., (b2) (u , u) ∈ ). It remains to show that u ∈ lowX, i.e., Γ u ⊆ X. Let u ∈ Γ u , i.e., (u , u ) ∈ . Hence, (u , u) ∈ by (b2) and transitivity of . That is, u ∈ Γ u. In virtue of (b1), u ∈ X which ends the proof of this part of (b). For the remaining part of (b) it suﬃces to show that (upp ◦ upp)X ⊆ uppX, i.e., upp(uppX) ⊆ uppX, and to apply the deﬁnition of . To this end, let u ∈ upp(uppX). By the deﬁnition of upp, Γ u ∩ uppX = ∅. Hence, there is (b3) u ∈ Γ u such that u ∈ uppX. By the deﬁnition of upp, Γ u ∩ X = ∅. Hence, there is (b4) u ∈ Γ u such that (b5) u ∈ X. Note that (u , u) ∈ and (u , u ) ∈ by (b3) and (b4), respectively. Hence, by transitivity of ,

78

A. Gomoli´ nska

(u , u) ∈ as well. That is, u ∈ Γ u. As a consequence, Γ u ∩ X = ∅ by (b5). By the deﬁnition of upp, u ∈ uppX.

Some comments can be useful. Due to transitivity of , the -deﬁnable version of the lower approximation of a concept is included in the lower approximation of that concept by (a). Moreover, the lower approximation of a concept is included in the lower approximation of the lower approximation of that concept, whereas the Pawlak upper approximation of the same form of the upper approximation of a concept is included in the Pawlak upper approximation of that concept by (b). Unfortunately, a similar result does not seem to hold for the Skowron–Stepaniuk upper approximation. In the sequel, assuming reﬂexivity and transitivity of , the number of diﬀerent compositions of approximation mappings may substantially be reduced thanks to (c) since, then, both versions of the lower approximation are equal to the composition of the lower approximation with itself, and the Pawlak upper approximation is equal to the composition of the very form of the upper approximation with itself. 5.4

The Case of Symmetric Relations

In this section, we examine the case, where is symmetric. Then, for every u, u ∈ U , (u, u ) ∈ implies (u , u) ∈ . Immediately, = −1 and Γ = Γ ∗ . Relations which are both reﬂexive and symmetric are called tolerance relations. Obviously, every tolerance relation is also a similarity relation7 . The sets of all symmetric relations and all tolerance relations on U will be denoted by SYM(U ) and TL(U ), respectively. Proposition 10. The following properties hold: (a) f = f ∗ & f ∪ = upp ◦ f (b) upp ◦ low low ◦ upp (c) ∈ TR(U ) ⇒ upp ◦ low low & upp∪ upp low ◦ upp & upp∪ ◦ f ∪ f ∪ Proof. We show (b), (c) only. Consider any concept X and any object u. In case (b), assume that u ∈ upp(lowX). By the deﬁnition of upp, Γ u∩lowX = ∅. Hence, there is u such that (b1) u ∈ Γ u and u ∈ lowX. From the latter, (b2) Γ u ⊆ X by the deﬁnition of low. In virtue of (b1), (u , u) ∈ . Hence, by symmetry of , (u, u ) ∈ . That is, u ∈ Γ u . Hence, (b3) u ∈ X due to (b2). We need to show that u ∈ low(uppX), i.e., Γ u ⊆ uppX by the deﬁnition of low. To this end, consider u ∈ Γ u (i.e., (u , u) ∈ ). By symmetry of , (u, u ) ∈ as well, i.e., u ∈ Γ u . Hence, Γ u ∩ X = ∅ by (b3). By the deﬁnition of upp, u ∈ uppX as required. Thus, we have proved that upp(lowX) ⊆ low(uppX), i.e., (upp ◦ low)X ⊆ (low ◦ upp)X. Immediately, upp ◦ low low ◦ upp by the deﬁnition of . 7

Tolerance relations will not be a subject to a separate study in this article. To list their properties, it suﬃces to merge the proposition below with the facts presented in Sect. 5.2.

A Comparison of Pawlak’s and Skowron–Stepaniuk’s Approximation

79

In case (c), assume additionally that is transitive. First, upp ◦ low low by the assumption, the 2nd part of (a), and Proposition 9a. Next, upp∪ upp by the assumption, the 2nd part of (a), and Proposition 9b. In the sequel, we show that uppX ⊆ (low ◦ upp)X, i.e., uppX ⊆ low(uppX). Then, upp low ◦ upp by the deﬁnition of . To this end, suppose that u ∈ uppX. By the deﬁnition of upp, Γ u ∩ X = ∅. Hence, there is u such that (c1) u ∈ X and u ∈ Γ u (i.e., (c2) (u , u) ∈ ). Consider any u ∈ Γ u (i.e., (u , u) ∈ ). By symmetry of , (u, u ) ∈ . Hence, (u , u ) ∈ by (c2) and transitivity of . In other words, u ∈ Γ u . Hence, Γ u ∩ X = ∅ in virtue of (c1). By the deﬁnition of upp, u ∈ uppX. Thus, we have proved that Γ u ⊆ uppX, i.e., u ∈ low(uppX) by the deﬁnition of low. For the remaining part of (c) note that upp∪ ◦ f ∪ = (upp ◦ upp) ◦ (upp ◦ f ) upp ◦ (upp ◦ f ) = (upp ◦ upp) ◦ f upp ◦ f = f ∪ by (a), transitivity of , Proposition 9b, and Proposition 1d.

Thus, whenever symmetry of is assumed, the ∗-versions of the basic approximation mappings coincide with the very mappings in virtue of (a). As a consequence, the -deﬁnable versions of the basic mappings are compositions of these mappings with the Pawlak upper approximation mapping. Due to (b), the Pawlak upper approximation of the lower approximation of a concept is included in the lower approximation of the Pawlak upper approximation of that concept. In virtue of (c), if is both symmetric and transitive, then — among others — the Pawlak upper approximation of the lower approximation of a concept is included in the lower approximation of that concept, the -deﬁnable version of the Pawlak upper approximation of a concept is included in the Pawlak upper approximation of that concept, and the latter one is included in the lower approximation of the Pawlak upper approximation of the very concept. 5.5

The Case of Equivalence Relations

Finally, we consider the case, where is an equivalence relation. In the context of Pawlak’s information systems and approximation spaces, equivalence relations on the set of objects are understood as indiscernibility relations. By deﬁnition, an equivalence relation is simultaneously reﬂexive, symmetric, and transitive, i.e., it is a transitive tolerance relation. Let EQ(U ) denote the set of all equivalence relations on U . Thus, EQ(U ) = TL(U ) ∩ TR(U ). Note that EQ(U ) = SER(U ) ∩ SYM(U ) ∩ TR(U ) as well. In the sequel, Γ u is called an equivalence class of u and it may be denoted by [u] (or simply, by [u] if is understood) along the standard lines. Note that ∀u, u ∈ U.(u ∈ Γ u ⇔ Γ u = Γ u). Proposition 11. It holds that: (a) f ∪ = f (b) low = low ◦ low = upp ◦ low low ◦ upp = upp ◦ upp = upp (c) low uppS ◦ low low ◦ uppS = uppS = upp ◦ uppS uppS ◦ upp

(18)

80

A. Gomoli´ nska

Proof. We show (a) only. Note that “” holds by reﬂexivity of and Proposition 8b. In the sequel, low∪ low by transitivity of and Proposition 9a. Moreover, upp∪ upp by symmetry and transitivity of , and Proposition 10c. Now, consider any concept X and any object u. To prove uppS∪ uppS , it sufﬁces to show that uppS∪ X ⊆ uppS X. Thus, suppose that u ∈ uppS∪ X. Hence, there is u such that (a1) u ∈ Γ u and u ∈ uppS X by the deﬁnition of uppS∪ . It follows from the latter that κ(Γ u , X) > 0 by the deﬁnition of uppS . By (18) and (a1), κ(Γ u, X) > 0. Finally, u ∈ uppS X by the deﬁnition of uppS .

Thus, whenever is an equivalence relation, the -deﬁnable versions of the basic approximation mappings coincide with the mappings by (a), respectively. As regarding the reduction in the number of diﬀerent compositions of the mappings considered, both the composition of the lower approximation mapping with itself and the composition of the lower approximation mapping with the Pawlak upper approximation mapping are equal (and hence, may be reduced) to the lower approximation mapping. Furthermore, both the composition of the Pawlak upper approximation mapping with the lower approximation mapping and the composition of the Pawlak upper approximation mapping with itself are equal to the Pawlak upper approximation mapping thanks to (b). Moreover by (c), both the composition of the Skowron–Stepaniuk upper approximation mapping with the lower approximation mapping and the composition of the Skowron–Stepaniuk upper approximation mapping with the Pawlak upper approximation mapping are equal to the Skowron–Stepaniuk upper approximation mapping. Finally, the lower approximation of a concept is included in the Skowron–Stepaniuk upper approximation of the lower approximation of that concept, the latter set is included in the Skowron–Stepaniuk upper approximation of the concept, and the very upper approximation of a concept is included in the Skowron–Stepaniuk upper approximation of the Pawlak upper approximation of that concept. Last but not least, observe that the upper approximation mappings uppS1 , uppS2 investigated in Example 1 are equal to upp if is an equivalence relation. Indeed, upp ◦ Γ = Γ then, and for any object u and any concept X, u ∈ uppS2 X if and only if Γ u ∩ upp∗ X = ∅ if and only if Γ u ∩ uppX = ∅ if and only if u ∈ upp(uppX) if and only if u ∈ uppX.

6

Summary

In this article, we studied and compared Pawlak’s rough approximation of concepts with Skowron–Stepaniuk’s approach within a general framework of approximation spaces of the form (U, , κ), where U is a non-empty set of objects, is a non-empty binary relation on U , and κ is a RIF satisfying rif 1 and rif 2 . The lower approximation mappings are the same in both approaches unlike the upper approximation ones8 . The latter mappings cannot be compared directly without additional assumptions made about κ. For the sake of illustration, we 8

The fact that Pawlak’s and Skowron–Stepaniuk’s upper approximations coincide for the standard RIF was known earlier.

A Comparison of Pawlak’s and Skowron–Stepaniuk’s Approximation

81

considered two special cases of κ and two corresponding Skowron–Stepaniuk upper approximation mappings in Example 1. In general, these two kinds of mappings are diﬀerent from each other and from Pawlak’s upper approximation mapping. However — as turned out — the three cases coincide if is an equivalence relation. In the paper presented, we compared Pawlak’s and Skowron–Stepaniuk’s upper approximation mappings indirectly by investigation of their properties. It is diﬃcult to say which mappings are totally better. While Pawlak’s approximation seems to be more suitable in some aspects, it is the Skowron–Stepaniuk approach which seems to provide more interesting results in other cases. As a side-eﬀect, we have obtained a fairly exhaustive list of basic mathematical properties of the mappings investigated, to be used in the future research and applications.

References 1. Pawlak, Z.: Rough set elements. In Polkowski, L., Skowron, A., eds.: Rough Sets in Knowledge Discovery 1. Volume 18 of Studies in Fuzziness and Soft Computing. Physica-Verlag, Heidelberg (1998) 10–30 2. Pawlak, Z.: Information systems – Theoretical foundations. Information Systems 6 (1981) 205–218 3. Pawlak, Z.: Rough sets. Computer and Information Sciences 11 (1982) 341–356 4. Pawlak, Z.: Information Systems. Theoretical Foundations (in Polish). Wydawnictwo Naukowo-Techniczne, Warsaw (1983) 5. Pawlak, Z.: Rough Sets. Theoretical Aspects of Reasoning About Data. Kluwer, Dordrecht (1991) 6. Pomykala, J.A.: Approximation operations in approximation space. Bull. Polish Acad. Sci. Math. 35 (1987) 653–662 7. Wybraniec-Skardowska, U.: On a generalization of approximation space. Bull. Polish Acad. Sci. Math. 37 (1989) 51–62 ˙ 8. Zakowski, W.: Approximations in the space (U, Π). Demonstratio Mathematica 16 (1983) 761–769 9. Skowron, A., Stepaniuk, J.: Generalized approximation spaces. In: Proc. 3rd Int. Workshop Rough Sets and Soft Computing, San Jose, USA, 1994, November 10-12. (1994) 156–163 10. Skowron, A., Stepaniuk, J.: Tolerance approximation spaces. Fundamenta Informaticae 27 (1996) 245–253 11. Gomoli´ nska, A.: Variable-precision compatibility spaces. Electronical Notices in Theoretical Computer Science 82 (2003) 1–12 http://www.elsevier.nl/locate/entcs/volume82.html. 12. Slowi´ nski, R., Vanderpooten, D.: Similarity relation as a basis for rough approximations. In Wang, P.P., ed.: Advances in Machine Intelligence and Soft Computing. Volume 4. Duke University Press (1997) 17–33 13. Yao, Y.Y., Wong, S.K.M., Lin, T.Y.: A review of rough set models. In Lin, T.Y., Cercone, N., eds.: Rough Sets and Data Mining: Analysis of Imprecise Data. Kluwer, Boston London Dordrecht (1997) 47–75 14. Ziarko, W.: Variable precision rough set model. J. Computer and System Sciences 46 (1993) 39–59 15. Ziarko, W.: Probabilistic decision tables in the variable precision rough set model. J. Comput. Intelligence 17 (2001) 593–603

82

A. Gomoli´ nska

16. Peters, J.F.: Approximation space for intelligent system design patterns. Engineering Applications of Artiﬁcial Intelligence 17 (2004) 1–8 17. Skowron, A.: Approximation spaces in rough neurocomputing. In Inuiguchi, M., Hirano, S., Tsumoto, S., eds.: Rough Set Theory and Granular Computing. Volume 125 of Studies in Fuzziness and Soft Computing. Springer-Verlag, Berlin Heidelberg (2003) 13–22 18. Skowron, A., Swiniarski, R., Synak, P.: Approximation spaces and information granulation. Transactions on Rough Sets III: Journal Subline to Lecture Notes in Computer Science 3400 (2005) 175–189 19. Polkowski, L., Skowron, A.: Rough mereology. Lecture Notes in Artiﬁcial Intelligence 869 (1994) 85–94 20. Polkowski, L., Skowron, A.: Rough mereology: A new paradigm for approximate reasoning. Int. J. Approximated Reasoning 15 (1996) 333–365 21. L ukasiewicz, J.: Die logischen Grundlagen der Wahrscheinlichkeitsrechnung. In Borkowski, L., ed.: Jan L ukasiewicz – Selected Works. North Holland, Polish Scientiﬁc Publ., Amsterdam London, Warsaw (1970) 16–63 First published in Krak´ ow in 1913. 22. Zadeh, L.A.: Outline of a new approach to the analysis of complex system and decision processes. IEEE Trans. on Systems, Man, and Cybernetics 3 (1973) 28–44

Data Preparation for Data Mining in Medical Data Sets Grzegorz Ilczuk1 and Alicja Wakulicz-Deja2 1

Siemens AG Medical Solutions, Allee am Roethelheimpark 2, 91052 Erlangen, Germany [email protected] 2 Institut of Informatics University of Silesia, Bedzinska 39, 41-200 Sosnowiec, Poland [email protected]

Abstract. Data preparation is a very important but also a time consuming part of a Data Mining process. In this paper we describe a hierarchical method of text classiﬁcation based on regular expressions. We use the presented method in our data mining system during a pre-processing stage to transform Latin free-text medical reports into a decision table. Such decision tables are used as an input for rough sets based rule induction subsystem. In this study we also compare accuracy and scalability of our method with a standard approach based on dictionary phrases. Keywords: rough sets, data preparation, regular expression.

1

Introduction

Preparation of data takes about 60% of a time needed for the whole Data Mining process and it is also deﬁned by Pyle as the most important part of a Data Exploration Process which leads to success [1]. This estimation is also valid in case of our Data Exploration system, where the entry stage of data processing is a key element for the further analysis. The mentioned Data Exploration system will be used in medicine (especially in cardiology) as a complete solution suitable for improving medical care and clinical work ﬂow through revealing new patterns and relations among data. Functional blocks of the system are: – Import subsystem-responsible for importing data from medical information systems into our storage subsystem – Data recognition subsystem-during this stage we use algorithms and methods described in this paper to transform the raw data to a form suitable for further Data Exploration – Data preprocessing-based on the statistical analysis of the transformed information noise and redundant data are removed [9] – Feature selection-this stage utilizes a few attribute reduction methods such as CFS (Correlation-based Feature Selection), Quickreduct and conjunction of these methods to select an optimal set of attributes for a further analysis J.F. Peters et al. (Eds.): Transactions on Rough Sets VI, LNCS 4374, pp. 83–93, 2007. c Springer-Verlag Berlin Heidelberg 2007

84

G. Ilczuk and A. Wakulicz-Deja

– Rule induction subsystem based on Rough Set Theory [6,7,11]. Early research on this area was described in [8,12] – Visualization of the knowledge discovery in a form easily understandable by humans for validating and extending of the collected knowledge [13] Most medical information useful for Data Mining is still written in form of freetext Latin reports. These reports are mostly used to extend a lapidary diagnosis written with statistical ICD-10 codes. There are some challenges to solve during analyzing such reports such as: diﬀerent descriptions for the same disease, non-standard abbreviations, misspelled words and a ﬂoating structure of such reports. In our ﬁrst solution of these problems we had used a phrase dictionary to map information from a report to an attribute. The main disadvantage of this approach was a lack of scalability and a diﬃcult maintenance. These facts leaded us to develop a diﬀerent approach. The proposed method and results achieved with it are presented in this study. Described in this paper technique is used in our Data Exploration system as a preprocessing step, which prepares data for rule induction. For generation of decision rules an own implementation of MLEM2 algorithm is used.

2

Rough Sets: Basic Notions and Medical Appliance

Developed by Pawlak and presented in 1982 Rough Sets theory is a mathematical approach to handle imprecision and uncertainty [4]. The main goal of rough set analysis is to synthesize approximation of concepts from the acquired data. Some basic deﬁnitions are presented below. Information system [4] is a pair A = (U, A) where U is a non-empty, ﬁnite set called the universe and A is a non-empty, ﬁnite set of attributes, i.e. a : U → Va for a ∈ A, where Va is called the value set of attribute a. Elements of U are called objects. The special case of information systems called decision system is deﬁned as A = (U, A ∪ {d}), where d ∈ A is a distinguished attribute called decision and elements of A are called conditions. A decision rule is deﬁned as r = (ai1 = v1 ) ∧ . . . ∧ (aim = vm ) ⇒ (d = k) where 1 ≤ i1 < . . . < im ≤ |A|, vi ∈ Vai . We say an object matches a rule if its attributes satisfy all atomic formulas (aij = vj ) of the rule. A rule is called minimal consistent with A when any decision rule r created from r by removing one of atomic formula of r is not consistent with A. In our Data Exploration system we use a modiﬁed version of LEM2 algoritm - MLEM2 to generate decision rules. LEM2 (Learning from Examples Module, version 2) algorithm was ﬁrstly presented in [14,15] and then implemented in [16]. LEM2 induces a rule set by exploring the space of blocks of attributevalue pairs to generate a local covering. Afterwards the found local covering is converted into the rule set. Following deﬁnitions must be quoted prior to deﬁne a local covering [11]. For a variable (attribute or decision) x and its value v, a block [(x, v)] of a variable-value pair (x, v) is the set of all cases for which variable x has value v.

Data Preparation for Data Mining in Medical Data Sets

85

Let B be a nonempty lower or upper approximation of a concept represented by a decision-value pair (d, w). Set B depends on a set T of attribute-value pairs (a, v) if and only if ∅ = [T ] =

[(a, v)] ⊆ B.

(1)

(a,v)∈T

Set T is a minimal complex of B if and only if B depends on T and no proper subset T , of T exists such that B depends on T , . Let T be a nonempty collection of nonempty sets of attribute-value pairs. Then T is a local covering of B if and only if the following conditions are satisﬁed: – each member T of T is a minimal complex of B, – T ∈T [T ] = B, and – T is minimal, i.e., T has the smallest possible number of members. Modiﬁed LEM2 (MLEM2) proposed by Grzymala-Busse in [11] in compare to LEM2 allows inducing rules from data containing numerical attributes without a need of a separate discretization step. Our implementation of MLEM2 algorithm induces decision rules from both lower approximation (certain rules) and upper approximation (possible rules). This technique allows us reasoning from ”real” data, which contains uncertain, noisy and redundant information. Decision rules are used in our system to present the extracted knowledge from medical data. This approach in medical domain has several advantages over other data mining techniques: – – – – –

Decision rules are easy to understand and verify Decision rules can be easily validated with existing knowledge Gathered decision rules can be modiﬁed and extended with a new knowledge If decision rules are used for classiﬁcation it is easy to explain the choice Simple structure of decision rules allows several ways of visualization

These advantages lead to a rapid growth of interest in appliance of rough set theory in medical domain. Many interesting case studies reported a successful appliance of rough set software systems. Some of them were: – Treatment of duodental ulcer by HSV described by Slowinski in [21,23] – Multistage analysis in progressive encephalopathy presented by Paszek in [20,22] – Preterm birth prediction researched by Grzymala-Busse [19] – Analysis of medical databases (headaches, CVD) [24] – Acute abdominal pain in childhood (MET system applied in Children’s Hospital of Eastern Ontario) [18] – Cardiac Tests analysis [25] More successful studies, not only from medical domain, are described in [17].

86

3

G. Ilczuk and A. Wakulicz-Deja

Regular Expression

The origins of, belonging to automata and formal language theory, regular expressions lie in 1940, when McCulloch and Pitts described a nervous system as a neurons in a small automata [3]. These models were then described by Kleene and Kozen using regular expression (regular set) notation [2]. Regular expressions consist of constants and operators that denote sets of strings and operations over these sets, respectively. Given a ﬁnite alphabet Σ the following constants are deﬁned: – empty set ∅ denoting the set ∅ – empty string denoting the set {} – literal character α in Σ denoting the set {α} Following operations are deﬁned: – concatenation RS denoting the set {αβ|α in R and β in S}. For example {”ab”, ”c”}{”d”, ”ef”} = {”abd”, ”abef”, ”cd”, ”cef”}. – alternation R|S denoting the set union of R and S. – Kleene star R∗ denoting the smallest superset of R that contains and is closed under string concatenation. This is the set of all strings that can be made by concatenating zero or more strings in R. For example, {”ab”,”c”}∗ = {, ”ab”, ”c”, ”abab”, ”abc”, ”cab”, ”cc”, ”ababab”, ...}. To avoid brackets it is assumed that the Kleene star has the highest priority, then concatenation and then set union. If there is no ambiguity then brackets may be omitted. For example, (ab)c is written as abc and a|(b(c∗)) can be written as a|bc∗.

4

Methods

In our research of analyzing medical data we would like to extend and complement information collected from clinical information systems in form of ICD-10 codes with additional information stored in free-text descriptions. A typical example of a such description is shown below: Status post implantationem pacemakeri VVI (1981, 1997) ppt. diss. A-V gr. III.Exhaustio pacemakeri. Morbus ischaemicus cordis. Insufficientia coronaria chronica CCS I. Myocardiopathia ischaemica in stadio comp. circulatoriae. Fibrillatio atriorum continua. Pacemaker dependent. A method suitable for our needs should therefore fulﬁll the following requirements: – it shall recognize misspelled and abbreviated words – it shall interpretable whole sentences

Data Preparation for Data Mining in Medical Data Sets

87

– it shall provide a back tracing so that an expert can always validate an assigned mapping – all mappings must done based 100% on information from an original text – it shall be easily maintainable and extendible With these requirements in mind we have developed a method which bases on a ﬁxed number of user deﬁned records each containing following three attributes: a level value (shown at ﬁgure 1 as ’LEVEL’), a mask coded using regular expressions for searching a phrase of text (’FIND TEXT’) and a string of text which will be used for replacing if the searched phrase is found (’REPLACE FOUND TEXT’). Deﬁned records are sorted incrementally based on their level value, so that, when the algorithm starts a group of records having the lowest value can be ﬁrstly selected. During the next step for each record from the group the algorithm tries to replace a found text with a speciﬁed string. When all records are processed then a next group of records with a next higher level value is selected and the process of searching/replacing text repeats. This algorithm ends when the last record from the group with the highest level value is processed. A simpliﬁed, but based on a real implementation, example is shown at ﬁgure 1.

Fig. 1. Sample of records used by the algorithm

From this ﬁgure it can be seen that the lowest level value is 10, so that the algorithm begins to select a group of records having this level value. In our case it is only one record which replace all found Roman numbers with the following schema ’’ for example a number ’II’ will be replaced to and

88

G. Ilczuk and A. Wakulicz-Deja

’IV’ results in . After this processing a next higher level value is selected (50) together with a group of records having the same value of their level attribute. In the shown example there are only one such record. But it is possible and common that, there are a lot of records which have the same level value and thus can be processed in parallel. It is important to note that a deﬁnition of a mask used for searching (ﬁeld ’FIND TEXT’) contains not only a correct version of a phrase but also several misspelled combinations stored using the alternation operation of regular expressions for example (pectoralis|pectoris|...). If a phase is found then it will be replaced with a speciﬁed replace string independently if it was written correctly or incorrectly. This allows to correct a simple type errors and focus on sentence analysis. As an example records with their level value 100 and 110 can be used. These two records search for a combination of symbolic replacements previously replaced by records with the level value 50, so that these two records can correctly assign an ’<END I20 0> code not only to a properly written ’angina pectoris’ diagnose but also to a whole bunch of misspelled combinations of these two words as for example ’angin pectoralis’. The described algorithm has following advantages: – it allows ﬁltering of redundant and noisy information at entry processing stage – it correctly recognizes misspelled diagnoses – the process of interpreting whole sentences is simpliﬁed because only connections between symbolic phrases must be analyzed and not all possible combinations which can be found in an input text – it is possible to stop the algorithm at stage and analyze or eventually correct the replacing process – in our implementation we only use already found in input text combinations of words what decreases a possibility of false positive recognitions In the next section we will compare the recognition accuracy and scalability of the described algorithm with our previous dictionary based algorithm.

5

Dataset Preparation and Experimental Environment

Data used in our research was obtained from the Cardiology Department of Silesian Medical Academy in Katowice - the leading Electrocardiology Department in Poland specializing in hospitalization of severe heart diseases. For our experiments we took a data set of 4000 patients hospitalized in this Department between 2003 and 2005. This data were imported into a PostgreSQL database and then divided in eight groups (G-C1, ..., G-C8), where G-C1 contained ﬁrst 500 records from the database and each next group had 500 more records then the previous group, so that the last group G-C8 contained all 4000 records. Each record in a group contained a single free-text report which was then analyzed of the presence of one of the following diseases: – Essential (primary) hypertension - if found mapped to I10 code – Old myocardial infarction - if found mapped to I25-2 code

Data Preparation for Data Mining in Medical Data Sets

– – – –

89

Atrioventricular block, ﬁrst degree - if found mapped to I44-0 code Atrioventricular block, second degree - if found mapped to I44-1 code Atrioventricular block, complete - if found mapped to I44-2 code Sick sinus syndrome - if found mapped to I49-5 code

We implemented the presented algorithm in Java version 1.5. and used the Java implementation of regular expressions from the ’java.util.regex.Pattern’ class.

6

Results

Results presented the table 1 show an absolute number of cases recognized by the described in this paper method within each of the tested group. These results are additionally compared with the dictionary method and this comparison is shown as a number in brackets, where a positive number means a number of cases additionally recognized by the method based on regular expressions. Visualization of these numbers is shown at ﬁgure 2, where it can be seen that the proposed method recognized more cases then the dictionary method but with a diﬀerent, depending on a selected disease, characteristic. For hypertension and old myocardial infarction a number of additionally recognized cases is rather low what can be attributable to the fact, that the most diagnosis variants are already covered by the dictionary method. Recognition of atrioventricular block poses a bigger challenge, so that a diﬀerence in a number of recognized cases for all three types of this disease oscillates between 20-40% additional cases identiﬁed by the proposed method. The most spectacular results were achieved for recognizing Sick sinus syndrome what can be assignable with a huge number of possible combinations used to specify this diagnosis. These combinations were better covered by regular expressions and a diﬀerence to the dictionary method was almost 42%. It can be also seen, that a number of identiﬁed cases, shown at ﬁgure 3, increased for all tested diseases almost linearly. This satisfactory result shows a good ability of the presented method to recognize new records with a relatively small number of deﬁnitions (500 regular expressions compared to more then 4800 dictionary phrases). We had also randomly selected a set of 100 records and with a help from domain experts from the Cardiology Department manually identiﬁed them for three diseases. These numbers were then compared with a results achieved by both the regular expression and the dictionary method. This comparison is shown in the table 2. From this table it can be seen that for a relatively small group of records the method based on regular expression recognized all hypertension and atrioventricular block (ﬁrst degree) cases. Of course it will be only a matter of additional time eﬀort needed to extend the recognition accuracy of the dictionary method but this is exactly the advantage of the proposed algorithm, which with a signiﬁcant smaller number of records presents better scalability and in case of new data also a better update ability.

90

G. Ilczuk and A. Wakulicz-Deja

Number of records additionaly recognized by the new model

800 I10 I25-2 I44-0 I44-1 I44-2 I49-5

700

600

500

400

300

200

100

0

-100 500

1000

1500

2000

2500 3000 Number of records

3500

4000

4500

Fig. 2. Additionally recognized cases by the method based on regular expressions

3000

Number of recognized cases

2500

I10 I25-2 I44-0 I44-1 I44-2 I49-5

2000

1500

1000

500

0 500

1000

1500

2000

2500 3000 Number of records

3500

4000

Fig. 3. Number of recognized diseases by the regular expression method

4500

Data Preparation for Data Mining in Medical Data Sets

91

Table 1. Number of recognized cases by the proposed method Group Number of I10 records

I25-2

I44-0

I44-1

I44-2

I49-5

G-C1 G-C2 G-C3 G-C4 G-C5 G-C6 G-C7 G-C8

91 (-15) 190 (-13) 276 (3) 368 (25) 442 (30) 524 (44) 600 (40) 704 (43)

45 (0) 95 (12) 134 (36) 183 (61) 236 (96) 305 (120) 373 (125) 427 (131)

82 (3) 161 (31) 248 (99) 329 (158) 410 (220) 502 (258) 590 (267) 678 (276)

135 (-4) 245 (17) 354 (68) 483 (131) 604 (192) 728 (238) 863 (237) 1007 (233)

220 (72) 414 (150) 618 (241) 824 (338) 1029 (441) 1265 (550) 1493 (640) 1693 (714)

500 1000 1500 2000 2500 3000 3500 4000

325 (1) 636 (3) 978 (12) 1314 (21) 1645 (30) 1959 (38) 2275 (41) 2616 (41)

Table 2. Recognition accuracy comparison between methods

7

Disease Number of Regular exp. cases found cases

Regular exp. accuracy [%]

Dictionary Dictionary found cases

accuracy [%]

I10 I44-0 I44-1

100.0 100.0 93.3

61 9 10

100.0 81.9 66.7

61 11 15

61 11 14

Conclusions

In this paper we presented an algorithm for recognition of free-text Latin medical reports which is based on hierarchically organized records. These records use regular expressions to ﬁnd a speciﬁed phrase in an input text and replace it with a user deﬁned text. The hierarchically organized records convert an input text step by step replacing ﬁrstly simple words into symbolic phrases then these symbolic phrases into more complicated expressions and at the end the whole sentences are mapped to user deﬁned codes. Such codes can be then easily used to construct a decision table used by next data mining algorithms. Our experiments shown that the presented method achieves better recognition accuracy then the method based on ﬁxed dictionary phrases and this result can be achieved with a signiﬁcant smaller number of records used for deﬁnition. This small number of easily modiﬁable and very ﬂexible records is truly an advantage of the described method. Our idea to reduce the complexity of recognizing Latin diagnosis through deﬁning a short parts of the whole sentence using regular expressions and then to join hierarchically such pieces of information together allowed us to cover with a ﬁnite, small number of records a huge number of possible combinations. This advantage and the fact that the presented method fulﬁll all the speciﬁed

92

G. Ilczuk and A. Wakulicz-Deja

requirements it is used in our data exploration system during a preprocessing stage for processing not only Latin free-text reports but also laboratory, electrocardiogram (ECG) and cardiovascular ultrasound descriptions.

Acknowledgements We would like to thank Rafal Mlynarski from the Cardiology Department of Silesian Medical Academy in Katowice, Poland for providing us the data and giving us feedbacks.

References 1. Pyle, D.: Data preparation for data mining. Morgan Kaufmann, San Francisco (1999) 2. Kozen, D.: On Kleene Algebras and Closed Semirings. In: Mathematical Foundations of Computer Science, Bansk´ a Bystrica (1990) 26–47 3. McCulloch, W. and Pitts, W.: A logical calculus of the ideas immanent in nervous activity. In: Bulletin of Mathematical Biophysics, (1943) 115–133 4. Pawlak, Z.: Rough sets. International Journal of Computer and Information Science 11 (1982) 341–356 5. Sipser, M.: Introduction to the Theory of Computation. Course Technology, (2006) 6. Pawlak, Z.: Knowledge and Uncertainty: A Rough Set Approach. SOFTEKS Workshop on Incompleteness and Uncertainty in Information Systems (1993) 34–42 7. Pawlak, Z. and Grzymala-Busse, J. W. and Slowinski, R. and Ziarko, W.: Rough Sets. Commun. ACM 38 (1995) 88–95 8. Ilczuk, G. and Wakulicz-Deja, A.: Rough Sets Approach to Medical Diagnosis System. In: AWIC 2005, Lodz (2005) 204–210 9. Ilczuk, G. and Wakulicz-Deja, A.: Attribute Selection and Rule Generation Techniques for Medical Diagnosis Systems. In: RSFDGrC 2005, Regina (2005) 352–361 10. Wakulicz-Deja, A. and Paszek, P.: Applying Rough Set Theory to Multi Stage Medical Diagnosing. Fundam. Inform. 54 (2003) 387–408 11. Grzymala-Busse, J. W.: MLEM2 - Discretization During Rule Induction. In: IIS 2003, Zakopane (2003) 499–508 12. Ilczuk, G. and Mlynarski, R. and Wakulicz-Deja, A. and Drzewiecka, A. and Kargul, W.: Rough Sets Techniques for Medical Diagnosis Systems. In: Computers in Cardiology 2005, Lyon (2005) 837–840 13. Mlynarski, R. and Ilczuk, G. and Wakulicz-Deja, A. and Kargul, W.: Automated Decision Support and Guideline Veriﬁcation in Clinical Practice. In: Computers in Cardiology 2005, Lyon (2005) 375–378 14. Chan, C. C. and Grzymala-Busse, J. W.: On the two local inductive algorithms: PRISM and LEM2. Foundations of Computing and Decision Sciences 19 (1994) 185–203 15. Chan, C. C. and Grzymala-Busse, J. W.: On the attribute redundancy and the learning programs ID3, PRISM, and LEM2.Department of Computer Science, University of Kansas,TR-91-14, (1991) 16. Grzymala-Busse, J. W.: A new version of the rule induction system LERS. Fundam. Inform. 31 (1997) 27–39

Data Preparation for Data Mining in Medical Data Sets

93

17. Komorowski, H. J. and Pawlak, Z. and Polkowski, L. T. and Skowron, A.: Rough Sets: A Tutorial. Springer-Verlag, Singapore (1999) 18. Farion, K. and Michalowski, W. and Slowinski, R. and Wilk, S. and Rubin, S.: Rough Set Methodology in Clinical Practice: Controlled Hospital Trial of the MET System. Rough Sets and Current Trends in Computing. 3066 (2004) 805–814 19. Grzymala-Busse, J. W. and Goodwin, L. K.: Predicting pre-term birth risk using machine learning from data with missing values. Bulletin of the International Rough Set Society (IRSS). 1 (1997) 17–21 20. Paszek, P. and Wakulicz-Deja, A.: The Application of Support Diagnose in Mitochondrial Encephalomyopathies. Rough Sets and Current Trends in Computing. 2475 (2002) 586–593 21. Pawlak, Z. and Slowinski, K. and Slowinski, R.: Rough Classiﬁcation of Patients After Highly Selective Vagotomy for Duodenal Ulcer. International Journal of ManMachine Studies. 24 (1986) 413–433 22. Tsumoto, S. and Wakulicz-Deja, A. and Boryczka, M. and Paszek, P.: Discretization of continuous attributes on decision system in mitochondrial encephalomyopathies. Proceedings of the First International Conference on Rough Sets and Current Trends in Computing. 1424 (1998) 483–490 23. Slowinski, K. and Slowinski, R. and Stefanowski, J.: Rough sets approach to analysis of data from peritoneal lavage in acute pancreatitis. Medical Informatics. 13 (1988) 143–159 24. Tsumoto, S. and Tanaka, H.: Induction of Disease Description based on Rough Sets. 1st Online Workshop on Soft Computing. (1996) 19–30 25. Komorowski, H.J. and Øhrn, A.: Modelling prognostic power of cardiac tests using rough sets. Artiﬁcial Intelligence in Medicine. 15 (1999) 167–191

A Wistech Paradigm for Intelligent Systems Andrzej Jankowski1,2 and Andrzej Skowron3 1

Institute of Decision Processes Support 2 AdgaM Solutions Sp. z o.o. W¸awozowa 9 lok. 64, 02-796 Warsaw, Poland [email protected] 3 Institute of Mathematics, Warsaw University Banacha 2, 02-097 Warsaw, Poland [email protected] If controversies were to arise, there would be no more need of disputation between two philosophers than between two accountants. For it would suﬃce to take their pencils in their hands, and say to each other: ‘Let us calculate’. – Gottfried Wilhelm Leibniz, Dissertio de Arte Combinatoria (Leipzig, 1666). ... Languages are the best mirror of the human mind, and that a precise analysis of the signiﬁcation of words would tell us more than anything else about the operations of the understanding. – Gottfried Wilhelm Leibniz, New Essays on Human Understanding (1705) Translated and edited by Peter Remnant and Jonathan Bennett Cambridge: Cambridge UP, 1982 Abstract. The problem considered in this article is how does one go about discovering and designing intelligent systems. The solution to this problem is considered in the context of what is known as wisdom technology (wistech), an important computing and reasoning paradigm for intelligent systems. A rough-granular approach to wistech is proposed for developing one of its possible foundations. The proposed approach is, in a sense, the result of the evolution of computation models developed in the Rasiowa–Pawlak school. We also present a long-term program for implementation of what is known as a wisdom engine. The program is deﬁned in the framework of cooperation of many Research & Development (R & D) institutions and is based on a wistech network (WN) organization. Keywords: wisdom technology, adaptive rough-granular computing, rough sets, wisdom engine, open innovation, wisdom network. J.F. Peters et al. (Eds.): Transactions on Rough Sets VI, LNCS 4374, pp. 94–132, 2007. c Springer-Verlag Berlin Heidelberg 2007

A Wistech Paradigm for Intelligent Systems

1

95

Introduction

Huge technological changes occurred during the second half of the 20th century aﬀecting every one of us. These changes aﬀect practically all objects manufactured by man such as spoons, clothing, books, and space rockets. There are many indications that we are currently witnessing the onset of an era of radical changes. These radical changes depend on the further advancement of technology to acquire, represent, store, process, discover, communicate and learn wisdom. In this paper, we call this technology wisdom technology (or wistech, for short). The term wisdom commonly means “judging rightly” [50]. This common notion can be reﬁned. By wisdom, we understand an adaptive ability to make judgements correctly to a satisfactory degree (in particular, correct decisions) having in mind real-life constraints. One of the basic objectives of the paper is to indicate the potential directions for the design and implementation of wistech computation models. An important aspect of wistech is that the complexity and uncertainty of real-life constraints mean that in practise we must reconcile ourselves to the fact that our judgements are based on non-crisp concepts and also do not take into account all the knowledge accumulated and available to us. This is why consequences of our judgements are usually imperfect. But as a consolation, we also learn to improve the quality of our judgements via observation and analysis of our experience during interaction with the environment. Satisfactory decision-making levels can be achieved as a result of improved judgements. The intuitive nature of wisdom understood in this way can be expressed metaphorically as shown in (1). wisdom = KSN + AJ + IP,

(1)

where KSN, AJ, IP denote knowledge sources network, adaptive judgement, and interactive processes, respectively. The combination of the technologies represented in (1) oﬀers an intuitive starting point for a variety of approaches to designing and implementing computational models for wistech. In this paper, (1) is called the wisdom equation. There are many ways to build wistech computational models. In this paper, the focus is on an adaptive rough-granular approach. The issues discussed in this article are relevant for the current research directions (see, e.g., [16,15,31,38,51,90,108] and the literature cited in these articles). This paper is organized as follows.

2

Wisdom Technology

This section brieﬂy introduces the wistech paradigm. 2.1

What Do We Mean by Wistech?

On the one hand, the idea expressed by (1) (the wisdom equation paradigm) is a step in the direction of a new philosophy for the use of computing machines

96

A. Jankowski and A. Skowron

in our daily life, referred to as ubiquitous computing (see [66]). This paradigm is strongly connected with various applications of autonomic computing [64]. On the other hand, it should be emphasized that the idea of integrating many basic AI concepts (e.g., interaction, knowledge, network, adaptation, assessment, pattern recognition, learning, network, simulation of behavior in an uncertain environment, planning and problem solving) is as old as the history of AI itself. Many examples of such an approach adopted by researchers in the middle of the 20th century can be found in [27]. This research was intensively continued in the second half of the 20th century. For example, the abstracts of thousands of interesting reports from the years 1954 -1985 can be found in [91,92]. This paper contains the conclusions of the authors’ experiences during numerous practical projects implementing wistech technologies in speciﬁc applications, e.g., fraud detection (MERIX – a prototype system for Bank of America), dialogue based search engine (EXCAVIO – intelligent search engine), UAV control (WITAS project), Intelligent marketing (data mining and optimization system for Ford Motor Company, General Motors), robotics, EVOLUTIONARY CHECKERS (adaptive checker R&D program at the University of North Carolina at Charlotte) and many other applications. These experiences are summarized by the authors in the metaphoric wisdom equation (1). This equation can also be illustrated using the following diagram presented in Figure 1. In Figure 1 the term ‘data’ is understood as a stream of symbols without any interpretation of their meaning. From the perspective of the metaphor expressed in the wisdom equation (1), wistech can be perceived as the integration of three technologies (corresponding to three components in the wisdom equation (1)). At the current stage two of them seem to be conceptually relatively clear, namely 1. knowledge sources network – by knowledge we traditionally understand every organized set of information along with the inference rules; in this context one can easily imagine the following examples illustrating the concept of knowledge sources network: – representation of states of reality perceived by our senses (or observed by the “receptors” of another observer) are integrated as a whole in our minds in a network of sources of knowledge and then stored in some part of our additional memory, – a network of knowledge levels represented by agents in some multi-agent system and the level of knowledge about the environment registered by means of receptors; 2. interactive processes – interaction understood as a sequence of stimuli and reactions over time; examples are: – the dialogue of two people, – a sequence of actions and reactions between an unmanned aircraft and the environment in which the ﬂight takes place, or – a sequence of movements during some multi-player game. Far more diﬃcult conceptually seems to be the concept of adaptive judgement distinguishing wisdom from the general concept of problem solving. Intuitions behind this concept can be expressed as follows:

A Wistech Paradigm for Intelligent Systems

Wistech

Database Technology

HIERARCHY

wisdom = knowledge sources network + adaptive judgment + interactive processes knowledge = information + information relationships + inference rules

LEVELS

Information Technology

TECHNOLOGY

Knowledge Management Technology

97

information = data + interpretation

data

COMPLEXITY LEVELS OF THE SOLUTION PROBLEM SUPPORT

Perception

Prediction

Questions: about data values and additionally questions about data context, like: Who? What? When? Where? How Much?

Questions: information type questions and additionally questions about explanation and prediction, like: How? Why? What if?

Objects: Data and data explanation by a description, picture or other presentations.

Objects: Information and rules for information transformation (reasoning, ...), constrains, relationships between concepts, ideas and thought patterns.

Time context: Usually information is a posteriori, known after the fact. Measures: Logical values, uncertainty, completeness, amount.

Time context: Usually knowledge is a priori, known before the fact and provides its meaning. Measures: Efficiency of problem solutions by applying theory to information, quality of problem solutions.

Questions: knowledge type questions and additionally questions about correct judgments and decisions, action/ interaction planning / executing and justification, like: What to do? Why to do it? When to do it? How to do it? Objects: Knowledge and correct judgments, decisions based on a hierarchy of being values or believes, action plans, incorporation of vision, design, plans and implementation standards based on being preferences. Time context: Usually deals with management interactions with environment to achieve future objectives. Measures: Priorities, culture values, profits, quality of action results and plan implementation.

Understanding Questions:

Objects:

Time context:

Measures:

Fig. 1. Wisdom equation context

1. adaptive judgement – understood here as arriving at decisions resulting from the evaluation of patterns observed in sample objects. This form of judgement is made possible by mechanisms in a metalanguage (meta-reasoning) which on the basis of selection of available sources of knowledge and on the basis of understanding of history of interactive processes and their current status enable us to perform the following activities under real-life constraints: – identiﬁcation and judgement of importance (for future judgement) of sample phenomena, available for observation, in the surrounding environment;

98

A. Jankowski and A. Skowron

– planning current priorities for actions to be taken (in particular, on the basis of understanding of history of interactive processes and their current status) toward making optimal judgements; – selection of fragments of ordered knowledge (hierarchies of information and judgement strategies) satisfactory for making a decision at the planned time (a decision here is understood as a commencing interaction with the environment or as selecting the future course to make judgements); – prediction of important consequences of the planned interaction of processes; – adaptive learning and, in particular, reaching conclusions deduced from patterns observed in sample objects leading to adaptive improvement in the adaptive judgement process. One of the main barriers hindering an acceleration in the development of wistech applications lies in developing satisfactory computation models implementing the functioning of “adaptive judgement”. This diﬃculty primarily consists in overcoming the complexity of the process of integrating the local assimilation and processing of changing non-crisp and incomplete concepts necessary to make correct judgements. In other words, we are only able to model tested phenomena using local (subjective) models and interactions between them. In practical applications, usually, we are not able to give global models of analyzed phenomena (see, e.g., [110,62,64,45,25,21]). However, we can only approximate global models by integrating the various incomplete perspectives of problem perception. One of the potential computation models for “adaptive judgement” might be the rough-granular approach. 2.2

Main Diﬀerences Between Wisdom and Inference Engine

In natural language, the concept of wisdom is used in various semantic contexts. In particular, it is frequently semantically associated with such concepts as inference, reasoning, deduction, problem solving, judging rightly as a result of pattern recognition, common sense reasoning, reasoning by analogy, and others. As a consequence this semantic proximity may lead to misunderstandings. For example, one could begin to wonder what the diﬀerence is between the widely known and applied concept in AI of “inference engine” and the concept of “wisdom engine” deﬁned in this paper? In order to avoid this type of misunderstanding it is worth explaining the basic diﬀerence between the understanding of wisdom and such concepts as inference, reasoning, deduction and others. Above all, let as start with explaining how we understand the diﬀerence between problem solving and wisdom. The widespread concept of problem solving is described as some slight modiﬁcation of this notion deﬁned in the context of solving mathematical problems by George P´ olya in [84]. The concept of problem solving is understood in [84] as the following set of activities: 1. First, you have to understand the problem. 2. After understanding, then make a plan.

A Wistech Paradigm for Intelligent Systems

99

3. Carry out the plan. 4. Look back on your work. How could it be better? An attempt at explaining the concept of wisdom can be taken using the concept of problem solving in the following manner: wisdom is the ability to identify important problems, search for suﬃciently correct solutions to them, having in mind real life, available knowledge sources, personal experience, constraints, etc. Having in mind this understanding of wisdom we get at once the ﬁrst important diﬀerence. Namely, in the problem solving process we do not have the following important wisdom factor in the above sequence (1-4) of activities: 0. Learning to recognize patterns that identify important problems and problem solution constraints. Certainly, this is not the only diﬀerence. Therefore, one can illustrate the general diﬀerence between the concept of problem solving and wisdom as the diﬀerence between the concept of ﬂying in an artiﬁcially controlled environment (e.g., using a ﬂying simulator and problem solving procedures) and the concept of ﬂying Boeing 767 aeroplane in real-life dangerous environment (wisdom in a particular domain). One can therefore think that wisdom is very similar to the ability of problem solving in a particular domain of application, which in the context of the world of computing machines is frequently understood as an inference engine. The commonly accepted deﬁnition of the concept of inference engine can be found for example in Wikipedia (http://en.wikipedia.org/wiki/Inference engine). It refers to understanding of “problem solving” in the spirit of the book [84]. It reads as follows: An inference engine is a computer program that tries to derive answers from a knowledge base. It is the “brain” that expert systems use to reason about the information in the knowledge base, for the ultimate purpose of formulating new conclusions. An inference engine has three main elements. They are: 1. An interpreter. The interpreter executes the chosen agenda items by applying the corresponding base rules. 2. A scheduler. The scheduler maintains control over the agenda by estimating the eﬀects of applying inference rules in light of item priorities or other criteria on the agenda. 3. A consistency enforcer. The consistency enforcer attempts to maintain a consistent representation of the emerging solution. In other words, the concept of inference engine relates to generating strategies for the inference planning from potentially varied sources of knowledge which are in interaction together. So this concept is conceptually related to the following two elements of the wisdom equation: 1. knowledge sources network, 2. interactive processes.

100

A. Jankowski and A. Skowron

However, it should be remembered that wisdom in our understanding is not only some general concept of inference. The basic characteristic of wisdom, distinguishing this concept from the general understanding of inference, is adaptive ability to make correct judgements having in mind real-life constraints. The signiﬁcant characteristic diﬀerentiating wisdom from the general understanding of such concepts as problem solving or inference engine is adaptive judgement. In analogy to what we did in the case of problem solving, we can now attempt to explain the concept of wisdom based on the notion of an inference engine in the following manner: Wisdom is an inference engine interacting with a reallife environment, which is able to identify important problems and to ﬁnd for them suﬃciently correct solutions having in mind real-life constraints, available knowledge sources and personal experience. In this case, one can also illustrate the diﬀerence between the concept of inference engine and the concept of wisdom using the metaphor of ﬂying a plane. One could ask the question of which is the more general concept: wisdom or problem solving? Wisdom is a concept carrying a certain additional structure of adaptive judgement which in a continuously improving manner assists us in identifying the most important problem to resolve in a given set of constraints and what an acceptable compromise between the quality of the solution and the possibility of achieving a better solution is. Therefore, the question of what the more general concept is closely resembles the question from mathematics: What is the more general concept in mathematics: the concept of a ﬁeld (problem solving), or the concept of the vector space over a ﬁeld (wisdom understood as problem solving + adaptive judgement)? The vector space is a richer mathematical structure due to the action on vectors. Analogously to wisdom it is a richer process (it includes adaptive judgement - a kind of meta-judgement that encompasses recognition of patterns common to a set of sample objects that leads to judgements relating to problem solving). On the other hand, research into single-dimensional space can be treated as the research of ﬁelds. In this sense, the concept of vector space over a ﬁeld is more general than the concept of a ﬁeld. 2.3

Why Does Wistech Seem to Be One of the Most Important Future Technologies?

Nobody today doubts that technologies based on computing machines are among the most important technology groups of the 20th century, and, to a considerable degree, have been instrumental in the progress of other technologies. Analyzing the stages in the development of computing machines, one can quite clearly distinguish the following three stages in their development in the 20th century: 1. Database Technology (gathering and processing of transaction data). 2. Information Technology (understood as adding to the database technology the ability to automate analysis, processing and visualization of information). 3. Knowledge Management Technology (understood as systems supporting organization of large data sets and the automatic support for knowledge processing and discovery (see, e.g., [59,18])).

A Wistech Paradigm for Intelligent Systems

101

The three stages of development in computing machine technology show us the trends for the further development in applications of these technologies. These trends can be easily imagined using the further advancement of complexity of information processing (Shannon Dimension) and advancement of complexity of dialogue intelligence (Turing Dimension), viz., – Shannon Dimension level of information processing complexity (representation, search, use); – Turing Dimension the complexity of queries that a machine is capable of understanding and answering correctly. One of the objectives of AI is for computing machines to reach the point in Turing Dimension that is wellknown Turing Test (see [114]). In this framework, the development trends in the application of computing machines technology can be illustrated in Figure 2.

Technology

Additional

Shannon

Turing

attributes

Dimensions

Dimensions

Database

data is the

How to

Technology

most basic

represent

SQL

level

information?

Information

information =

Where to find

Who? What?

Technology

data +

information?

When? Where?

interpretation

How much?

Knowledge

knowledge =

How to use

How? Why?

Management

information +

information?

What if?

Technology

information relationships + inference rules

Fig. 2. Computing machines technology

Immediately from the beginning of the new millennium one can see more and more clearly the following new application of computing machine technology, viz., wisdom technology (wistech) which put simply can be presented in table (see Figure 3, being an extension of the table presented in Figure 2). In other words, the trends in the development of technology of computing machines can be presented using the so-called DIKW hierarchy (i.e., Data, Information, Knowledge, Wisdom). Intuitively speaking, each level of the DIKW hierarchy adds certain attributes over and above the previous one. The hierarchy is presented graphically in Figure 4.

102

A. Jankowski and A. Skowron Technology

Additional

Shannon

attributes

Dimensions

Wisdom

Wisdom equation, Learn when to

Technology

i.e. wisdom

(Wistech)

knowledge

=

use

Turing Dimensions How to make

information correct judgements (in.

sources network

Learn how

+ adaptive

to get important correct

judgment +

information

particular

decisions)

interactive

heaving in mind

processes

real-life constraints?

Fig. 3. Computing machine technology (continued)

DIKW hierarchy can be traced back to the well-known poem by T. S. Eliot, “The Rock”, written in 1932. He wrote: Where is the life we have lost in living? Where is the wisdom we have lost in knowledge? Where is the knowledge we have lost in information? It is a truism to state that the eﬀects of any activity depend to a decisive degree on the wisdom of the decisions taken, both in the start, during the implementation, improvement and completion of the activity. The main objective of wistech is to automate support for the process leading to wise actions. These activities cover all areas of man’s activities, from the economy, through medicine, education, research, development, etc. In this context, one can clearly see how important role may have the wistech development in the future. The following comment from G.W. Leibniz on the idea to automate the processing of concepts representing thoughts should not surprise us either: No one else, I believe, has noticed this, because if they had ... they would have dropped everything in order to deal with it; because there is nothing greater that man could do. 2.4

A General Approach to Wistech Computation Model

In order to create general Wistech computational models let us start an analysis of the concept of adaptive judgement. For better familiarization of adaptive judgement we shall use the visualization of processes based on the IDEFO standard. Put simply, this means of visualisation is described in the diagram presented in Figure 5. An intrinsic part of the concept of judgement is relating it to the entity implementing the judgement. Intuitively this can be a person, animal, machine, abstract agent, society of agents, etc. In general, we shall call the entity making

A Wistech Paradigm for Intelligent Systems

103

Fig. 4. DIKW hierarchy

Controls constrain and direct activities

Inputs are transformed into Output

Movements or activity result items

Controls

Activity Inputs

Outputs

Mechanisms

Physical aspects of the activity e.g people, machines, tools

Activity, process or subprocess

Fig. 5. Activity

a judgement the judge. We shall also assume that knowledge sources network is divided into external sources, i.e., sources of knowledge that are also available to other judges, internal sources, which are only available to the speciﬁc judge in question.

104

A. Jankowski and A. Skowron Evolution rules

Real Life Constrain

Reception of External interactive process Explanation and modification proposals for External Communication Language

Input wisdom communication messages from external knowledge sources network

Adaptive Judgment

External environment conditions („weather”)

Output wisdom communication messages to external knowledge sources network - in particular knowledge messages and proposals of modifications of External Communication Language (for example new communication rules, new concepts, names, ...) External actions in order to implement desired interactive processes

Implementation issues

Fig. 6. The ﬁrst level of the model

The ﬁrst level of the model is presented in Figure 6. Of course, successive levels of the model are more complex. Its details may depend on the assumed paradigms for the implementation of adaptive judgement. However, these details should include such elements as: 1. Learning of the External Communication Language understood as a language based on concepts used to communicate and process knowledge with a network of external sources of knowledge; 2. Learning of the Internal Communication Language understood as a hierarchy of meta-languages based on concepts used to process and improve External Communication Language and a language based on concepts used to communicate and process knowledge with a network of internal sources of knowledge; 3. Receiving in memory signals from signal receptors and interactive processes and expressing their signiﬁcance in the External Communication Language and the Internal Communication Language; 4. Planning the current priorities for internal actions (mainly related to the processing of wisdom) on the basis of an assessment in relation to the hierarchy of values controlling the adaptive judgement process; 5. Selection of fragments of ordered knowledge (hierarchies of information and judgement strategies) suﬃcient to take a decision at the planned time (a decision here is understood as commencing interaction with the environment or selecting the future course to resolve the problem);

A Wistech Paradigm for Intelligent Systems

105

6. Output wisdom communication messages to external knowledge sources network, in particular, knowledge messages and proposals of modiﬁcations of the External Communication Language (e.g., new communication rules, new concepts, names); 7. External actions in order to implement the desired interactive processes. All elements occurring in the above list are very complex and important but the following two problems are particularly important for adaptive judgement computational models: 1. Concept learning and integration - this is the problem of computational models for implementation of learning concepts important for the representation, processing and communicating of wisdom and, in particular, this relates to learning of concepts improving the quality of approximation of the integration of incomplete local perceptions of a problem (arising during local assimilation and processing of vague and incomplete concepts (see, e.g., [78,79])). 2. Judge hierarchy of habit habits controls - this is the problem of computational models for implementation of process of the functioning of a hierarchy of habit controls by a judge controlling the judgement process in an adaptive way. Now, we sketch the idea of a framework for solution of the problem of implementation of judge hierarchy of habit controls. In this paper, we treat a concept of habit as an elementary and repeatable part of behavioral pattern. In this context, the meaning of elementary should be considered by comparison to the required reasoning (knowledge usage) complexity necessary for the behavioral pattern implementation. In other words, by a habit we mean any regularly repeated behavioral pattern that requires little or no reasoning eﬀort (knowledge usage). In general, any behavioral pattern could be treated as a sequence of habits and other activities which use knowledge intensively. Among such activities those leading to new habits are especially important. We assume that such habit processing is controlled by so-called habit controls which support the following aspects of adaptive judgement process for a considered situation by a judge: 1. Continuous habit prioritization to be used in a particular situation after identiﬁcation of habits. This is a prioritization from the point of view of the following three criteria: – The predicted consequences of the phenomena observed in a considered situation; – Knowledge available to a judge; – The actual plans of a judge’s action. 2. Knowledge prioritization is used if we do not identify any habit to be used in a considered situation, then we have to make prioritization of pieces of available knowledge which could be used to choose the best habit or for a construction of a new habit for the considered situation. 3. Habit control assessment for continuous improvement of adaptive judgement process and for construction of new habits and habit controls.

106

A. Jankowski and A. Skowron

SIGNAL FOR ACTION

JUDGE judging in order to take action

ACTION TAKEN

Fig. 7. Judge diagram

As it can be seen from the above considerations, one of the key components of wistech, judge hierarchy of habit control, is essential for optimal decision making and is closely correlated with the knowledge held and interactions with the environment. Judge hierarchy also means the desire of the judge to satisfy his/her needs in interactions with his/her environment. Put very simply, the judge receives and sends out signals according to the diagram presented in Figure 7. The interior of the box is the place for the judge to process signals and to take an action. By the judge environment adaptation we understand the interaction of the following two adaptive processes: 1. adaptation of the environment, in which the judge lives to the needs and objectives of the judge so as to best ﬁt the needs and objectives of the environment, 2. adaptation of the internal processes taking place in a judge in such a way as to best realize his/her needs and objectives based on the resources available in the environment. The judge environment adaptation is the basis for computational models of judge learning. The key part of this is the evolution of judge hierarchy of habit controls. The judge hierarchy of habits controls constitutes a catalyst for evolutionary processes in the environment, and also constitutes an approach to expressing various paradigms of computation models to be used in the machine implementation of this concept. For example, these paradigms can be based on the metaphorically understood principle of Newtonian dynamics (e.g., action = reaction), thermodynamics (e.g., increase in entropy of information), quantum mechanics (the principle of it being impossible to determine location and speed simultaneously) and quantum computational models [44], psychology (e.g., based on metaphorical understanding of Maslow’s hierarchy of needs; see also [53,80,40]). Particularly worthy of attention in relation to wistech is the metaphoric approach to Maslow’s hierarchy of needs in reference to the abstractly understood community of agents. Put simply, this hierarchy looks as in Figure 8. It could be used for direct constructions of computational models of judge hierarchy of habit controls.

A Wistech Paradigm for Intelligent Systems

PHYSICAL

EMOTIONAL

MENTAL

SPIRITUAL

Transcendental

Self-fulfillment

Aesthetic Needs Cognitive Needs

Respect and an appreciation of one's own value Allegiance to a group and the desire for love

Security and intactness

Physiological needs

107

. Transcendental (the highest level, confirming the existence of higher being) - spiritual needs. Self-realisation (the need to have objectives, the need to satisfy one's own potential) - the needs of self-realization - are expressed in the striving of man to develop his possibilities; they constitute a means to satisfying physiological needs or reactions compensating for the failure to satisfy other needs. Aesthetic needs (the need for harmony and beauty) Cognitive needs (the need for knowledge, understanding, novelty)

Respect (the need to believe in oneself, a feeling of one's own value, competencies, authority) - the need of recognition (respect) and prestige in one's own eyes and the eyes of other people; desiring power, achievement and freedom; a need for respect and recognition from other people, good social standing, fame, dominance, attention. Allegiance (the need for ties, affiliations, love and to be loved) occur in attempts to overcome loneliness, elimination and alienation, a tendency to make close intimate relations, participation in the life of a group, in efforts to create gangs.

Security needs (defined by dependence, care and support, protection, need of comfort, peace, freedom from fear) stimulates for action, secures intactness; they reveal when the existing habits are becoming less useful. Physiological needs (lowest level needs such as food, water, oxygen, sexual needs, lack of tension) if not satisfied they dominate over all other needs, taking them on the back seat; decide on human behavior processes

Hierarchy of Human Needs A. Maslow (about 1934)

Fig. 8. The Maslow Hierarchy of human needs (about 1934) as an example of judge hierarchy of habit controls

2.5

A Rough-Granular Computing Approach to Wistech Computation Models

In this section, we outline basic ideas for the rough-granular approach to wisdom. 2.5.1 Evolution of Reasoning Computation Models in the Rasiowa–Pawlak School By the Rasiowa–Pawlak school we mean a continuation of approaches to computational models of approximate reasoning developed by Rasiowa [86], Pawlak [74,87], and their students. In some sense, it is a continuation of ideas initiated by Leibniz, Boole and currently continued in a variety of forms over the world. Of course, the Rasiowa–Pawlak school is also some kind of continuation of the Polish School of Mathematics and Logics. The achievements of this school led to the development of the modern understanding of the basic computational aspects of logic, epistemology, ontology, foundations of mathematics and natural deduction (S. Banach, S. Eilenberg, R. Ingarden, S. Ja´skowski, K. Kuratowski,

108

A. Jankowski and A. Skowron

S. Le´sniewski, A. Lindenbaum, J. L ukasiewicz, S. Mazur, A. Mostowski, H. Rasiowa, R. Sikorski, W. Sierpi´ nski, A. Tarski, S. Ulam, and many others). Two fundamental tools of the Rasiowa–Pawlak school are the following: – Computation models of a logical concept (especially of such concepts as deduction or algebraic many-valued models for classical, modal, and constructive mathematics). The Rasiowa–Pawlak approach is based on the method of treating the sets of logically equivalent statements (or formulas) as abstract algebras known as the Lindenbaum–Tarski algebras. – Computation models of vague concept. L ukasiewicz originally has proposed to treat uncertainty (or vague concepts) in logic as concepts of many-valued logic. However, software developed for today’s computers is based on two-valued Boolean algebra. Therefore it is more practical to treat uncertainty and vagueness using the classical logic concept based on two-valued Boolean algebra. The concept of a rough set introduced by Pawlak [74] and developed in the Rasiowa–Pawlak school is based on the classical two-valued logic and, hence, the rough set approach is important and suitable for the applications mentioned above. The rough set approach intended to deal with uncertainty and vagueness has been developed to deal with uncertainty and vagueness. The rough set approach makes it possible to reason precisely about approximations of vague concepts. These approximations are tentative, subjective, and varying accordingly to changes in the environment [75,76,77,8]. Both the above mentioned fundamental tools can be applied in many contexts. It is interesting to illustrate evolution of the both above fundamental tools from the Rasiowa–Pawlak school perspective (see Figure 9 and Figure 10). 2.5.2 Rough-Granular Computing (RGC) Solving complex problems by multi-agent systems in distributed environments requires approximate reasoning methods based on new computing paradigms. One such emerging recently computing paradigm is RGC. Computations in RGC are performed on information granules representing often vague, partially speciﬁed, and compound concepts delivered by agents engaged in tasks such as knowledge representation, communication with other agents, and reasoning. We discuss the rough-granular approach for modeling computations in complex adaptive systems and multiagent systems. Information granules are any objects constructed when modeling of computations, and in approximating compound concepts, and approximate reasoning about these concepts. Information granules are constructed in an optimization process based on the minimal length principle. This process is aiming at constructing approximations of concepts satisfying some (vague and/or uncertain) constraints. Examples of information granules are information systems and decision systems, elementary information granules deﬁned by indiscernibility neighborhoods, families of elementary granules (e.g., partitions and coverings),

A Wistech Paradigm for Intelligent Systems Domain & Operators

Natural Numbers Calculus

Algebra of subsets

Boolean Algebra

Logical concepts Semantical models in Lindenbaum – of constructive Tarski algebra mathematics

X is smaller than Y

X is a subset of Y

X is smaller than Y in Boolean algebra

Y can be deduced from Y

0

Zero

Empty set

The smallest element

Falsity

1

One

Full set

The biggest element

Truth

+

Addition

Join of two sets

Maximum

Disjunction

*

Multiplication

Intersection of Minimum two sets

Conjunction

Exponentiation of X to power Y

Join of (–Y) and X

Join of (–Y) and X

Implication (Y implies X)

Mod (X)

Modulo X calculus

Logical values

True False

Quotient algebra of the filter generated by set X True False

Quotient Boolean algebra of the filter generated by set X True False

Lindenbaum – Tarski algebra for a theory generated by a set of axioms X Algebra of logical values

X t+ (u |=+ deg u )} for u ∈ NInf . Granule u votes “for” granules from N + (u) (relative to t+ ) (see [46]). Let us assume N − (u) = {u ∈ D : ∃deg < t− (u |=− deg u )}

for u ∈ NInf . Then granule u votes “against” granules from N − (u) (relative to t− ). We assume that B is a distinguished set of information granules called behavioral patterns of J (e.g., decisions, actions, plans [34,115]) and Lab : D −→ B is the (partial) labeling function assigning the behavioral patterns to (some) information granules from D. • S is one of the judgemental strategies of J making it possible to select a particular behavioral pattern as a reaction to the perceived information about the environment. In particular, S uses granules from Lab(N + (u)) and Lab(N − (u)), where u = Inf (x) and x is the current state of the environment, and the labeling of these sets of granules by behavioral patterns. Observe that the strategy S should resolve conﬂicts arising due to the fact that information granules should satisfy some constraints. For example, some information granules cannot be matched by one information granule to a degree higher than a given threshold t+ . • Quality strategy of J. Q is the quality strategy of J for estimation of the closeness (similarity) between granules. The closeness estimation is

A Wistech Paradigm for Intelligent Systems

115

based on arguments “for” and “against” the satisﬁability of the compound concept of “closeness” represented by Q. In this judgement J uses relevant granules from available granules representing knowledge accessible for J, often distributed among other judges, as well as the relationships between granules represented by matching degrees. • Adaptation strategy of J. Adap is the adaptation strategy transforming a tuple − (N, G, Inf, B, Lab, |=+ deg , |=deg , S, Q) into a new such tuple. Observe that judgements performed by J during adaptation can, in particular, lead to construction of new granules (e.g., through cooperation with other judges [2]), changing some strategies such as the matching strategy, the labeling strategy, the selection strategy for relevant behavioral patterns, and the strategy for estimation of closeness of granules. Adap can also be changed, e.g., by tuning some of its parameters. 2.5.5 Basic Cycle of Judge Each judge realizes some goals using behavioral patterns. The basic cycle of each judge J is the following: 1. Step 1: Initialization. − (N, G, Inf, B, Lab, |=+ deg , |=deg , S, Q) := − (N0 , G0 , Inf0 , B0 , Lab0 , |=+ deg,0 , |=deg,0 , S0 , Q0 ). 2. Step 2: Perception granule construction by J representing the current state. u := Inf (x); where u is the granule representing perception by J of the current environment state x. 3. Step 3: J selects the relevant granules from N + (u), N − (u) and performs judgements to select (construct) the relevant behavior b toward achieving the current goal (target). During selection of b the judge J is also predicting the information Infpred (b, x) returned from EN VJ as a reaction to the behavior b applied to the current state x of EN VJ . This is realized by another special judgemental strategy of J. By applying S to Lab(N + (u)) and Lab(N − (u)) J searches for a relevant behavior b. 4. Step 4: Estimation of the closeness. The judge J uses the quality measure Q for estimation of the closeness (similarity) of Infpred (b, x) and Infreal (b, x) by Q(Infpred (b, x), Infreal (b, x)), where Infreal (b, x) is information about the real reaction of the environment in state x to the behavior b.

116

A. Jankowski and A. Skowron

5. Step 5: J uses a special judgemental strategy in testing whether the closeness is satisfactory. If the closeness is satisfactory, then J continues from Step2; otherwise J goes to the next step. 6. Step 6: Adaptation step. − (N, G, Inf, B, Lab, |=+ deg , |=deg , S, Q) := − Adapt(N, G, Inf, B, Lab, |=+ deg , |=deg , S, Q). 7. Step 7: Go to Step 2. All constructive objects involved in computations realized by means of the above judgement schemes are information granules. 2.5.6 Remark on Task Solving by Systems of Judges The above examples illustrate the complexity and richness of the information granule spaces we deal with when modeling adaptive processes and reasoning about such processes. Systems of judges solve tasks by searching in the information granule spaces for information granules satisfying the task speciﬁcation to a satisfactory degree (not necessarily exactly), i.e., matching information granules representing the task speciﬁcation to a satisfactory degree. The requirement of “matching to a degree” used instead of “matching exactly” often makes searching for solutions feasible in information granule spaces [122]. In a number of papers (see, e.g., [99,105,106]), we have developed methods for construction of information granules (satisfying a given speciﬁcation to a satisfactory degree) by means of operations on information systems called constrained sums. In particular, this approach proved to be general enough for modeling compound spatio-temporal information granules (e.g., information granules representing processes or behavioral patterns speciﬁed by vague concepts) and interactions between them.

3

Wistech Network (WN)

In this section, we discuss shortly the organization of cooperation for the projects based on wistech. 3.1

What We Mean by Wistech Network

The huge complexity of the problem of designing eﬀective wistech computation models means that wistech progress signiﬁcantly depends on forming eﬀective and systematic cooperation between the numerous interdisciplinary teams verifying the Wistech calculation models developed in practical experiments. Moreover, in order to make a really essential progress in wistech it is important to involve the best possible specialists for making it possible to combine in wistech based projects knowledge of such areas as: psychology, sociology, ethics and domain dependent knowledge, e.g., neuroscience, medicine, economics,

A Wistech Paradigm for Intelligent Systems

117

security, law, robotics, telecommunications, banking. This research, like all other research, requires a signiﬁcant eﬀort in other fundamental sciences, such as logic, epistemology, ontology, mathematics, computer science, philosophy and others. Of course such activity is very expensive. Moreover, in general, research of this type does not translate directly into economic results. No private company can aﬀord to implement such extensive research by itself. It is also unlikely that there would be any signiﬁcant commitment by government agencies in the coordination and development of research on such a wide scale. Unfortunately, current attempts at extending the international coordination of such type of research are not eﬀective. A dilemma therefore arises whether to develop wistech within the framework of expensive and highly risky closed research programs, or to support open programs in which the costs and risk are spread among many entities? It is our opinion that both directions are equally important and the key to the success is an environment for creating and developing harmony mechanisms between open and closed research (see [19]). In [19], among others, the contrasting principles of closed and open innovation are clariﬁed (see Figure 11). At the current stage of building an environment for creating and developing harmony mechanisms between open and closed research it is very important to develop a powerful framework for eﬀective Open Innovation Wistech R&D network. The current stage of development in wistech above all requires the development of coordinated interdisciplinary basic research with a well-coordinated and easily accessible environment for experiments. Such activities are not possible in hermetically sealed companies, which are paralyzed by security procedures and guided by the criterion of rapid economic return. This is also why it is proposed to start up mechanisms for the systematized and relatively coordinated cooperation of centers interested in developing Wistech under a Wistech Network (WN) cooperating with one another in accordance with jointly perfected open principles based on Open Innovation Principles. It is worth stressing that organizations preferring Closed Innovation Principles may also draw great beneﬁts from active participation in WN. This participation gives the possibility of testing solutions that have little chance of giving rapid market results, and also in the case of the appearance of such opportunities they can be translated into economic results in accordance with the principles accepted. At the same time, in the case of basic research, which in general does not translate directly into market eﬀects, the understanding of progress in basic research gives greater opportunities for developing new market applications of one’s own. A further great beneﬁt of active participation in WN should be the possibility of comparing the various paradigms for building calculation models for Wistech. The times have long since gone when people believed that there is only one perfect paradigm in AI. Hybrid solutions adapted to the speciﬁc nature of the sphere of application dominate in applications. Hybrid applications themselves also use a variety of construction paradigms in a platform for integrating various approaches. Similarly we are also assuming that the WN environment would be represented in

118

A. Jankowski and A. Skowron Contrasting Principles of Closed and Open Innovation

Closed Innovation Principles

Open Innovation Principles

The smart people in our field work

Not all smart people work for us.

for us.

We need to work with smart people inside and outside of our company.

To profit from R&D, we must

External R&D can create significant

discover it, develop it, and ship

value; internal R&D in needed to

it ourselves.

claim some portion of that value.

If we discover it ourselves, we

We don’t have to originate the

will get it to the market first.

research to profit from it.

The company that gets an innovation Building a better business model is to the market first will win.

better than getting to the market first.

If we create the most and the best

If we make the best use of internal

ideas in the industry, we will win. and external ideas, we will win. We should control our intellectual

We should profit from others’ use

properties (IP), so that our

of our IP, and we should buy

competitors don’t profit from our

others’ IP whenever it advances our

ideas.

own business model.

Fig. 11. Contrasting principles of closed and open innovation [19]

the form of a sub-network with various paradigms for the construction of an integration platform. WN would provide the data and criteria to assess the results of experiments used for the assessment of various paradigms. In the remainder of this work we present, among others, a proposal to start up a sub-network based on a paradigm for the integration of various technologies based on an adaptive rough-granular computing approach (RGC). 3.2

A Potential Example Scenario of WN Establishment

3.2.1 WN Long-Term Vision and Role The basic objectives of WN are supporting open innovation and the development of wistech and its applications through: 1. creating new paradigms and trends in Wistech and its applications, 2. creating a platform (e.g. intranet, symposia, training programs, e-learning, etc.) for communication and the exchange of knowledge and experience on the practical applications and achievements of basic research, 3. preparing educational and research programs, 4. starting up projects for speciﬁc practical applications, as well as for basic research,

A Wistech Paradigm for Intelligent Systems

119

5. establishing the conditions and criteria used to compare the quality of various approaches to Wistech (especially having in mind applications in medicine, economy, agriculture, energy and forex market), 6. popularization of Wistech. 3.2.2 WN Organization and Financial Support We assume that participation in WN is absolutely voluntary in nature, and WN itself also does not assume any additional ﬁnancial fees or obligatory participation in conferences. The organization is open in nature and any person or organization can take part in it. The form of organization is based on communities cooperating together, which jointly use and develop open software (see, e.g., http://www.opensource.org/). At the same time we assume that at some stage WN may take part in commercial projects. The project participants will mutually agree upon the principles for cooperation in every such case. It is expected that in the long-term some products or components created by WN will function according to the principles of open software (e.g. similar to the principles of http://www.opensource.org/). We continue to assume the organization of working groups in the network which would deal with jointly agreed packets of problems and projects. It is expected in our exemplary scenario that WN will develop in accordance with the stages for development of a mature organization modeled on the ideas of Carnegie Mellon Capability Maturity Model (http://www.sei.cmu.edu/cmm/). This model consists of the six stages presented in Figure 12 and Figure 13. The basic assumption to WN is the realization of projects ﬁnanced by WN participants who cover the costs and risk of their own activities in the network. It is also assumed that in WN there will be several specialist centers which will coordinate the activities in individual areas (competency centers), e.g. the multi-agent approach, the rough mereology approach. The coordination work of these centers would be ﬁnanced from voluntary ﬁnancial contributions from participants of the group in question. It follows from this that the intensity and quality of work in a given group will to a large degree depend on the level of ﬁnancial support from participants in the group.

4

Wisdom Engine

We discuss some exemplary projects proposed as pilot projects in development of wistech. 4.1

Wisdom Engine Concept

By wisdom engine we understand a machine system which implements the concept of wisdom. In other words, the basic functions of the wisdom engine would be acquiring, processing, discovering, learning and communicating wisdom. One of the main ﬁrst objectives of WN can be to create an open international R&D environment for the design and implementation of the concept of universal domainindependent wisdom engine. A universal wisdom engine implementation should

120

A. Jankowski and A. Skowron

Name of stage ESTABLISHMENT

INITIAL

Organization Starting up the first projects in the network and defining the principles for the cooperation of the first group of participants who confirm their participation in WN. Starting up the first forms of communication.

Content Developing the initial catalogue of paradigms for approaches to development of wistech (e.g., multiagents, evolution, symbolic processing, neural nets, statistics, adaptive rough granular approach, formal concepts, ontology engineering, information semiotics, cognitive and epistemological approach, etc., and their combinations). The preliminary allocation Developing a common of categorized paradigms language to describe the for approaches to wistech concepts relating to to their respective starting up, implementing and closing projects in WN. competency centers. Allocating a paradigm to a competency center, e.g. multi-agent approach, adaptive rough granular approach, etc. This does not mean that at a given competency center only and exclusively this method will be developed. On the contrary, it is assumed that every competency center will develop hybrid solutions combining various approaches. At the same time, a competency center will particularly strongly develop aspects relating to the paradigms allocated to this center.

Fig. 12. Six stages of the Carnegie Mellon Capability Maturity Model

be independent of any speciﬁc application domain. At the same time, functionality of the universal wisdom engine should enable the conﬁguration and tuning of modules for it in the form of a series of products dependent on speciﬁc application domains such as, e.g., medicine, economics, stock market, forex market, security, law, tourism, telecommunications, banking, job market. In particular universal wisdom engine should be able to learn domain knowledge by reading, discussing with experts and gathering wisdom from experience. Of course, the design and implementation of a universal wisdom engine is an extremely diﬃcult task and probably unrealistic today in a short term. First of all, we have to do some experiments with several application domains and several diﬀerent paradigms for wistech implementation. Based on an analysis of the results of such experiments we can create a more general wistech ontology which should provide a better formal framework for the implementation of a universal wisdom engine.

A Wistech Paradigm for Intelligent Systems REPEATABLE

121

Establishing the principles for

Establishing the mutually tied

selecting good practices

objectives to achieve at the

specific for the implementation

individual competency centers

of a project in wistech,

in order to verify the

designed to repeat the

effectiveness and

successes of projects realized

possibilities of developing

in similar conditions and to

various approaches.

avoid failures. Establishing the list of first conditions and criteria used to compare the quality of various approaches to wistech.

DEFINED

Putting in writing and the

Starting up the first projects

effective implementation of a

realized in the common

list of joint standards for

standards by a variety of

organization and management of

centers within the WN

projects specific to wistech, that will be binding for the WTN community. MEASURABLE

Enhancing the standards arising

Starting up mechanisms for

at the previous stage to

competitiveness between

include sets of measurable

communities working on various

indices used to verify and

approaches to wistech in the

optimize the benefits to costs

network.

of wistech projects.

CONTINUOUS IMPROVEMENT

Enhancing the standards and

Developing the optimum methods

indices defined at the

for harmonious co-operation

MEASURABLE stage to set out in

between WN and commercial

writing and effectively

companies.

implement procedures for continuously improving the functioning of WN.

Fig. 13. Six stages of the Carnegie Mellon Capability Maturity Model (continued)

Thus, it is assumed that in parallel with the work on a universal concept of a wisdom engine, work would also be conducted on utilizing the wisdom engine in selected areas of application, e.g., medicine, economics, stock market, forex market, security, law, tourism, telecommunications, banking, or job market. The long-term vision is as follows: “wisdom engineers” will receive the task to create the conﬁguration for the wisdom engine for applications in a speciﬁc ﬁeld of life, and then, after having carried out the necessary analytical and design work, to

122

A. Jankowski and A. Skowron

conﬁgure the wisdom engine and to enter the necessary initial data. The wisdom engine should have properties for self-growth and adaptation to changing conditions of its environment, as well as advances in wisdom in the ﬁelds of application. This is why one should strongly emphasize the planned property of automatic adaptation of the system – a feature not taken into account in the construction of the numerous systems in the past that were intended to perform similar tasks. A classic example here is the long-standing MYCIN project implemented by Stanford University. The implementation of the idea expressed by the wisdom equation is very diﬃcult and it would be unreasonable to expect its full implementation in a short period of time. We assume that the creativity cycle for the ﬁrst product prototypes implementing this concept would take several years of intensive work with cooperation of product managers, scientists, engineers, programmers and domain experts. On the other hand, it is not desirable to implement such long projects without any clear interim eﬀects. This is why we assume that the wisdom engine implementation project would go through several phases. For example, initially we assume they will go through ﬁve phases in the implementation of the wisdom engine. We propose a route, to achieving the target wisdom engine products through continuously improving intermediary products that meet successive expansions in functionality. The ﬁve phases are called as follows: 1. 2. 3. 4. 5.

Summary, Spider, Conceptual Clustering and Integration, Wisdom Extraction, Wisdom Assistant.

The eﬀect of each of these phases will be a prototype product that after acceptance would be interesting for the WN community. Stated in simple terms the functional eﬀects of the individual phases would be as presented in Figure 14. 4.2

Examples of Wisdom Engine Domain-Dependent Product Lines

The above ﬁve phases (i.e., Summary, Spider, Conceptual Clustering and Integration, Wisdom Extraction, and Wisdom Assistant) should be applied to several directions for potential product lines which would be developed in the WN. Of course, it can theoretically be any product relating to applications in robotics, unmanned aircraft, space rockets, etc. However, if we wish to have as many people as possible cooperating in the WN, then the product lines must be chosen so that experimenting with them does not prove expensive. On the other hand, these product lines must be suﬃciently attractive so as to interest as many people as possible. We propose that these product lines relate to applications in such areas as medicine, economics, the stock market, forex market, security, law, tourism, telecommunications, banking, job market and others. The list of products that could be expanded in accordance with the above scheme is potentially unlimited. The proposals for the descriptions of speciﬁc

A Wistech Paradigm for Intelligent Systems Phase

Key functions

Summary

Spider

Conceptual Clustering and Integration key document document concept searching clustering extraction related to based on key key concept concept and and indexing Integration documents

123

Adaptive Wisdom Extraction

Adaptive Wisdom Assistant

extracting data structures, information structures, knowledge structures and adaptive wisdom structures from documents, in particular generating thesauruses, conceptual hierarchies and constraints to acceptable solutions

user query / answering processing in order to support users in solving their problems as effectively as possible

Fig. 14. Functional eﬀects of the individual phases

Summary

Spider

Document Manager

Phase/Product

Document Summary

Document Spider

Conceptual Clustering Document Conceptual Clustering

Job Market

Job Market Summary

Job Market Spider

Job Market Conceptual Clustering

Brand Monitoring

Brand Monitoring Summary

Brand Monitoring Spider

Brand Monitoring Conceptual Clustering

World Communication

World Communication Summary

World World Communication Communication Spider Conceptual Clustering

World Forex

World Forex Summary

World Forex Spider

World Stock Market

World Stock Market Summary

World Stock World Stock Market Spider Market Conceptual Clustering

World Tourist

World Tourist Summary

World Tourist World Tourist Spider Conceptual Clustering

Physician

Physician Summary

Physician Spider

Lawyer

Lawyer Summary

Lawyer Spider Lawyer Conceptual Clustering

Economy Monitoring

Economy Monitoring Summary

Economy Monitoring Spider

World Forex Conceptual Clustering

Physician Conceptual Clustering

Economy Monitoring Conceptual Clustering

Wisdom Extraction Document adaptive wisdom Extraction Job Market adaptive wisdom Extraction Brand Monitoring adaptive wisdom Extraction World Communication adaptive wisdom Extraction World Forex adaptive wisdom Extraction World Stock Market adaptive wisdom Extraction World Tourist adaptive wisdom Extraction Physician adaptive wisdom Extraction Lawyer adaptive wisdom Extraction Economy Monitoring adaptive wisdom Extraction

Wisdom Assistant Document adaptive wisdom Assistant Job Market adaptive wisdom Assistant Brand Monitoring adaptive wisdom Assistant World Communication adaptive wisdom Assistant World Forex adaptive wisdom Assistant World Stock Market adaptive wisdom Assistant World Tourist adaptive wisdom Assistant Physician adaptive wisdom Assistant Lawyer adaptive wisdom Assistant Economy Monitoring adaptive wisdom Assistant

Fig. 15. Proposed products

products, included in the later part of this report, should be treated as ﬂexible and primarily constitute material for discussion, and not a ﬁnal decision. On the other hand, the list of products described is not entirely accidental in nature.

A. Jankowski and A. Skowron

Job Market

Document Manager

Product / Phase

124

Summary

Spider

Conceptual Clustering

Adaptive Wisdom Adaptive Wisdom Extraction Assistant

automatic summarizing of a document and groups of documents, the contents of which are not connected with any specific field

automatic searching and downloading of any documents

conceptual clustering of documents on any subject

automatic summarizing of a document and groups of documents relating to job market, carried out from the perspective of the following groups of users: potential employers and potential employees

automatic searching and downloading of documents relating to job market, carried out with particular emphasis on the needs of the following groups of users: potential employers and potential employees

conceptual clustering of documents relating to job market with particular emphasis on the specific nature of queries submitted by the following types of users: potential employers and potential employees

extracting data structures, information structures, knowledge structures and adaptive wisdom structures from documents, in particular generating thesauruses, conceptual hierarchies and constraints to acceptable solutions in any domain extracting data structures, information structures, knowledge structures and adaptive wisdom structures from documents, in particular generating thesauruses, conceptual hierarchies and constraints to acceptable solutions in job market domain

general user query / answering processing in order to support users in solving their problems as effectively as possible

job market related to user query / answering processing in order to support users in solving their problems as effectively as possible

Fig. 16. Functionality of individual products

This is because they form a certain logical continuity, connected both with the degree of diﬃculty in successive products and current preferences resulting from the previous experiences of the human resources that would be engaged to carry out the work on individual products. The initial selection of product lines is as follows: – – – – – – – – – –

Document Manager, Job Market, Brand Monitoring, World Communication, World Forex, World Stock Market, World Tourist, Physician, Lawyer, Economy Monitoring.

World Communication

Brand Monitoring

Product / Phase

A Wistech Paradigm for Intelligent Systems

125

Summary

Spider

Conceptual Clustering

Adaptive Wisdom Adaptive Wisdom Extraction Assistant

automatic summarizing of a document and groups of documents relating to brand, carried out from the perspective of the following groups of users: brand owners, detectives looking for frauds, buyers

automatic searching and downloading of documents relating to brand, carried out with particular emphasis on the needs of the following groups of users: brand owners, detectives looking for frauds, buyers

conceptual clustering of documents relating to brand with particular emphasis on the specific nature of queries submitted by the following types of users: brand owners, detectives looking for frauds, buyers

extracting data structures, information structures, knowledge structures and adaptive wisdom structures from documents, in particular generating thesauruses, conceptual hierarchies and constraints to acceptable solutions in brand monitoring domain

brand monitoring related to user query / answering processing in order to support users in solving their problems as effectively as possible

automatic summarizing of a document and groups of documents relating to communication , carried out from the perspective of people looking for optimal connections

automatic searching and downloading of documents relating to communication carried out with particular emphasis on the needs of the following groups of users: people looking for optimal connections

conceptual clustering of documents relating to communication with particular emphasis on the specific nature of queries submitted by the following types of users: people looking for optimal connections

extracting data structures, information structures, knowledge structures and adaptive wisdom structures from documents, in particular generating thesauruses, conceptual hierarchies and constraints to acceptable solutions in communication domain

communication related to user query / answering processing in order to support users in solving their problems as effectively as possible

Fig. 17. Functionality of individual products (continued)

This initial selection for the product list generates several dozen of products that would be the eﬀect of work on the individual phases of implementing each of the products, i.e., Summary, Spider, Conceptual Clustering, Wisdom Extraction, Wisdom Assistant. We present the proposed products in Figure 15. The scope of the program described in this paper should be considered as dynamic and more as a basis for further discussion than a ﬁnal version of the speciﬁc deﬁnitions of the projects. This is why the innovative ideas presented and the vision for their implementation do not contain any detailed cost beneﬁts analysis. It will only be possible to specify revenues, costs and cash ﬂow forecasts with any accuracy after the planned scope of work and the role of the WN has stabilized. As there are as yet no ﬁnal decisions on the scope of operations or role of the WN, this means that at the current stage it is impossible to

126

A. Jankowski and A. Skowron

precisely estimate the planned requirements for human resources. This is why in this document we present only the general human resources requirements and a description of the general mechanisms for acquiring these resources to implement WN. Each of these products would have their own individual functionality which would result from adapting the wisdom engine to the speciﬁc characteristics of their specialist ﬁelds. Figure 16 and Figure 17 show the functionality of the individual products.

5

Conclusions

We have discussed the main features of wistech and its importance for further progress in the development of intelligent systems. The proposed approach is based on Rough Granular Computing (RGC). One of the central problems of science today is to develop methods for approximation of compound vague concepts and approximate reasoning about them [32,81]. Today, we do not have yet satisfactory tools for discovery of relevant patterns for approximation of compound concepts directly from sample objects. However, we have developed methods for compound concept approximation using sample objects and domain knowledge acquired from experts (this is the approach pioneered by Zdzislaw Pawlak in [73]). The performed experiments based on approximation of concept ontology (see, e.g., [3,5,6,7,8,22,68,69,70], [78,79,93,94,95,98,99], [100,101,102,105,106]) showed that domain knowledge enables to discover relevant patterns in sample objects for compound concept approximation. Our approach to compound concept approximation and approximate reasoning about compound concepts is based on the rough-granular approach. One of the RGC challenges is to develop approximate reasoning techniques for reasoning about dynamics of distributed systems of judges. These techniques should be based on systems of evolving local perception logics rather than on a global logic [94,95]. Approximate reasoning about global behavior of judges’ system is infeasible without methods for approximation of compound vague concepts and approximate reasoning about them. One can observe here an analogy to phenomena related to the emergent patters in complex adaptive systems [21]. Let us observe that judges can be organized into a hierarchical structure, i.e., one judge can represent a coalition of judges in interaction with other agents existing in the environment [2,56,62]. Such judges representing coalitions play an important role in hierarchical reasoning about behavior of judges’ populations. Strategies for coalition formation and cooperation [2,62,64] are of critical importance in designing systems of judges with dynamics satisfying to a satisfactory degree a given speciﬁcation. Developing strategies for discovery of information granules representing relevant (for the given speciﬁcation) coalitions and cooperation protocols is another challenge for RGC.

A Wistech Paradigm for Intelligent Systems

127

RGC will become more and more important for analysis and synthesis of the discussed compound adaptive processes.The impact of RGC on real-life applications will be determined by techniques based on the rough-granular approach to modeling of relevant computations on compound information granules and methods for approximate reasoning about complex adaptive processes over such information granules. RGC techniques for modeling of complex processes will also have impact on the development of new non-conventional computation models.

Acknowledgments The research of Andrzej Jankowski was supported by Institute of Decision Process Support. The research of Andrzej Skowron has been supported by the grant 3 T11C 002 26 from Ministry of Scientiﬁc Research and Information Technology of the Republic of Poland. Many thanks to Professors James Peters and Anna Gomoli´ nska for their incisive comments and for suggesting many helpful ways to improve this article.

References 1. G. Antoniou, F. van Harmelen, A Semantic Web Primer (Cooperative Information Systems) The MIT Press, 2004. 2. R. Axelrod, The Complexity of Cooperation, Princeton, NJ: Princeton University Press, 1997. 3. A. Bargiela, W. Pedrycz, Granular Computing: An Introduction, Dordrecht: Kluwer Academic Publishers, 2003. 4. J. Barwise, J. Seligman, Information Flow: The Logic of Distributed Systems, Cambridge University Press, 1997. 5. J. G. Bazan, J. F. Peters, and A. Skowron, Behavioral pattern identiﬁcation through rough set modelling, in [71], pp. 688–697. 6. J. Bazan, A. Skowron, On-line elimination of non-relevant parts of complex objects in behavioral pattern identiﬁcation, in [111], pp. 720–725. 7. J. Bazan, A. Skowron, Classiﬁers based on approximate reasoning schemes, in Monitoring, Security, and Rescue Tasks in Multiagent Systems (MSRAS 2004), B. Dunin-Keplicz, A. Jankowski, A. Skowron, and M. Szczuka, Eds., Advances in Soft Computing, pp. 191-202, Heidelberg: Springer, 2005. 8. J. Bazan, A. Skowron, R. Swiniarski, Rough sets and vague concept approximation: From sample approximation to adaptive learning, Transactions on Rough Sets V: Journal Subline, Lecture Notes in Computer Science, vol. 3100, pp. 39-63, Heidelberg: Springer, 2006. 9. R. Baeza-Yates, B. Ribeiro-Neto, Modern Information Retrieval, Addison Wesley, 1999. 10. M. W. Berry, Survey of Text Mining : Clustering, Classiﬁcation, and Retrieval, Springer, 2003. 11. R. Brachman, H. Levesque, Knowledge Representation and Reasoning, Morgan Kaufmann, 2004. 12. L. Breiman, Statistical modeling: The two Cultures, Statistical Science 16(3) (2001) 199–231.

128

A. Jankowski and A. Skowron

13. S. Brin, L. Page, The Anatomy of a Large-Scale Hypertextual Web Search Engine, Stanford University, 1998. 14. C. Carpineto, G. Romano, Concept Data Analysis: Theory and Applications, John Wiley & Sons, 2004. 15. N. L. Cassimatis, A cognitive substrate for achieving human-level intelligence, AI Magazine 27(2) (2006) 45-56. 16. N. L. Cassimatis, E. T. Mueller, P. H. Winston, Achieving human-level intelligence through integrated systems and research, AI Magazine 27(2) (2006) 12-14. 17. S. Chakrabarti, Mining the Web: Analysis of Hypertext and Semi Structured Data, The Morgan Kaufmann Series in Data Management Systems, Morgan Kaufmann, 2002. 18. Mu-Y. Chen, An-P. Chen, Knowledge management performance evaluation: a decade review from 1995 to 2004, Journal of Information Science 32 (1) 2006 17-38. 19. H. W. Chesbrough, Open Innovation: The New Imperative for Creating and Profiting from Technology, Cambridge MA: Harvard Business School Publishing, 2003. 20. J. Coleman, Introducing Speech and Language Processing (Cambridge Introductions to Language and Linguistics), Cambridge University Press, 2005. 21. A. Desai, Adaptive complex enterprises, Communications ACM 48(5) (2005) 3235. 22. P. Doherty, W. L ukaszewicz, A. Skowron, and A. Szalas, Knowledge Engineering: A Rough Set Approach, Studies in Fuzziness and Soft Computing, vol. 202, Heidelberg: Springer, 2006. 23. R. Dornfest, T. Calishain, Wistech Network Hacks O’Reilly Media, Inc.; 2004. 24. R. Duda, P. Hart, and R. Stork, Pattern Classiﬁcation, New York, NY: John Wiley & Sons, 2002. 25. B. Dunin-K¸eplicz, A. Jankowski, A. Skowron, M. Szczuka, Monitoring, Security, and Rescue Tasks in Multiagent Systems (MSRAS’2004), Series in Soft Computing, Heidelberg: Springer, 2005. 26. A. E. Eiben, J. E. Smith, Introduction to Evolutionary Computing, Natural Computing Series, Springer, 2003. 27. E. Feigenbaum, J. Feldman (Eds.), Computers and Thought, New York: McGraw Hill, 1963. 28. S. Feldman, Why Search is Not Enough (white paper), IDC, 2003. 29. S. Feldman, Enterprise Search Technology: Information Disasters and the High Cost of Not Finding Information (Special IDC Report), IDC, 2004. 30. D. Fensel, Ontologies: A Silver Bullet for Knowledge Management and Electronic Commerce, Springer, 2003. 31. K. D. Forbus, T. R. Hinrisch, Companion cognitive systems: A step toward human-level AI, AI Magazine 27(2) (2006) 83-95. 32. M. Gell-Mann, The Quark and the Jaguar, NY: Freeman and Co., 1994. 33. J. H. Friedman, T. Hastie, R. Tibshirani, The Elements of Statistical Learning: Data Mining, Inference, and Prediction, Heidelberg: Springer, 2001. 34. M. Ghallab, D. Nau, and P. Traverso, Automated Planning: Theory and Practice, CA: Morgan Kaufmann, 2004. 35. S. Ghemawat, H. Gobioﬀ, Shun-Tak Leung, The Wistech Network File System, Wistech Network, 2005. 36. D. E. Goldberg, Genetic Algorithms in Search, Optimization, and Machine Learning, Addison-Wesley Professional, 1989.

A Wistech Paradigm for Intelligent Systems

129

37. A. Gomez-Perez, O. Corcho, M. Fernandez-Lopez, Ontological Engineering with examples from the areas of Knowledge Management, e-Commerce and the Semantic Web (Advanced Information and Knowledge Processing), Springer, 2004. 38. R. Granger, Engines of the brain: The computational instruction set of human cognition, AI Magazine 27(2) (2006) 15-31. 39. S. Grimes, The Developing Text Mining Market, A white paper prepared for Text Mining Summit 2005, Boston, June 7-8 2005, Alta Plana Corporation, 2005. 40. A. OHagan, C. E. Buck, A. Daneshkhah, J. R. Eiser, P. H. Garthwaite, D. J. Jenkinson, J. E. Oakley, T. Rakow, Uncertain Judgements: Eliciting Expert Probabilities, Wiley, New York, 2006. 41. J. Heaton, Programming Spiders, Bots, and Aggregators in Java, Sybex, 2002. 42. K. Hemenway, T. Calishain, Spidering Hacks O’Reilly Media, Inc.; 2003. 43. M. Henzinger, S. Lawrence, Extracting knowledge from the World Wide Web, Wistech Network, 2004. 44. M. Hirvensalo, Quantum Computing, Springer-Verlag, Heidelberg 2001. 45. M. N. Huhns, M. P. Singh, Readings in Agents, Morgan Kaufmann, 1998. 46. P. Jackson, I. Moulinier, Natural Language Processing for Online Applications: Text Retrieval, Extraction, and Categorization (Natural Language Processing, 5), John Benjamins Publishing Co, 2002. 47. Z. Janiszewski, On needs of mathematics in Poland (O potrzebach matematyki w Polsce) (in Polish), In: Nauka Polska. Jej Potrzeby, Organizacja i Rozw´ oj, Warszawa, 1918; see also reprint in Wiadomo´sci Matematyczne VII (1963) 3-8. 48. A. Jankowski, An alternative characterization of elementary logic, Bull. Acad. Pol. Sci., Ser. Math. Astr. Phys.XXX (1-2) (1982) 9-13. 49. A. Jankowski, Galois structures, Studia Logica 44(2) (1985) 109-124. 50. S. Johnson, Dictionary of the English Language in Which the Words are Deduced from Their Originals, and Illustrated in their Diﬀerent Signiﬁcations by Examples from the Best Writers, 2 Volumes. London: F.C. and J. Rivington, 1816. 51. R. M. Jones, R. E. Wray, Comparative analysis of frameworks for knowledgeintensive intelligent agents, AI Magazine 27(2) (2006) 57-70. 52. D. Jurafsky, J. H. Martin, Speech and Language Processing: An Introduction to Natural Language Processing, Computational Linguistics and Speech Recognition, Prentice Hall, 2000. 53. D. Kahneman, P. Slovic, A. Tversky, A. (Eds.). Judgement under Uncertainty: Heuristics and Biases, Cambridge University Press, New York, 1982. 54. R. Keefe, Theories of Vagueness, Cambridge, UK: Cambridge Studies in Philosophy, 2000. ˙ 55. W. Kloesgen, J. Zytkow, Handbook of Knowledge Discovery and Data Mining, New York: Oxford University Press, 2002. 56. S. Kraus, Strategic Negotiations in Multiagent Environments, Massachusetts: The MIT Press, 2001. 57. J. Lambek, P. J. Scott, Introduction to Higher-Order Categorical Logic (Cambridge Studies in Advanced Mathematics 7), Cambridge University Press, 1986. 58. M. Lamb, Build Your Own Army of Web Bots Within 24 Hours (Army of Web Bots Series) Authorhouse, 2003. 59. P. Langley, H. A. Simon, G. L. Bradshaw, J. M. Zytkow, Scientiﬁc Discovery: Computational Explorations of the Creative Processes, MIT Press, 1987. 60. P. Langley, Cognitive architectures and general intelligent systems, AI Magazine 27(2) (2006) 33-44. 61. G. W. Leibniz, New Essays on Human Understanding, Cambridge UP, 1982.

130

A. Jankowski and A. Skowron

62. J. Liu, Autonomous Agents and Multi-Agent Systems: Explorations in Learning, Self-Organization and Adaptive Computation, Singapore: World Scientiﬁc Publishing, 2001. 63. J. Liu, L. K. Daneshmend, Spatial Reasoning and Planning: Geometry, Mechanism, and Motion, Springer, 2003, Hardcover. 64. J. Liu, X. Jin, K. Ch. Tsui, Autonomy Oriented Computing: From Problem Solving to Complex Systems Modeling, Heidelberg: Kluwer Academic Publisher/Springer, 2005. 65. S. MacLane, I. Moerdijk, Sheaves in Geometry and Logic: A First Introduction to Topos Theory (Universitext), Springer, 1994. 66. A. Madhavapeddy, N. Ludlam, Ubiquitious Computing needs to catch up with Ubiquitous Media, University of Cambridge Computer Laboratory, Interceptor Communications Ltd., 2005. 67. C. D. Manning, H. Sch¨ utze, H., Foundations of Statistical Natural Language Processing, The MIT Press, 1999. 68. S. H. Nguyen, J. Bazan, A. Skowron, and H. S. Nguyen, Layered learning for concept synthesis, Transactions on Rough Sets I: Journal Subline, Lecture Notes in Computer Science, vol. 3100, pp. 187-208, Heidelberg: Springer, 2004. 69. T. T. Nguyen, Eliciting domain knowledge in handwritten digit recognition, in [71] 762-767. 70. S. K. Pal, L. Polkowski, and A. Skowron (Eds.), Rough-Neural Computing: Techniques for Computing with Words, Cognitive Technologies, Heidelberg: SpringerVerlag, 2004. 71. S. K. Pal, S. Bandoyopadhay, and S. Biswas (Eds.), Proceedings of the First International Conference on Pattern Recognition and Machine Intelligence (PReMI’05), December 18-22, 2005, Indian Statistical Institute, Kolkata, Lecture Notes in Computer Science vol. 3776, Heidelberg: Springer, 2005. 72. T. B. Passin, Explorer’s Guide to the Semantic Web Mining Publications, 2004. 73. Pawlak, Z.: Classiﬁcation of objects by means of attributes, Research Report PAS 429, Institute of Computer Science, Polish Academy of Sciences, ISSN 138-0648, January (1981). 74. Z. Pawlak, Rough Sets: Theoretical Aspects of Reasoning about Data, System Theory, Knowledge Engineering and Problem Solving 9, Dordrecht: Kluwer Academic Publishers, 1991. 75. Z. Pawlak, A. Skowron, Rudiments of rough sets. Information Sciences. An International Journal. 177(1) (2007) 3-27. 76. Z. Pawlak, A. Skowron, Rough sets: Some extensions. Information Sciences. An International Journal. 177(1) (2007) 28-40. 77. Z. Pawlak, A. Skowron, Rough sets and Boolean reasoning. Information Sciences. An International Journal. 177(1) (2007) 41-73. 78. J. F.Peters, Rough ethology: Toward a biologically-inspired study of collective behavior in intelligent systems with approximation spaces. Transactions on Rough Sets III: Journal Subline, Lecture Notes in Computer Science, vol. 3400, pp. 153174, Heidelberg: Springer, 2005. 79. J. F. Peters, C. Henry, C., Reinforcement learning with approximation spaces. Fundamenta Informaticae 71(2-3) (2006) 323-349. 80. S. Plous, The Psychology of Judgement and Decision Making, McGraw-Hill, New York, 1993. 81. T. Poggio, S. Smale, The mathematics of learning: Dealing with data, Notices of the AMS 50(5) (2003) 537-544.

A Wistech Paradigm for Intelligent Systems

131

82. L. Polkowski, A. Skowron, Rough mereology: A new paradigm for approximate reasoning, International Journal of Approximate Reasoning 15 (1996) 333-365. 83. L. Polkowski, S. Tsumoto, T. Y. Lin (Eds.), Rough Set Methods and Applications: New Developments in Knowledge Discovery in Information Systems, Studies in Fuzziness and Soft Computing vol. 56, Physica-Verlag Heidelberg, 2000. 84. G. P´ olya, How to Solve It, 2nd ed., Princeton University Press, 1957; see also http://en.wikipedia.org/wiki/How to Solve It. 85. H. Rasiowa, W. Marek, On reaching consensus by groups of intelligent agents, In: Z. W. Ras (Ed.), Methodologies for Intelligent Systems, North-Holland, Amsterdam, 1989, 234-243. 86. H. Rasiowa, Algebraic Models of Logics, Warsaw University, 2001. 87. H. Rasiowa, R. Sikorski, The Mathematics of Metamathematics, Monograﬁe Matematyczne vol. 41, PWN Warsaw, 1963. 88. S. Read, Thinking about Logic. An Introduction to the Philosophy of Logic, Oxford, New York: Oxford University Press, 1995. 89. P. Saint-Dizier, E. Viegas, B. Boguraev, S. Bird, D. Hindle, M. Kay, D. McDonald, H. Uszkoreit, Y. Wilks, Computational Lexical Semantics (Studies in Natural Language Processing), Cambridge University Press, 2005. 90. C. Schlenoﬀ, J. Albus, E. Messina, A. J. Barbera, R. Madhavan, S. Balakirsky, Using 4D/RCS to address AI knowledge integration, AI Magazine 27(2) (2006) 71-81. 91. Scientiﬁc Datalink, The Scientiﬁc DataLink index to artiﬁcial intelligence research, 1954-1984, Scientiﬁc DataLink, 1985. 92. Scientiﬁc Datalink, The Scientiﬁc DataLink index to artiﬁcial intelligence research, 1985 Supplement, Scientiﬁc DataLink, 1985. 93. A. Skowron, Approximate reasoning in distributed environments, in N. Zhong, J. Liu (Eds.), Intelligent Technologies for Information Analysis, Heidelberg: Springer, pp. 433-474. 94. A. Skowron, Perception logic in intelligent systems (keynote talk), In: S. Blair et al (Eds.), Proceedings of the 8th Joint Conference on Information Sciences (JCIS 2005), July 21-26, 2005, Salt Lake City, Utah, USA, X-CD Technologies: A Conference & Management Company, ISBN 0-9707890-3-3, 15 Coldwater Road, Toronto, Ontario, M3B 1Y8, 2005, pp. 1-5. 95. A. Skowron, Rough sets in perception-based computing (keynote talk), in [71], pp. 21-29. 96. A. Skowron, R. Agrawal, M. Luck, T. Yamaguchi, O. Morizet-Mahoudeaux, J. Liu, N. Zhong (Eds.), Proceedings of the 2005 IEEE/WIC/ACM International Conference on WEB Intelligence, Compiegne, France, September 19-22, 2005, IEEE Computer Society Press, Los Alamitos, CA, 2005, pp. 1-819. 97. A. Skowron, J.-P. Barthes, L. Jain, R. Sun, P. Morizet-Mahoudeaux, J. Liu, N. Zhong (Eds.), Proceedings of the 2005 IEEE/WIC/ACM International Conference on Intelligent Agent Technology, Compiegne, France, September 19-22, 2005, IEEE Computer Society Press, Los Alamitos, CA, 2005, pp. 1-766. 98. A. Skowron, J. Stepaniuk, Tolerance approximation spaces, Fundamenta Informaticae 27 (1996) 245-253. 99. A. Skowron, J. Stepaniuk, Information granules: Towards foundations of granular computing, International Journal of Intelligent Systems 16(1) (2001) 57-86. 100. A. Skowron, J. Stepaniuk, Information granules and rough-neural computing, in [70], pp. 43-84. 101. A. Skowron, P. Synak, Complex patterns, Fundamenta Informaticae 60(1-4) (2004) 351-366.

132

A. Jankowski and A. Skowron

102. A. Skowron, R. Swiniarski, and P. Synak, Approximation spaces and information granulation, Transactions on Rough Sets III: Journal Subline, Lecture Notes in Computer Science, vol. 3400, pp. 175-189, Heidelberg: Springer, 2005. 103. J. F. Sowa, Knowledge Representation: Logical, Philosophical, and Computational Foundations, Course Technology, 1999. 104. S. Staab, R. Studer, Handbook on Ontologies, in International Handbooks on Information Systems, Heidelberg: Springer 2004. 105. J. Stepaniuk, J. Bazan, and A. Skowron, Modelling complex patterns by information systems, Fundamenta Informaticae 67 (1-3) (2005) 203-217. 106. J. Stepaniuk, A. Skowron, J. Peters, and R. Swiniarski, Calculi of approximation spaces, Fundamenta Informaticae 72 (1-3) (2006) 363-378. 107. P. Stone, Layered Learning in Multi-Agent Systems: A Winning Approach to Robotic Soccer, Cambridge, MA: The MIT Press, 2000. 108. W. Swartout, J. Gratch, R. W. Hill, E. Hovy, S. Marsella, J. Rickel, D. Traum, Towards virtual humans, AI Magazine 27(2) (2006) 96-108. 109. R. Sun (Ed.), Cognition and Multi-Agent Interaction. From Cognitive Modeling to Social Simulation. New York, NY: Cambridge University Press, 2006. 110. K. Sycara, Multiagent systems, in AI Magazine, Summer 1998, 79-92. ´ ezak, J. T. Yao, J. F. Peters, W. Ziarko, and X. Hu (Eds.), Proceedings of 111. D. Sl¸ the 10th International Conference on Rough Sets, Fuzzy Sets, Data Mining, and Granular Computing (RSFDGrC’2005), Regina, Canada, August 31-September 3, 2005, Part II, Lecture Notes in Artiﬁcial Intelligence, vol. 3642, Heidelberg: Springer, 2005. 112. A. S. Troelstra, H. Schwichtenberg, Basic Proof Theory, Cambridge University Press, 2000. 113. A. S. Troelstra, D. Van Dalen, Constructivism in Mathematics: An Introduction, Studies in Logic and the Foundations of Mathematics Vol. 1 & 2, Elsevier Science Publishing Company, 1988. 114. A. Turing, Computing machinery and intelligence, Mind LIX(236) (October 1950) 433-460. 115. W. Van Wezel, R. Jorna, and A. Meystel, Planning in Intelligent Systems: Aspects, Motivations, and Methods. Hoboken, New Jersey: John Wiley & Sons, 2006. 116. V. Vapnik, Statistical Learning Theory, New York: John Wiley & Sons, 1998. 117. S. Weiss, N. Indurkhya, T. Zhang, F. Damerau, Text Mining: Predictive Methods for Analyzing Unstructured Information, Springer, 2004. 118. I. H. Witten, E. Frank, Data Mining: Practical Machine Learning Tools and Techniques, Second Edition, Morgan Kaufmann Series in Data Management Systems, Morgan Kaufmann, 2005. 119. I. H. Witten, A. Moﬀat, T. C. Bell, Managing Gigabytes: Compressing and Indexing Documents and Images, The Morgan Kaufmann Series in Multimedia and Information Systems, Morgan Kaufmann, 1999. 120. L. A. Zadeh, Fuzzy sets, Information and Control 8 (1965) 333-353. 121. L. A. Zadeh, From computing with numbers to computing with words - from manipulation of measurements to manipulation of perceptions, IEEE Transactions on Circuits and Systems - I: Fundamental Theory and Applications 45(1) (1999) 105-119. 122. L. A. Zadeh, A new direction in AI: Toward a computational theory of perceptions, AI Magazine 22 (1) (2001) 73-84. 123. W. Ziarko, Variable precision rough set model, Journal of Computer and System Sciences 46 (1993) 39-59.

The Domain of Acoustics Seen from the Rough Sets Perspective Bozena Kostek Multimedia Systems Department, Gdansk University of Technology and Excellence Center Communication Process: Hearing and Speech, PROKSIM, Warsaw, Poland [email protected]

Abstract. This research study presents rough set-based decision systems applications to the acoustical domain. Two areas are reviewed for this purpose, namely music information classiﬁcation and retrieval and noise control. The main aim of this paper is to show results of both measurements of the acoustic climate and a survey on noise threat, conducted in schools and students’ music clubs. The measurements of the acoustic climate employ multimedia noise monitoring system engineered at the Multimedia Systems Department of the Gdansk University of Technology. Physiological eﬀects of noise exposure are measured using pure tone audiometry and otoacoustic emission tests. All data are gathered in decision tables in order to explore the signiﬁcance of attributes related to hearing loss occurence and subjective factors that attribute to the noise annoyance. Future direction of experiments are shortly outlined in Summary.

1

Opening Thoughts

Before introducing the particular topic of research presented in the paper, I would like to share a few thoughts. This Section is devoted to some personal aspects of the research carried out by the author for many years. It concerns the fascination of the rough set methodology and the philosophy that lies behind it, and also (or rather in the ﬁrst place) the fascination of the rough set method creator, Professor Zdzislaw Pawlak [34,37]. His personality stands out very clearly amongst other researchers. It happened that his plenary talk I’ve listened to on the occasion of the 2nd International Conference on Rough Sets in Banﬀ guided me toward new interests, namely the applications of decisions rule-based systems which are formidably ﬁtted for uncertainty so often found in acoustics and its analysis. From this time on, we have met many times on various occasions, and I was always inspired by his presentations that led me into new directions and horizons. Professor Pawlak was a mentor to me and I beneﬁted greatly because he was very kind to write Foreword for my two books showing his interest in the research carried out by me and my colleagues. These books perhaps would not happen without his wise patronage. J.F. Peters et al. (Eds.): Transactions on Rough Sets VI, LNCS 4374, pp. 133–151, 2007. c Springer-Verlag Berlin Heidelberg 2007

134

B. Kostek

This is very valuable to me and I will be always grateful to him. Altogether within the rough set society, a clear interest appeared to pursue the rough set-acoustic applications [7,21,22], especially the domain of music evoked many research studies [2,3,8,9,12,13,16,17,18,21,22,23,24,25,29,30,31,40,41,42,43]. For some years now, many researches have published in the joint area of rough sets and acoustic/music, thus some of these names being recalled in References. Lately, also Chinese and Korean contributors to this domain appeared. As a result of the interest in this area a new domain of applications emerged, which focuses on interests such as musical instrument recognition based on timbre descriptors, musical phrase classiﬁcation based on its parameters or contour, melody classiﬁcation (e.g. query-by-humming systems), rhythm retrieval (diﬀerent approaches), high-level-based music retrieval such as looking for emotions in music or diﬀerences in expressiveness, music search based on listeners’ preferences, and others. One may also ﬁnd research studies which try to correlate low-level descriptor analysis to high-level human perception. The semantic description is becoming a basis of the next web generation, i.e., the Semantic Web. Several important concepts have been introduced recently by the researchers associated with the rough set community with regard to semantic data processing including techniques for computing with words [20,33]. Moreover, Zdzislaw Pawlak in his papers [35,36] promoted his new mathematical model of ﬂow networks which can be applied to mining knowledge in databases. Such topics are reﬂected also in papers that followed Prof. Pawlak’s original idea on ﬂow graphs [9,26]. Studies performed on the verge of two domains: soft computing (and particularly rough sets) and acoustics enabled the author to apply for many research grants and many of these projects have been successfully awarded. Once again, the current research is also a good example of the need for employing decision systems to the area which at ﬁrst glance seems far away from the soft computing interests.

2

Introduction

This paper deals with a particular topic which is noise threat-related. As indicated in numerous reports, noise threats occur very frequently nowadays. Occupational exposure limits (OELs) for noise are typically given as the maximum duration of exposure permitted for various noise levels. Environmental noise regulations usually specify a maximum outdoor level of 60 to 65 dB(A), while occupational safety organizations recommend that the maximum exposure to noise is 40 hours per week at 85 to 90 dB(A). For every additional 3 dB(A), the maximum exposure time is reduced by a factor of 2, e.g. 20 hours per week at 88 dB(A). Sometimes, a factor of 2 per additional 5 dB(A) is used. However, these occupational regulations are recognized by the health literature as inadequate to protect against hearing loss and other health eﬀects, especially for sensitive individuals, adverse subjective eﬀects might be expected to appear earlier than for others [4,5,38].

The Domain of Acoustics Seen from the Rough Sets Perspective

135

The background of this study is the fact that younger and younger people experience a noticeable loss in hearing. In previous decades a procedure was established within the audiology ﬁeld, that a group of young males of the age between 18-21 could constitute a reference group for hearing measurements. However, during the last decade numerous studies have shown that this statement is no longer valid. Also, hearing characteristics of students measured during psychoacoustic laboratory sessions at the Multimedia Systems Department have shown that students of this age typically have a threshold shift at 6 kHz. On average, this accounts to 20 dB HL (hearing loss), which is for this age rather unexpected taking into account that students that have any history of ear illnesses were excluded from the experiments. That is why the starting point is to look for the causes of loss in hearing in younger groups of population. The study aimed at showing results of a survey on noise threat which was conducted in schools and students’ music clubs. Noise has an enormous impact on health and life quality of human beings. Noise pollution in Poland is greater than in others UE countries, moreover recently it has been reported to be on the increase [14]. Taking into account the European 2002/49/WE directive related to the control and assessment of environmental noise, monitoring of these threats becomes a necessity [38]. That is why a thorough study on many aspects of noise was envisioned and is carried out for some time at the Multimedia Systems Department [6,10,27,28]. First of all, measurements of the acoustic climate that employed telemetry stations for continuous noise monitoring engineered at the Multimedia Systems Department were conducted. Also, physiological eﬀects of noise were measured among pupils and students. Hearing tests were performed twice, before and after the exposure to noise. For this purpose a so-called distortion product otoacoustic emission method (DPOAE) was utilized. As derived from numerous studies, otoacoustic emission is treated as an early indicator of the occurrence of hearing loss for which reason this method was chosen. The obtained results of noise measurements revealed that an unfavorable noise climate was found in examined schools and music clubs. This was also conﬁrmed by the results of a subjective examination. For the latter purpose students and pupils ﬁlled in a questionnaire expressing their feelings as to noise presence and its annoyance. The noise dose analysis based on average time spent by pupils in schools was also calculated. It revealed that noise in schools did not constitute a risk to the pupils’ hearing system, however, it may be considered as an essential source of annoyance. On the other hand, noise in music clubs surpassed all permitted noise limits, thus could be treated as dangerous to hearing. Hearing tests revealed changes in the cochlea activity of examined students, also the Tinnitus (ringing in the ear) eﬀect was experienced temporarily. In addition, noise annoyance and noise threat criteria and analysis were proposed and veriﬁed based on the acquired and analyzed data. All factors recognized in the study constitute the basis of two types of decision tables that were created. The ﬁrst one consists of the attributes derived from the measurements and calculation of the noise dose, also the presence or absence of the Tinnitus (ringing in the ear) eﬀect is included in this table. The second

136

B. Kostek

decision table gathers data from the survey on noise annoyance. The conditional attributes are subject-driven in this case. Examples of the questions included in this survey are shown in the following sections. The paper aims at showing that a complex and thorough study may lead to better understanding of noise threats and the correlation between the measurement data and the survey responses concerning noise annoyance. Another aim is to show that current regulations are not adequate to predict PTS (Permanent Threshold Shift) early. Finally the data dependency is analyzed, in the reduced database, to ﬁnd the minimal subset of attributes called reduct. The analysis of collected data is done by employing the rough set decision system.

3

Multimedia Noise Monitoring System

The MNMS (Multimedia Noise Monitoring System), developed at the Multimedia Systems Department of the Gdansk University of Technology enables to proceed with the environmental noise measurements in cities on an unparalleled scale now. In general, the MNMS consists of a central database which serves as a repository of measurement results, and numerous equipment tools which execute noise meter functions. One of the proposed devices is a mobile noise monitoring station. The station realizes all measuring functions typical for a sound level meter. It also includes special solutions for long-term measurements and introduces a new type of noise indicators. The application of wireless data transmission technology enables to send data to the server and to remotely control the performance of the station. Since this subject was already published [6], thus its main features are only outlined above.

4

Noise and Hearing Measurements

The noise measurement results, obtained by means of the MNMS, are presented below. The measurements were done in selected schools, musical clubs, and during a musical band rehearsals. Participation in music bands concerts and staying in students’ clubs are a common way of entertainment amongst students. This is why the investigation was carried out also in these locations. The acquired data were utilized to perform the noise dose analysis. This is done to determine the noise exposure of a person staying in the considered places. In selected cases (i.e. schools and musical clubs), the noise dose analysis was expanded by the assessment of hearing. To achieve this, a so-called distortion product otoacoustic emission (DPOAE) measurement and pure tone audiometry were applied. Hearing was examined twice. First, directly before the exposure to noise of a given type, and then immediately after. The performed analysis combined the obtained noise and hearing measurement results. Hearing examinations employed the DPOAE method using GSI 60 DPOAE system. The following parameters of the stimuli were used during tests: L1 equals 65 dB, L2 equals 55 dB, f2/f1 = 1.2, DP frequency (geometric mean): 1062, 1312, 1562, 1812, 2187, 2625, 3062, 3687, 4375, 5187, 6187, 7375 Hz. A DP signal level

The Domain of Acoustics Seen from the Rough Sets Perspective

137

and a noise ﬂoor for every stimuli were registered. The test result was accepted if the diﬀerence between the evoked otoacoustic emission signals and the noise ﬂoor was not less than 10 dB. For pure tone audiometry only selected frequencies were examined: 1000, 1500, 2000, 3000, 4000, 6000, 8000 Hz. The stimuli for each frequency were presented starting from the minimal loudness. The reason of such a selection of parameters was because the noise impact on the hearing system is the strongest for middle and high frequencies. The test was carried out in rooms specially adapted for this purpose. Some measurements performed in schools were interfered with sounds coming from adjoining rooms. Typical response to stimuli is shown in Figure 1. The DP otoacoustic response was found at 3187 Hz, and because the diﬀerence between the noise level and DP signal is larger than 10 dB, thus this measurement is accepted.

Fig. 1. Example of the DPOAE analysis

The following noise parameters LAF min , LAeq , LAmax (see the Equation and deﬁnitions below) were measured independently over broadband and in one-third octave bands. A histogram of time history of LAF instantaneous levels was also calculated. A measuring microphone was located 1.9 m above the ﬂoor level for every measurement. For all measuring series, a place where people gather most often was selected. This was to determine correctly a real noise dose to which they were exposed. N 1 0.1·LAdBi 10 (1) Leq = 10 log N i=1 where: Leq – A-weighted equivalent continuous noise level, N –number of LAdBi values, LAdBi – A-weighted instantaneous sound levels,

138

B. Kostek

On the other hand, LAF min , LAF max denote the lowest and highest A-weighted sound levels for fast time weighting that occurred during the measurement. 4.1

Noise Dose Analysis

The evaluation of both the occupational noise exposure and the risk of developing a permanent hearing loss that may result from the noise exposure are shown in Fig. 2.

Fig. 2. Evaluation of the occupational noise exposure (LEX,8h (D) – daily noise exposure level, LAmax – maximum sound level in working conditions, LCpeak – peak sound level)

The presented evaluation of the occupational noise exposure and the risk of developing a permanent hearing loss is based on noise regulations [4]. The regulations recommend a limit for occupational noise exposure of 85 decibels, A-weighted, as an 8-hour time-weighted average [85 dBA as an 8-hr TWA]). This enables to evaluate whether occupational noise may cause hearing loss, and (or) whether personal a hearing protector (i.e. earmuﬀs, formable earplugs, earplugs, etc.) is required. The A ﬁlter is a weighting curve which approximates equal loudness perception characteristics of the human hearing for pure tones with reference to 40 dB SPL at 1 KHz. It is worth noticing that this curve was established for pure tones, and a potential noise is typically broadband, which means that A-weighting may not provide the best estimate of potential noiseinduced hearing loss. The formal deﬁnition of the noise dose deﬁnes the dose as the amount of actual exposure relative to the amount of allowable exposure, and for which 100% and above represents exposures that are hazardous. The noise dose is calculated according to the following formula:

The Domain of Acoustics Seen from the Rough Sets Perspective

D = [C1 /T1 + C2 /T2 + ... + Cn /Tn ]100%

139

(2)

where D is a dose in the allowable percent, Cn refers to the total time of exposure at a speciﬁed noise level, and Tn denotes the exposure time at which noise for this level becomes hazardous. This deﬁnition allows to calculate a T W A (Timeweighted average) from a Noise Dose: T W A = 90 + 16.61log(D/100)

(3)

where T W A is the 8-hour time-weighted average noise exposure, and D denotes the dose. Unfortunately, the deﬁnition given above is arguable since other formulas of the T W A are also known. The above regulations are recalled before the noise dose calculated for this study is presented, because it is to show that the norms may not provide the best estimate of the potential occurrence of the permanent hearing loss. One may ﬁnd also such conclusions in many research studies [1,15,11,32,39]. Time of the noise exposure for the presented activities is much longer in real conditions than in a time-controlled experiment. A simple survey that included questions about how long pupils/students stay in clubs, play or listen to loud music, stay in school, etc. was also carried out. On the basis of the answers, an average time of the exposure for diﬀerent type of activities was speciﬁed. The total time of the noise exposure in schools, clubs and rehearsing musicians’ is respectively equal to 3600 s, 13500 s, and 5400 s. Based on the assumption that in the indicated places the noise climate is the same, it is possible to obtain the noise dose for people staying in these places. The noise dose for school amounts to not more than 26%, for rehearsing in a musical band to 673%, for club No.1 - 506% and for club No. 2 - 1191% of the daily dose. 4.2

Noise Investigation Results

The obtained noise measurement results are presented in Table 1. Noise investigation was performed in three diﬀerent schools. They diﬀered from each other in the age of the pupils. The youngest pupils attended a primary school (school No. 1). The second school was for children between the age of 13 and 15. The third school was a high school attended by the youth aged from 16 to 19. The biggest noise occurred at the primary school. This is because small children are the main source of noise in schools. They behave extremely vigorously at this age. This entailed a very high noise level. In this school, additional source of noise was loud music played from loudspeakers. In high school No. 3 the noise was produced by loud conversations. It should also be mentioned that in all investigated schools there was no suﬃcient absorption materials covering walls and ceilings, which fact further increased sound level. The fourth measuring series was done during a rehearsal of a small students’ music band. The band consisted of a drummer, a bass player and a keyboard player. This measurement revealed high dynamics of noise level. This is because the musicians often paused and

140

B. Kostek

Table 1. Noise measurement results. Time of noise exposure is expressed in seconds, noise levels in dB and noise dose in per cent of allowed daily noise dose. Measurem. No. Exposure LAF min 1 School No. 1 67.4 2 School No. 2 67.2 3 School No. 3 72.0 4 Music band 52.5 5 Club No. 1 76.2 6 Club No. 2 68.9

LAeq 89.0 85.5 83.6 100.5 95.3 99.0

LAF max 105.5 106.8 97.4 114.4 108.2 114.2

Exposure Time Noise Dose 600 5.2 900 3.5 600 1.5 4058 506.1 4529 169.9 5330 470.0

Fig. 3. LAF histogram measured during lessons

Fig. 4. LAF histogram measured during breaks

consulted each. The 5th and 6th measurement series were carried out in two students’ clubs. They diﬀered in the type of music. In the ﬁrst club a pop music was dominating, while in the second only rock was played. The results of the

The Domain of Acoustics Seen from the Rough Sets Perspective

141

noise dose analysis for the measured time exposures are presented in Table 1. Figures 3 and 4 show histograms measured during lessons and breaks in schools. The breaks were 15 minutes long. The results of noise measurements during two breaks were similar (see Fig. 4). Leq for the breaks was approx. 85dB. Sounds louder than 90 dB appeared at 3-4% of measured time span, however a noise level of 122 dB was also registered. On the other hand, noise during lessons turned out to be less annoying, Leq equaled to 62 dB. In all examined places, a very high noise level was observed. Staying for a long time in places such as music clubs, discotheques, etc. where noise level reaches the values shown in Table 1, can be hazardous to hearing (i.e. may produce a permanent threshold shift). 4.3

Hearing Measurement Results

Several dozens of persons took part in the presented hearing tests. The total number of people examined for diﬀerent types of exposure is presented in Table 2. Table 2 also includes the results of average changes of a hearing threshold (pure tone audiometry), and the results of the DPOAE tests. The average changes of a hearing threshold after the exposure to noise for individual frequencies obtained for pure tone audiometry are presented in Table 3. Two diﬀerent aspects were taken into consideration while analyzing the DPOAE results. First, the number Table 2. Results of hearing testing (in [%])

Measurem. No. 1 2 3 4 5 6

Pure tone audiometry No. of Decrease of Increase of persons threshold threshold 9 21.4 37.3 10 14.3 62.9 12 12.5 75.0

No change 41.3 22.8 12.5

DPOAE test results No. of +Pass -Pass persons 10 11.0 13.6 5 10.0 19.2 5 3.3 12.5 9 5.1 11.6 11 3.4 10.6 12 4.5 20.5

No change 75.4 70.8 84.2 83.3 86.0 75.0

Table 3. The average changes of hearing threshold for pure tone audiometry (in [dB]) Type of noise exposure Music band

Club No. 1

Club No. 2

L R AVG L R AVG L R AVG

1000 2.8 0.0 1.4 2.5 -1.5 0.5 2.9 3.8 3.3

1500 0.0 -1.1 -0.6 6.5 3 4.75 6.7 6.7 6.7

2000 2.2 1.1 1.7 7.5 7 7.25 8.8 10.0 9.4

3000 2.2 0.6 1.4 10.5 10.5 10.5 10.4 12.1 11.3

4000 3.3 2.2 2.8 11.5 10.5 11.0 15.9 12.9 14.4

6000 1.7 3.3 2.5 5 5.5 5.25 12.1 10.8 11.5

8000 1.1 0.5 0.8 1 3.5 2.25 -1.7 0.8 -0.4

142

B. Kostek

Table 4. The average changes of DP signal level for particular types of exposures (in [%]) Measurement No. 1 2 3 Increase 30.3 27.5 36.7 Decrease 28.1 30.0 34.2 No change 41.6 42.5 29.1

4 25.0 44.4 30.6

5 27.3 45.4 27.3

6 16.3 53.5 30.2

of “passed” and “failed” tests for the DPOAE examination were determined. The result of the ﬁrst examination served as reference. The symbol “+Pass” indicates that a pupil failed the ﬁrst examination and passed the second one. The symbol “–Pass” signiﬁes a reverse situation (a test passed in the ﬁrst examination and failed after the exposure to noise). The results are presented in Table 2, in the “DPOAE test results” column. The second kind of analysis determined how the DP signal level changed under the inﬂuence of the exposure to noise. The results of this analysis are presented in Table 4. As seen from the tables, in most cases the hearing threshold was increased. It means that almost every person had a TTS after noise exposure. Also some people reported the perception of the Tinnitus eﬀect as well. The most signiﬁcant TTS occurred for 4000 Hz. The data obtained from DPOAE tests conﬁrm negative after-eﬀects of the noise exposure analysis acquired by means of pure tone audiometry.

5

Psychophysiological Noise Dosimeter

Methods of the estimation of noise-induced hearing loss presented in this paper are based mainly on the equal energy hypothesis [5]. This approach focuses on an assessment of the quantity of energy which aﬀects the hearing system. The time characteristics of noise is neglected, and the main emphasis is placed on the

Fig. 5. General scheme of the psychophysiological noise dosimeter

The Domain of Acoustics Seen from the Rough Sets Perspective

143

assessment of the equivalent sound level value. However, in many cases this approach could be insuﬃcient. The analysis of numerous literature data, including testing of the exposure to noise from diﬀerent sources, provides a knowledge that time characteristics and noise spectrum have an essential signiﬁcance in generating loss in hearing [1,11,15]. Taking these data into consideration, a method of estimating the risk of hearing impairment has been proposed by J. Kotus, the author’s Ph.D. student. The proposed model is based on a modiﬁed Johnston’s psychoacoustical model [19], which provides a global distribution of a basilar membrane deﬂection within critical bands. In Figure 5, a scheme of the psychophysiological noise dosimeter is presented. The model is based on the analysis of a basilar membrane’s answer to noise in critical bands. First, the power spectrum of noise is determined. Afterwards, it is corrected by taking into account the transition from the outer to the inner ear. Subsequently, particular spectrum coeﬃcients are grouped into critical bands, according to the bark scale. Then the noise level in particular critical bands is calculated. The result let us assess the extent to which the basilar membrane is stimulated. Its answer is determined by the multiplication of the instate stimulation level value by the characteristics of hearing ﬁlters for the particular critical band. A basilar membrane displacement value obtained in this way is exponentially averaged. This action reﬂects the inertia of the processes occurring in the inner ear. The obtained averaged values are used for the assessment of the asymptotic hearing threshold shift. Finally, these values are subjected to exponential averaging, that reﬂects a process of the hearing threshold shift. Therefore this model enables to assess TTS in critical bands and the recovery time of a hearing threshold to its initial value. The model enables to determine the hearing threshold shift for a given noise during the exposure. The initial simulation for two selected noise exposures was done (series 4 and 6). A theoretical time series of noise was created based on the histogram of the LAF levels. The obtained TTS levels for a particular measurement series amounted to: TTS = 13.5 dB (measurement No. 4), and TTS = 18.5 dB (measurement No. 6). A comparison of the noise exposure for series 4 and 6 has shown that the time characteristic of noise inﬂuences the occurrence and level of the TTS. The results obtained through the use of the presented model, conﬁrmed a greater TTS for the 6th exposition (club No. 2). Overall, this conﬁrms harmful eﬀects of noise.

6

Survey on Noise Annoyance

An objective noise measurement was extended by a subjective measurement by means of a dedicated survey. The survey consisted of three parts. The ﬁrst part involved getting information such as age, sex, class, school. The second part included questions about noise in places of residence and exposure to noise related to musical preferences. The last part concentrated on noise climates in schools in typical circumstances (lessons, breaks, etc.). In the following Figures (Figs. 6-10) some sample results are shown.

144

B. Kostek

Fig. 6. Main type of noise sources in the neighborhood during day and night

Fig. 7. Subjective evaluation of noise annoyance during day (CL – comfortable)

Fig. 8. Subjective evaluation of noise annoyance during night

Main type of noise sources in the neighborhood As seen in Figure 6 the main sources of noise during day and night are communication and the neighborhood. They constitute 60% of all noise sources. On the other hand, industrial noise does not present a threat to the place of living.

The Domain of Acoustics Seen from the Rough Sets Perspective

145

Subjective evaluation of noise annoyance Overall, 8 % of asked pupils evaluated their place of living as quiet, 40% said that the neighborhood is either too loud or loud. On the other hand, during the night most places were evaluated as quiet, apart from special events occurring during the night. It should be mentioned that 70% of our responders live in towns or in close proximity to towns. Subjective evaluation of noise annoyance in schools Almost 70% of pupils evaluate noise as too loud during breaks, and the remaining 30% as loud. As to noise evaluation during lessons pupils diﬀer in their opinions.

Fig. 9. Subjective evaluation of noise annoyance during breaks

Fig. 10. Subjective evaluation of noise annoyance during lessons

Evaluation of noise sources related to personal way of living In addition, pupils were asked how often they listen to music and how loud this music is, and also how often they use personal music players and listen to music via headphones. It occurs that on average younger groups of pupils do not often listen to loud music, contrarily to older pupils and students. The latter group

146

B. Kostek

also attends musical events very frequently. As a result, one may notice problems with speech understanding when noise is present, which in students’ subjective opinions occurred in 50% of situations.

7

Data Analysis Employing Rough Sest

Regulations and gathered data enabled to deﬁne attributes that may provide ways to assess the risk of developing PTS (Permanent Threshold Shift) and noise annoyance. Conditional attributes from the decision table (Table 5) are derived from objective measurements and noise dose calculation. One of the attributes driven more subjectively, namely the absence and presence of the Tinnitus eﬀect after the exposure to noise, is also taken into consideration. As mentioned before, otoacoustic emission is treated as an early indicator of hearing loss, that is why the DPOAE analysis should be included in the decision table. In a way the decision table follows results included in Tables 1 and 2. Another decision table (Table 6) includes answers gathered at schools based on a questionnaire form published in the Internet (MNMS service [7]). The decision attribute are in this case noise annoyance/nuisance or noise threat. A legal deﬁnition of a nuisance says that this is the noise that oﬀends or upsets the receiver because it is occurring at the wrong time, in the wrong place, or is of a character that annoys due to excessive tonal components or impulses. When looking at the table, one may expect that sensitivity to noise is more frequent among children and increases with age. This could be explained by the fact that small children get easily tired in noisy conditions, and on the other hand, people active professionally prefer some rest after work, the same as older people. This means that the age of the survey respondents’ should be taken into consideration. Also, it is quite obvious that females are less noisy than males while entertaining. When analyzing other attributes from Table 5 the relation between them and the decision attribute is not always that clear, and since this is a subjectively-driven questionnaire many contradictions may occur in the acquired data. A good example of such a confusing attribute is ‘hearing’. Hearing impairments of the inner ear are characterized by the so-called recruitment phenomenon, thus people with the inner-ear hearing loss may suﬀer more in noisy conditions. This means that they do not hear soft sounds, and perceive loud sounds as too loud. This phenomenon is due to an abnormally rapid rise of the loudness sensation with only a small increase in sound stimulus intensity. In this early study, the survey did not include questions related to for example social class, since it was directed at speciﬁc groups (pupils attending schools and students). The observations and earlier researches on noise annoyance indicate that these factors should also be taken into account. All these factors constitute huge data, especially as the survey will be progressively ﬁlled in by the Internet respondents. This enables to consolidate data on noise annoyance on a very large scale. The only way to analyze subjectivelydriven data is to use a soft computing approach, and particularly the rough set

The Domain of Acoustics Seen from the Rough Sets Perspective

147

method which seems to be most suitable for this problem. The rough set-based analysis of the decision tables enables to derive rules in which attributes and their values support a claim that noise dose analysis based on average daily exposure is not only insuﬃcient to assess noise annoyance, but also to predict the risk of permanent threshold shift early enough. There are a lot of data in which we see noise events recalled by the survey respondents even if they live in a quiet neighborhood with no evident source of noise during the day and night. This means that they evaluate the noise annoyance or nuisance as high because of noise events that happen sporadically and were of short duration. In such a case, factors derived from the noise dose analysis will not exceed any norm or regulations, but still these noise events may be highly stressful and harmful. Table 5. Decision table based on measurements and calculation according to noise regulations (denotation same as before, in addition: Ts denotes absence or presence of the Tinnitus eﬀect) No. 1 2 ... n

LAF min 67.4 52.5 ... ...

Leq 89 100.5 ... ...

LAF max 105.5 114.4 ... ...

Exposure Time . . . 600 4058 ... ... ... ...

D 5.2 506.1 ... ...

DPOAE ... ... ... ...

TTS YES YES ... ...

Ts YES YES ... ...

PTS NO YES ... ...

Table 6. Decision table based on survey (denotations: StoN – sensitivity to noise, H – hearing, SC – social class, WP/S – type of work performed/school, SR/D – Survey results/neighborhood day, SR/N – Survey results/neighborhood night, SR/NS – Survey results/noise source, NE/AdNS – Noise events/additional noise sources, NA/N – noise annoyance/nuissance) No. Age Sex StoN H SC WP/S SR/D SSR/N SR/NS 1 41-50 M high good high educat. quiet quiet no 2 10-15 F high good NO high noisy quiet road school traﬃc ... ... ... ... ... ... ... ... ... ... n ... ... ... ... ... ... ... ... ...

. . . NE/AdNS NA/N . . . dogs HIGH . . . ambul. HIGH siren ... ... ... ... ... ...

A rule prototype derived from the rough set-based analysis is presented below: (attribute 1) = (val 1) ∧ ... ∧ (attribute k) = (val k) => (P T SY ES/N O )

(4)

It should be mentioned that some of the numerical attributes require quantization, however in the case of the noise-related attributes the quantization should be based on norms and standards. This means a process of replacing the original values of the input data with for example the number of an interval to which a selected parameter value belongs should be performed taking into account speciﬁc constraints. For example, it is well-known that values of LAF max , Leq , Exposure

148

B. Kostek

Time exceeding the limits for which these values start to be harmful constitute the cut-points. It is obvious that after rule derivation not all attributes are to be retained and some of the rules are not longer valid. Most important attributes are to be found in reducts. Preliminary experiments show that amongst the most signiﬁcant attributes resulting from measurements one may ﬁnd: LAF max , Leq , Exposure Time and TTS, though some of these parameters are interrelated, and are eliminated in the rough set-based analysis, reducing the number of the correlated attributes. On the other hand, looking into the Decision Table based on survey results, one may ﬁnd that such attributes as: age of a person, speciﬁc type of noise events, duration of these events, neighborhood noise background are present in the evaluation of noise annoyance/nuissance. Having the rough set analysis performed and rules derived based on both approaches (objective and subjective), this may constitute a scientiﬁc basis of an advertising campaign against noise pollution and its adverse consequences, which in addition may contribute to better regulations on noise. This is why noise monitoring stations should be installed in all agglomerations larger than 250,000 inhabitants in the coming year, thus both measurement quantities shown in a form of noise maps and subjective opinions would be available for the same place, in future. However, it is worth noticing that without changing our habits, regulations may still be ineﬀective as seen from the survey.

8

Summary

To sum up, on the basis of the investigations, it was conﬁrmed that noise climates in schools is adverse to pupils’ and teachers’ health. The main reasons of the high noise level in schools are: the behavior of pupils, loudspeaker systems and low sound absorption of the classrooms and corridors. The data analysis of the hearing measurements at schools does not conﬁrm negative inﬂuence of noise on the hearing system. Especially because the time of exposure to noise is too short to produce measurable changes in the activity of the inner ear. Noise measured during of the students’ music band rehearsals and in clubs reaches very high levels and exceeds all related norms. Measurements of the hearing characteristics of people working or entertaining in these places conﬁrmed harmful eﬀects of noise. A signiﬁcant TTS and the reduction of the DP level were observed. A comparison of the noise exposure for series 4 and 6 has shown that the time characteristic of noise inﬂuences the occurrence of the TTS. The results obtained through the use of the presented model, conﬁrmed a greater TTS for the 6th exposition (club No. 2). In addition, the Tinnitus eﬀect was perceived by some students. The data presented are very complex, interrelated, and in some cases contradictory, thus for an adequate analysis they require the use of a decision system. For this particular problem, it is thought that the rough set method is the most suitable solution for the analysis, since it allows to derive rules/reduct capable of identifying the most signiﬁcant attributes. It is then possible to determine the weights that should be taken into consideration when constructing a feature vector.

The Domain of Acoustics Seen from the Rough Sets Perspective

149

On the basis of the survey, it may be said that even if norms are not exceeded, noise can be still perceived as annoying, stressful and in consequence harmful. The analysis of the acquired data has shown that all measurement quantities should ﬁrst be included as conditional attributes in the constructed decision tables along with subjective factors. At the moment, two types of decision tables have been proposed. The ﬁrst one provides indications as to the signiﬁcance of some attributes in relation to the PTS occurrence (decision attribute). The second gathers results of the survey, and relates subjective factors to the noise annoyance. The results of this analysis may provide the basis for an adequate campaign against noise pollution and lead to better regulations. Acknowledgment. This work was supported by the Polish Ministry of Science and Education within the projects No. 3T11E02829 and No. R0201001. The author wishes to acknowledge her Ph.D. student J. Kotus for his valuable input to the presented research.

References 1. Borg, E., Engstrom, B.: Noise level, inner hair cell damage, audiometric features and equal-energy hypothesis, J Acoust. Soc. Am. 86 (5) (1989) 1776–1782 2. Budzynska, L., Jelonek, J., Lukasik, E., Slowinski, R.,: Supporting Experts in Ranking Generic Audio and Visual Objects, Proc. IEEE Workshop ”Signal Processing’2004”, Poznan (2004) 81–86. 3. Budzynska, L., Jelonek, J., Lukasik, E., Susmaga, R., Slowinski R.: Multistimulus ranking versus pairwise comparison in assessing quality of musical instruments sounds, 118 AES Convention Paper, 6482, Barcelona (2005) 4. Criteria for a recommended standard, Occupational Noise Exposure, U.S. Department of Health and Human Services (1998) 5. http://www.cdc.gov/niosh/98-126.html (CRITERIA FOR A RECOMMENDED STANDARD) 6. Czyzewski, A., Kostek, B., Skarzynski, H.: Intelligent System for Environmental Noise Monitoring, in Monitoring, Security, and Rescue Techniques in Multiagent Systems, Series: Advances in Soft Computing, Dunin-Keplicz, B.; Jankowski, A.; Skowron, A.; Szczuka, M.(eds.), chapter, 397–410 , XII, Springer Verlag, Heidelberg, New Yorkc(2005) 7. Czyzewski, A., Kostek, B., Skarzynski, H.: IT applications for the remote testing of communication senses” chapter in INFORMATION TECHNOLOGY SOLUTIONS FOR HEALTH CARE, Spinger-Verlag (2006) 8. Czyzewski, A., Szczerba M., Kostek B.: Musical Phrase Representation and Recognition by Means of Neural Networks and Rough Sets, Rough Set Theory and Applications (RSTA), vol. 1, 259-284, Advances in Rough Sets, Subseries of SpringerVerlag Lecture Notes in Computer Sciences, LNCS 3100, Transactions on Rough Sets, Grzymala-Busse, J.W., Kostek, B., Swiniarski, R.W., Szczuka M. (eds.) (2004) 9. Czyzewski A., Kostek B.: Musical Metadata Retrieval with Flow Graphs, in Rough Sets and Current Trends in Computing, RSCTC, Uppsala, Sweden, Lecture Notes in Atiﬁcial Intelligence, LNAI 3066, Springer Verlag, Berlin, Heidelberg, New York (2004) 691–698

150

B. Kostek

10. Czyzewski, A., Kotus, J., Kostek, B.,: Comparing Noise Levels and Audiometric Testing Results Employing IT Based Diagnostic Systems, The 33rd International Congress and Exposition on Noise Control Engineering INTERNOISE’2004, August 22-24, Prague (2004) 11. Dunn, D.E., Davis, R.R., Merry, C.J., Franks, J.R.: Hearing loss in the chinchilla from impact and continuous noise exposure, J Acoust. Soc. Am. 90 (4) (1991) 1979–1985 12. Dziubinski, M., Dalka, P., Kostek, B.: Estimation of Musical Sound Separation Algorithm Eﬀectiveness Employing Neural Networks, J. Intelligent Information Systems, Special Issue on Intelligent Multimedia Applications, 24, 2(2005) 133– 157 13. Dziubinski, M., Kostek, B.: Octave Error Immune and Instantaneous Pitch Detection Algorithm, J. of New Music Research, vol. 34, 292-273, Sept. 2005. 14. Engel, Z.W., Sadowski J., et al.: Noise protection in Poland in European Legislation, The Committee on Acoustics of the Polish Academy of Science & CIOP-PIB, Warsaw, (2005) (in Polish) 15. Henderson, D., Hamernik, R.P.: Impulse noise: Critical review, J Acoust. Soc. Am. 80(2) (1986) 569–584 16. Hippe, M.P.: Towards the Classiﬁcation of Musical Works: A Rough Set Approach Third International Conference, RSCTC 2002, Malvern, PA, USA, October 14-16, 2002. Proceedings Editors: J.J. Alpigini, J.F. Peters, A. Skowron, N. Zhong (eds.) (2002) 546-553 17. Jelonek, J., Lukasik, E., Naganowski, A., Slowinski, R.: Inferring Decision Rules from Jurys’ Ranking of Competing Violins, Proc. Stockholm Music Acoustic Conference, KTH, Stockholm (2003) 75–78 18. Jelonek, J., Lukasik, E., Naganowski, A., Slowinski, R.: Inducing jury’s preferences in terms of acoustic features of violin sounds, Lecture Notes in Computer Science, LNCS 3070, Springer (2004) 492–497 19. Johnston, J.D.: Transform Coding of Audio Signals Using Perceptual Noise Criteria. IEEE Journal on Selected Areas in Communications, vol. 6(2) (1988) 314–323 20. Komorowski, J, Pawlak, Z, Polkowski, L, Skowron, A. Rough Sets: A Tuto-rial. In: Pal SK, Skowron A (eds) Rough Fuzzy Hybridization: A New Trend in DecisionMaking. Springer-Verlag (1998), 3–98 21. Kostek, B.: Soft Computing in Acoustics, Applications of Neural Networks, Fuzzy Logic and Rough Sets to Musical Acoustics, Physica Verlag, Heidelberg, New York (1999) 22. Kostek, B.: Perception-Based Data Processing in Acoustics. Applications to Music Information Retrieval and Psychophysiology of Hearing, Springer Verlag, Series on Cognitive Technologies, Berlin, Heidelberg, New York (2005) 23. Kostek, B.: Musical Instrument Classiﬁcation and Duet Analysis Employing Music Information Retrieval Techniques, Proc. of the IEEE, 92, 4 (2004) 712–729 24. Kostek, B.: Intelligent Multimedia Applications - Scanning the Issue, J. Intelligent Information Systems, Special Issue on Intelligent Multimedia Applications, 24, 2 (2005) 95–97 (Guest Editor) 25. Kostek, B., Wojcik, J.: Machine Learning System for Estimating the Rhythmic Salience of Sounds, International J. of Knowledge-based and Intelligent Engineering Systems, 9 (2005), 1–10

The Domain of Acoustics Seen from the Rough Sets Perspective

151

26. Kostek, B., Czyzewski, A.: Processing of Musical Metadata Employing Pawlak’s Flow Graphs, Rough Set Theory and Applications (RSTA), vol. 1, 285–305, Advances in Rough Sets, Subseries of Springer-Verlag Lecture Notes in Computer Sciences, LNCS 3100, Transactions on Rough Sets, Grzymala-Busse, J.W., Kostek, B., Swiniarski, R.W., Szczuka, M., (eds.) (2004) 27. Kotus, J., Kostek, B.: Investigation of Noise Threats and Their Impact on Hearing in Selected Schools, OSA’ 2006, Archives of Acoustics (2006) (in print). 28. Kotus, J.: Evaluation of Noise Threats and Their Impact on Hearing by Employing Teleinformatic Systems, (Kostek, B., supervisor) (2007) (in preparation). 29. Lukasik, E.: AMATI-Multimedia Database of Violin Sounds. In: Proc Stockholm Music Acoustics Conference, KTH Stockholm (2003a) 79–82 30. Lukasik, E.: Timbre Dissimilarity of Violins: Speciﬁc Case of Musical Instruments Identiﬁcation. Digital Media Processing for Multimedia Interactive Services, World Scientiﬁc, Singapore (2003b) 324–327 31. Lukasik, E., Susmaga, R.: Unsupervised Machine Learning Methods in Timbral Violin Characteristics Visualization. In: Proc Stockholm Music Acoustics Conference, KTH Stockholm (2003) 83–86 32. Melnick, W.: Human temporary threshold shift (TTS) and damage risk, J Acoust. Soc. Am. 90(1) (1991) 147–154 33. Pal, S.K., Polkowski, L., Skowron, A. Rough-Neural Computing. Techniques for Computing with Words. Springer Verlag, Berlin Heidelberg New York (2004) 34. Pawlak, Z.: Rough Sets. International J Computer and Information Sciences (1982) 35. Pawlak, Z.: Probability, Truth and Flow Graph. Electronic Notes in Theoretical Computer Science 82, International Workshop on Rough Sets in Knowledge Discovery and Soft Computing, Satellite event of ETAPS 2003, Elsevier, Warsaw (2003) 36. Pawlak, Z.: Elementary Rough Set Granules: Towards a Rough Set Processor. In: Pal SK, Polkowski L, Skowron A (eds) Rough-Neural Computing. Techniques for Computing with Words. Springer Verlag, Berlin Heidelberg New York, 5–13(2004) 37. Pawlak, Z.: A Treatise on Rough Sets. Transactions on Rough Sets IV, Peters, J.F., Skowron, A. (Eds) 1–17 (2005) 38. Polish Standard PN-N-01307 (1994), Permissible sound level values in work-places and general requirements concerning taking measurements (in Polish). 39. Seixas, N., et al.: Alternative Metrics for Noise Exposure Among Construction Workers, Ann Occup Hyg. 49 (2005) 493–502 40. A. Wieczorkowska, P. Synak, R. Lewis, Z. W. Ras, Creating Reliable Database for Experiments on Extracting Emotions from Music. In: M. A. Klopotek, S. Wierzchon, K. Trojanowski (eds.), Intelligent Information Processing and Web Mining, Proceedings of the International IIS: IIPWM’05 Conference, Gdansk, Poland Advances in Soft Computing, Springer (2005), 395-402 41. Wieczorkowska, A., Synak, P., Lewis, R., Ras, Z.W.: Extracting Emotions from Music Data, in: M.-S. Hacid, Murray, N.V., Ras Z.W., Tsumoto, S. (eds.), Foundations of Intelligent Systems, 15th International Symposium, ISMIS 2005, Saratoga Springs, NY, USA, 2005, Proceedings; LNAI 3488, Springer, 456-465 42. Wieczorkowska, A., Ras, Z.W.: Do We Need Automatic Indexing of Musical Instruments?, in: Warsaw IMTCI, International Workshop on Intelligent Media Technology for Communicative Intelligence, Warsaw, Poland, September 13–14, Proceedings, PJIIT - Publishing House (2004), 43–38 43. Wieczorkowska, A.: Towards Extracting Emotions from Music. In: Warsaw IMTCI, International Workshop on Intelligent Media Technology for Communicative Intelligence, Warsaw, Poland, September 13–14, Proceedings, PJIIT - Publishing House (2004) 181–183

Rule Evaluations, Attributes, and Rough Sets: Extension and a Case Study Jiye Li1 , Puntip Pattaraintakorn2, and Nick Cercone3 1

2

David R. Cheriton School of Computer Science, University of Waterloo 200 University Avenue West, Waterloo, Ontario, Canada N2L 3G1 [email protected] Department of Mathematics and Computer Science, Faculty of Science King Mongkut’s Institute of Technology Ladkrabang, Thailand [email protected] 3 Faculty of Science and Engineering, York University 4700 Keele Street, North York, Ontario, Canada M3J 1P3 [email protected]

Abstract. Manually evaluating important and interesting rules generated from data is generally infeasible due to the large number of rules extracted. Diﬀerent approaches such as rule interestingness measures and rule quality measures have been proposed and explored previously to extract interesting and high quality association rules and classiﬁcation rules. Rough sets theory was originally presented as an approach to approximate concepts under uncertainty. In this paper, we explore rough sets based rule evaluation approaches in knowledge discovery. We demonstrate rule evaluation approaches through a real-world geriatric care data set from Dalhousie Medical School. Rough set based rule evaluation approaches can be used in a straightforward way to rank the importance of the rules. One interesting system developed along these lies in HYRIS (HYbrid Rough sets Intelligent System). We introduce HYRIS through a case study on survival analysis using the geriatric care data set.

1

Introduction

The general models of knowledge discovery in databases (KDD) contains processes including data preprocessing, knowledge discovery algorithms, rule generations and evaluations. Rule evaluation is a signiﬁcant process in KDD. How to automatically extract important, representative rules to the human beings instead of selecting those useful rules manually are the main problems. Speciﬁc diﬃculties make the research of rule evaluation very challenging. One of the diﬃculties is that real-world large data sets normally contain missing attribute values. They may come from the collecting process, or redundant scientiﬁc tests, change of the experimental design, privacy concerns, ethnic issues, unknown data and so on. Discarding all the data containing the missing attribute values cannot fully preserve the characteristics of the original data, and J.F. Peters et al. (Eds.): Transactions on Rough Sets VI, LNCS 4374, pp. 152–171, 2007. c Springer-Verlag Berlin Heidelberg 2007

Rule Evaluations, Attributes, and Rough Sets: Extension and a Case Study

153

wastes part of the data collecting eﬀort. Knowledge generated from missing data may not fully represent the original data set, thus the discovery may not be as suﬃcient. Understanding and utilizing of original context and background knowledge to assign the missing values seem to be an optimal approach for handling missing attribute values. In reality, it is diﬃcult to know the original meaning for missing data from certain application domains. Another diﬃculty is that huge amount of rules are generated during the knowledge discovery process, and it is infeasible for humans to manually select useful and interesting knowledge from such rule sets. Rough sets theory, originally proposed in the 1980’s by Pawlak [1], was presented as an approach to approximate concepts under uncertainty. The theory has been widely used for attribute selection, data reduction, rule discovery and many knowledge discovery applications in the areas such as data mining, machine learning and medical diagnoses. We are interested in tackling diﬃcult problems in knowledge discovery from a rough sets perspective. In this paper, we introduce how rough sets based rule evaluations are utilized in knowledge discovery systems. Three representative approaches based on rough sets theory are introduced. The ﬁrst approach is to provide a rank of how important is each rule by rule importance measure (RIM) [2]. The second approach is to extract representative rules by considering rules as condition attributes in a decision table [3]. The third approach is applied to data containing missing values. This approach provides a prediction for all the missing values using frequent itemsets as a knowledge base. Rules generated from the complete data sets contain more useful information. The third approach can be used at the data preprocessing process, combining with the ﬁrst or second approach at the rule evaluation process to enhance extracting more important rules. It can also be used alone as preprocessing of missing attribute values. An interesting system based on this rule-enhanced knowledge discovery system, HYRIS (HYbrid Rough sets Intelligent System) [4], is developed. Case studies on using HYRIS on survival analysis are further demonstrated. We address particular problems from real-world data sets, using recent missing attribute value techniques and rule evaluations based on rough sets theory to facilitate the tasks of knowledge discovery. The rule discovery algorithm focuses on association rule algorithms, although it can be classiﬁcation algorithm, decision tree algorithm and other rule discovery algorithms from data mining and machine learning. We demonstrate the rule evaluation approaches using a real-world geriatric care medical data set. We discuss related work on rough sets theory, current knowledge discovery system based on rough sets, and rule evaluations in Section 2. Section 3 presents three rough sets based rule evaluations methods. We show experiments on the geriatric care data set in Section 4. Section 5 contains a case study of HYRIS system developed based on the proposed approaches, and experiments on survival analysis are demonstrated. Section 6 gives the concluding remarks.

154

2

J. Li, P. Pattaraintakorn, and N. Cercone

Related Work

We introduce related work to this paper including rough sets theory, knowledge discovery systems based on rough sets theory and existing rule evaluations approaches. 2.1

Rough Sets Theory

Rough sets theory, proposed in the 1980’s by Pawlak [1], has been used for attribute selection, rule discovery and many knowledge discovery applications in the areas such as data mining, machine learning and medical diagnoses. We brieﬂy introduce rough sets theory [1] as follows. U is the set of objects we are interested in, where U = φ. Deﬁnition 1. Equivalence Relation. Let R be an equivalence relation over U , then the family of all equivalence classes of R is represented by U/R. [x]R means a category in R containing an element x ∈ U . Suppose P ⊆ R, and P = φ, IN D(P ) is an equivalence relation over U . For any x ∈ U , the equivalence class of x of the relation IN D(P ) is denoted as [x]P . Deﬁnition 2. Lower Approximation and Upper Approximation. X is a subset of U , R is an equivalence relation, the lower approximation of X and the upper approximation of X is deﬁned as: RX = ∪{x ∈ U |[x]R ⊆ X}

(1)

RX = ∪{x ∈ U |[x]R ∩ X = φ}

(2)

respectively. From the original deﬁnitions [1], reduct and core are deﬁned as follows. R is an equivalence relation and let S ∈ R. We say, S is dispensable in R, if IN D(R) = IN D(R − {S}); S is indispensable in R if IN D(R) = IN D(R − {S}). We say R is independent if each S ∈ R is indispensable in R. Deﬁnition 3. Reduct. Q is a reduct of P if Q is independent, Q ⊆ P , and IN D(Q) = IN D(P ). An equivalence relation over a knowledge base can have many reducts. Deﬁnition 4. Core. The intersection of all the reducts of an equivalence relation P is deﬁned to be the Core, where Core(P ) = ∩All Reducts of P. Reduct and core are among the most important concepts in this theory. A reduct contains a subset of condition attributes that are suﬃcient enough to represent the whole data set. The reducts can be used in attribute selection process. There may exist more than one reduct for each decision table. Finding all the reduct sets for a data set is NP-hard [5]. Approximation algorithms are used to obtain reduct sets [6]. The intersection of all the possible reducts is called the core.

Rule Evaluations, Attributes, and Rough Sets: Extension and a Case Study

155

The core is contained in all the reduct sets, and it is the essential of the whole data. Any reduct generated from the original data set cannot exclude the core attributes. Reduct Generations. There are several reduct generation approaches, such as ROSETTA [7], RSES [8], ROSE2 [9], QuickReduct algorithm [10] and Hu et al. [11]’s reduct generation combining the relational algebra with the traditional rough sets theory. ROSETTA rough set system GUI version 1.4.41 [7] provides Genetic reducer, Johnson reducer, Holte1R reducer, Manual reducer, Dynamic reducer, RSES Exhaustive reducer and so on. Genetic reducer is an approximation algorithm based on genetic algorithm for multiple reducts generation. Johnson reducer generates only a single reduct with minimum length. In this research, we use both genetic and Johnson’s reduct generations to develop rule evaluations approaches. Core Generation. Hu et al. [11] introduced a core generation algorithm based on rough sets theory and eﬃcient database operations, without generating reducts. The algorithm is shown in Algorithm 1, where C is the set of condition attributes, and D is the set of decision attributes. Card denotes the count operation in databases, and Π denotes the projection operation in databases. Algorithm 1. Hu’s Core Generating Algorithm [11] input : Decision table T (C, D), C is the condition attributes set; D is the decision attribute set. output: Core, Core attributes set. Core ← φ; for each condition attribute A ∈ C do if Card(Π(C − A + D)) = Card(Π(C − A)) then Core = Core ∪ A; end end return Core;

This algorithm is developed to consider the eﬀect of each condition attribute on the decision attribute. The intuition is that, if the core attribute is removed from the decision table, the rest of the attributes will bring diﬀerent information to the decision making. Theoretical proof of this algorithm is provided in [11]. The algorithm takes advantage of eﬃcient database operations such as count and projection. This algorithm requires no inconsistency in the data set. 2.2

Rough Sets Based KDD Systems

We brieﬂy survey current rough sets based knowledge discovery systems. We discuss the individual functions of each system based on general characteristics, such as the input data sets, the preprocessing tasks, the related rough sets tasks, the rule generations and so on.

156

J. Li, P. Pattaraintakorn, and N. Cercone

1. ROSETTA. ROSETTA [7] software is a general purpose rough set toolkit for analyzing the tabular data, and is freely distributed. The downloadable versions for both the Windows and Linux operating systems are available. The software supports the complete data mining process, from data preprocessing, including processing incomplete data, data discretization, generating reduct sets which contain essential attributes for the given data set, to classiﬁcation, rule generation, and cross validation evaluation. Some discretization and reducts generation packages are from RSES library [8]. 2. RSES2.2. RSES [8] stands for Rough Set Exploration System. There are downloadable versions for both the Windows and Linux operating systems. It is still maintained and being developed. The system supports data preprocessing, handling incomplete data, discretization, data decomposition into parts that share the same properties, reducts generation, classiﬁcation, and cross validations and so on. 3. ROSE2. ROSE [9] stands for Rough Sets Data Explorer. This software is designed to process data with large boundary regions. The software supports data preprocessing, data discretization, handling missing values, core and reducts generation, classiﬁcations and rule generation, as well as evaluations. This software provides not only the classical rough set model, but also the variable precision model, which is not provided by [7] and [8]. 4. LERS. LERS [12] stands for Learning from Examples based on Rough Sets. It is not publicly available. The system was designed especially to process missing values of attributes and inconsistency in the data set. Certain rules and possible rules are both extracted based on the lower and upper approximations. In addition to the rough sets based systems mentioned above, there are other available knowledge discovery systems based on the methodologies of rough sets such as GROBIAN [13] and DBROUGH [14]. 2.3

Current Research on Rule Evaluations

Rule generation often brings a large amount of rules to analyze. However, only part of these rules are distinct, useful and interesting. How to select only useful, interesting rules among all the available rules to help people understand the knowledge in the data eﬀectively has drawn the attention of many researchers. Research on designing eﬀective measures to evaluate rules comes from statistic, machine learning, data mining and other ﬁelds. These measures fall into two categories of evaluation measures. Rule Interestingness Measures. One category of evaluating rules is to rank the rules by rule interestingness measures. Rules with higher interestingness measures are considered more interesting. The rule interestingness measures, originated from a variety of sources, have been widely used to extract interesting rules. Diﬀerent applications may have diﬀerent interestingness measures emphasizing on diﬀerent aspect of the applications. Hilderman provided an extensive survey

Rule Evaluations, Attributes, and Rough Sets: Extension and a Case Study

157

on the current interestingness measures [15] for diﬀerent data mining tasks. For example, support and conﬁdence are the most common interestingness measures to evaluate the association rules. Not all the interestingness measures generate the same rank of interestingness for the same set of rules. Depending on diﬀerent application purpose, appropriate rule interestingness measures should be selected to extract proper rules. More than one measure can be applied together to evaluate and explain the rules. Tan et. al. [16] evaluate twenty one measures in their comparative experiments and suggest diﬀerent usage domains for these measures. They provide several properties of the interestingness measures so that one can choose a proper measure for certain applications. Their experiments also imply that not all the variables perform equally good at capturing the dependencies among the variables. Furthermore, there is no measure that can perform constantly better than the others in all application domains. Diﬀerent measure is designed towards diﬀerent domains. Rule Quality Measures. The concept of rule quality measures was ﬁrst proposed by Bruha [17]. The motivation for exploring this measure is that decision rules are diﬀerent with diﬀerent predicting abilities, diﬀerent degrees to which people trust the rules and so on. Measures evaluating these diﬀerent characteristics should be used to help people understand and use the rules more eﬀectively. These measures have been known as rule quality measures. The rule quality measures are often applied in the post-pruning step during the rule extraction procedure [18]. For example, some measures are used to evaluate whether the rules overﬁt the data. When removing an attribute-value pair, the quality measure does not decrease in value, this pair is considered to be redundant and will be pruned. As one of the applications, rule generation system uses rule quality measures to determine the stopping criteria for the rule generations and extract high quality rules. In [19] twelve diﬀerent rule quality measures were studied and compared through the ELEM2 [18] system on their classiﬁcation accuracies. The measures include empirical measures, statistical measures and measures from information theory.

3

Rule Evaluation on Knowledge Discovery

In this section, we ﬁrst examine a current rough set knowledge discovery system, and suggest the importance of rule evaluations. We propose rule evaluation approaches and their functions in knowledge discovery systems. 3.1

Analyzing RSES – Rough Set Exploration System

We take the RSES [8] system as an example system, and study in more detail of the role of rule evaluations. We show that current systems are limited with regard to rule evaluation, and we emphasize the importance of rule evaluation in current knowledge discovery systems.

158

J. Li, P. Pattaraintakorn, and N. Cercone

RSES (Rough Set Exploration System) is a well developed knowledge discovery system focusing on data analysis and classiﬁcation tasks, which is currently under development. Figure 1 shows a use of the system on a heart disease data set for classiﬁcation rule generation.

Fig. 1. Using Rough Set Exploration System on heart data

The data input to RSES is in the form of decision table T = (C, D), where C is the condition attribute set and D is the decision attribute set. Preprocessing is conducted once the data is imported to the system, during which stage the missing attribute values are handled and discretization is performed if necessary as well. Reducts are then generated, classiﬁcation rules based on the reducts are extracted. RSES provides four approaches on processing missing attribute values, such as removing data records with missing values, assigning the most common values of the missing attribute within the same decision class and without the same decision class, and considering missing attribute values as a special value of the attribute [8]. These approaches are used during the data preprocessing stage in the system. Although these approaches are fast and can be directly applied in the data, they lack the ability of preserving the semantic meanings of the original data set. Missing values may be assigned, however, the ﬁlled values may not be able to fully represent what is missing in the data. RSES provides rule postprocessing, which are “rule ﬁlter”, “rule shorten” and “rule generalize”. “Rule ﬁlter” removes from the rule set rules that do not satisfy certain support. “Rule shorten” shortens the length of the rules according to certain parameters [8]. “Rule generalization” generalizes rules according to a system provided parameter on the precision level. Although these rule postprocessing approaches provide an easier presentation of all the rule sets, these approaches do not provide ways to evaluate which rules are more interesting,

Rule Evaluations, Attributes, and Rough Sets: Extension and a Case Study

159

and which rules have higher quality. These functions cannot provide a rank of rules according to a rule’s signiﬁcance to the users. 3.2

Enhanced Knowledge Discovery System Based on Rough Sets

We present a rough set based knowledge discovery system, as shown in Figure 2.

Original Data

ItemRSFit on predicting missing values

Preprocessing

Processed Data

Attribute Selection (Rough Sets)

Rule Generation

Defining Rule Template

Rule Importance Measures Rule Evaluation Reduct Rules

Knowledge

Knowledge Representation visualization

Fig. 2. The Knowledge Discovery Based on Rough Sets Theory

In this general purpose knowledge discovery system, data from diﬀerent application domains are ﬁrst imported into the system. Preprocessings including missing attribute values processing, discretization, are conducted in this stage. After the data is preprocessed attribute selections are conducted. Depending on the output, diﬀerent attribute selection approaches can be applied here. Rule generation algorithms extract rules. After the rule sets are obtained, the important postprocessing - rule evaluations are performed in this stage. Rules are ﬁnally represented, possibly visualized in a certain format, as knowledge to the end users. We introduce three approaches integrated into this general purpose KDD system as shown in Figure 2. The ﬁrst approach ItemRSFit is used in the data preprocessing stage. The second approach, rule importance measure is used to

160

J. Li, P. Pattaraintakorn, and N. Cercone

rank rules during the rule evaluation process. The third approach of extracting reduct rules is also used during the rule evaluation stage. We will elaborate these approaches in the following. I. Predicting missing attribute values based on Frequent Itemset. ItemRSFit approach is a recently developed approach on predicting missing attribute values based on association rules algorithm and rough sets theory. It has been shown on both large scale real world data set and UCI machine learning data sets on the improved prediction accuracies. ItemRSFit approach is an integration of two other approaches from association rule algorithm and rough sets theory. Priori to the association rule generation, frequent itemsets are generated based on the item-item relations from the large data set according to a certain support. Thus the frequent itemsets of a data set represent strong correlations between diﬀerent items, and the itemsets represent probabilities for one or more items existing together in the current transaction. When considering a certain data set as a transaction data set, the implications from frequent itemsets can be used to ﬁnd to which attribute value the missing attribute is strongly connected. Thus the frequent itemset can be used for predicting the missing values. We call this approach “itemset-approach” for prediction. The larger the frequent itemsets used for the prediction, the more information from the data set itself will be available for prediction, hence the higher the accuracy will be obtained. However, generating frequent itemset for large data set is time-consuming. Although itemsets with higher support need less computation time, they restrict item-item relationships, therefore not all the missing values can be predicted. In order to balance the tradeoﬀ between computation time and the percentage of the applicable prediction, another approach must be taken into consideration. A reduct contains a subset of condition attributes that are suﬃcient enough to represent the whole data set. The intersection of all the possible reduct is the core. Therefore the attributes contained in the reduct or core are more important and representative than the rest of the attributes. Thus by examining only attributes within the same core or reduct to ﬁnd the similar attribute value pairs for the data instance containing the missing attribute values, we can assign the most relevant value for the missing attribute. Since this method only considers a subset of the data set, which is either the core or the reduct, the prediction is quite fast. This approach “RSFit” is recently proposed in [20], and it is an alternative approach designed for fast prediction. It can be used to predict missing attributes that cannot be predicted by the frequent itemset. We integrate the prediction based on frequent itemset and RSFit approach into a new approach ItemRSFit to predict missing attribute values. Frequent itemsets are used to predict missing values ﬁrst, and RSFit approach is used to predict the rest of the missing values that cannot be predicted by the frequent itemsets. This integrated approach can predict missing values from the data itself, therefore less noise is brought into the original data. The details on the ItemRSFit approach is presented in [21].

Rule Evaluations, Attributes, and Rough Sets: Extension and a Case Study

161

Properly processed data can improve the quality of the generated knowledge. Therefore the ItemRSFit approach is used in this system at the preprocessing stage. It helps to preserve the qualities of the original input data to this system, thus facilitate the rule evaluation process. II. Rule Importance Measures. Rule importance measure [2] is developed to provide a diverse rank of how important the association rules are, although this approach can also be applied to rules generated by other rule discovery algorithms. Association rules algorithm can be applied on this transaction data set to generate rules, which have condition attributes on the antecedent part and decision attributes on the consequent part of the rules. Rules generated from diﬀerent reduct sets can contain diﬀerent representative information. If only one reduct set is being considered to generate rules, other important information might be omitted. Using multiple reducts, some rules will be generated more frequently than other rules. We consider the rules that are generated more frequently more important. The Rule Importance is deﬁned to be important by the following deﬁnition. Deﬁnition 5. If a rule is generated more frequently across diﬀerent rule sets, we say this rule is more important than rules generated less frequently across those same rule sets. Rule importance measure is deﬁned as follows, Deﬁnition 6 Number of times a rule appears in all the generated rules from the reduct sets . Rule Importance Measure = Number of reduct sets The deﬁnition of the rule importance measure can be elaborated by Eq. 3. Let n be the number of reducts generated from the decision table T (C, D). Let RuleSets be the n rule sets generated based on the n reducts. rulesetj ∈ RuleSets (1 ≤ j ≤ n) denotes individual rule sets containing rules generated based on reducts. rulei (1 ≤ i ≤ m) denotes the individual rule from RuleSets. RIMi represents the rule importance measure for the individual rule. Thus the rule importance measures can be computed by the following |{rulesetj ∈ RuleSets|rulei ∈ rulesetj }| . (3) n The details of how to use rule importance measures can be found in [2]. Rule importance measure can be integrated into the current rough sets based knowledge discovery system to be used during the rule evaluation process. A list of ranked important rules can therefore be presented with their rule importance measures to facilitate the understanding of the extracted knowledge. RIMi =

162

J. Li, P. Pattaraintakorn, and N. Cercone

III. Extracting Reduct Rules. In [3] a method of discovering and ranking important rules by considering rules as attributes was introduced. The motivation comes from the concept of reduct. A reduct of a decision table contains attributes that can fully represent the original knowledge. If a reduct is given, rules extracted based on this reduct are representative of the original decision table. Can we take advantage of the concept of a reduct to discover important rules? We construct a new decision table Am×(n+1) , where each record from the original decision table u0 , u1 , ..., um−1 are the rows, and the columns of this new table consists of Rule0 , Rule1, ..., Rulen−1 and the decision attribute. We say a rule can be applied to a record in the decision table if both the antecedent and the consequent of the rule appear together in the record. For each Rulej (j ∈ [0, ..., n−1]), we assign 1 to cell A[i, j] (i ∈ [0, ..., m−1]) if the rule Rulej can be applied to the record ui . We set 0 to A[i, j] otherwise. The decision attribute A[i, n] (i ∈ [0, ..., m − 1]) remains the same as the original values of the decision attribute in the original decision table. Eq. 4 shows the conditions for the value assignments of the new decision table. ⎧ ⎨ 1, if j < n and Rulej can be applied to ui A[i, j] = 0, if j < n and Rulej cannot be applied to ui ⎩ di , if j = n and di is the corresponding decision attributes for ui (4) where i ∈ [0, ..., m − 1], j ∈ [0, ..., n − 1]. We further deﬁne Reduct Rule Set and Core Rule Set. Deﬁnition 7. Reduct Rule Set. We deﬁne a reduct generated from the new decision table A as Reduct Rule Set. A Reduct Rule Set contains Reduct Rules. The Reduct Rules are representative rules that can fully describe the decision attribute. Deﬁnition 8. Core Rule Set. We deﬁne the intersection of all the Reduct Rule Sets generated from this new decision table A as Core Rule Set. A Core Rule Set contains Core Rules. The Core Rules are contained in every Reduct Rule Set. By considering rules as attributes, reducts generated from the new decision table contain all the important attributes, which represent the important rules generated from the original data set; and it excludes the less important attributes. Core attributes from the new decision table A contain the most important attributes, which represent the most important rules. Other Enhancements. The three approaches discussed in our research have shown to eﬀectively evaluate rules. There are other techniques that can be used along with these approaches in Figure 2. For example, during the rule generation

Rule Evaluations, Attributes, and Rough Sets: Extension and a Case Study

163

process, properly deﬁned rule templates can not only reduce the computation of rule generations, but it also ensures high quality rules, or interesting rules generated according to the application purposes. Important attributes, such as probe attributes (discussed in Section 5) can be deﬁned in the data preprocessing stage for generating rules containing such attributes for generating expected rules. Our motivation is, proposing approaches to enhance the current knowledge discovery system, to facilitate the knowledge discovery process on discovering more interesting and higher quality rules.

4

Experiments

We demonstrate, through a series of experiments, that systems improved by the proposed rule evaluation approaches can help humans discover and understand more important rules. 4.1

Specifying Rule Templates

Apriori association rules algorithm is used to generate rules. Because our interest is to make decisions or recommendations based on the condition attributes, we are looking for rules with only decision attributes on the consequent part. Therefore, we specify the following 2 rule templates to extract rules we want as shown by Template 5, and to subsume rules as shown by Template 6. Attribute1 , Attribute2 , . . . , Attributen → DecisionAttribute

(5)

Template 5 speciﬁes only decision attributes can be on the consequent part of a rule, and Attribute1 , Attribute2 , . . . , Attributen lead to a decision of DecisionAttribute. We specify the rules to be removed or subsumed using Template 6. For example, given rule Attribute1 , Attribute2 → DecisionAttribute

(6)

the following rules Attribute1 , Attribute2 , Attribute3 → DecisionAttribute

(7)

Attribute1 , Attribute2 , Attribute6 → DecisionAttribute

(8)

can be removed because they are subsumed by Template 6. Take the geriatric care data in Table 1 as an example, in the rule set, a rule shown as Eq. 9 exists SeriousChestP roblem → Death

(9)

the following rule is removed because it is subsumed. SeriousChestP roblem, T akeM edicineP roblem → Death

(10)

164

4.2

J. Li, P. Pattaraintakorn, and N. Cercone

Geriatric Care Data Set

We perform experiments on a geriatric care data set as shown in Table 1. This data set is an actual data set from Dalhousie University Faculty of Medicine to determine the survival status of a patient giving all the symptoms he or she shows. The data set contains 8, 547 patient records with 44 symptoms and their survival status (dead or alive). We use survival status as the decision attribute, and the 44 symptoms of a patient as condition attributes, which includes education level, the eyesight, hearing, be able to walk, be able to manage his/her own meals, live alone, cough, high blood pressure, heart problem, cough, gender, the age of the patient at investigation and so on.1 There is no missing value in this data set. There are 12 inconsistent data entries in the medical data set. After removing these instances, the data contains 8, 535 records. 2 Table 1. Geriatric Care Data Set edulevel eyesight hearing health 0.6364 0.25 0.50 0.25 0.7273 0.50 0.25 0.25 0.9091 0.25 0.50 0.00 0.5455 0.25 0.25 0.50 0.4545 0.25 0.25 0.25 0.2727 0.00 0.00 0.25 0.0000 0.25 0.25 0.25 0.8182 0.00 0.50 0.00 ... ... ... ...

trouble livealone cough hbp heart 0.00 0.00 0.00 0.00 0.00 0.50 0.00 0.00 0.00 0.00 0.00 0.00 0.00 1.00 1.00 0.00 1.00 1.00 0.00 0.00 0.00 1.00 0.00 1.00 0.00 0.50 1.00 0.00 1.00 0.00 0.00 0.00 0.00 0.00 1.00 0.00 0.00 0.00 1.00 0.00 ... ... ... ... ...

stroke 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 ...

... ... ... ... ... ... ... ... ... ...

sex livedead 1 0 2 0 1 0 2 0 2 0 2 0 1 0 2 0 ... ...

The ItemRSFit approach is implemented by Perl and the experiments are conducted on Sun Fire V880, four 900Mhz UltraSPARC III processors. We use apriori frequent itemset generation [23] to generate frequent 5-itemset. The core generation in RSFit approach is implemented with Perl combining the SQL queries accessing MySQL (version 4.0.12). ROSETTA software [7] is used for reduct generation. 4.3

Experiments on Predicting Missing Attribute Values

In order to show the ItemRSFit approach obtains better prediction accuracy than the existing approach (i.e., RSFit), we perform the experiments on the geriatric care data set by randomly selecting 150 missing values from the original data. We then apply both RSFit approach and ItemRSFit approach on predicting missing values, and compare the accuracy of the prediction. Figure 3 demonstrates the comparison predicting abilities between RSFit and ItemRSFit approaches. We can see from the ﬁgure that the smaller the support is, the more accurate the 1 2

Refer to [22] for details about this data set. Notice from our previous experiments that core generation algorithm cannot return correct core attributes when the data set contains inconsistent data entries.

Rule Evaluations, Attributes, and Rough Sets: Extension and a Case Study

165

100 RSFit ItemRSFit 95

Accuracy (%)

90

85

80

75

70 10

20

30

40

50 Support (%)

60

70

80

90

Fig. 3. Accuracy Comparisons for Geriatric Care Data with 150 Missing Attribute Values

prediction of the missing attribute values for the ItemRSFit approach obtains; whereas for the RSFit approach, the accuracy remains the same as the value of support gets smaller; and the accuracy obtained by RSFit is always lower than the ItemRSFit approach. This result demonstrates that frequent itemsets as knowledge base can be eﬀectively applied for predicting missing attribute values. 4.4

Experiments on Rule Importance Measure

In our experiment, we use the genetic algorithm to generate multiple reduct sets with the option of full discernibility. The apriori algorithm [23] for large item sets generation. The core attributes for this data set are eartrouble, livealone, heart, highbloodpressure, eyetrouble, hearing, sex, health, educationlevel, chest, housework, diabetes, dental, studyage. Table 2. Reduct Sets for the Geriatric Care Data Set after Preprocessing No. Reduct Sets 1 {edulevel,eyesight,hearing,shopping,housewk,health,trouble,livealone, cough,sneeze,hbp,heart,arthriti,eyetroub,eartroub,dental, chest,kidney,diabetes,feet,nerves,skin,studyage,sex} 2 {edulevel,eyesight,hearing,phoneuse,meal,housewk,health,trouble,livealon, cough,sneeze,hbp,heart,arthriti,evetroub,eartroub,dental, chest,bladder,diabetes,feet,nerves,skin,studyage,sex} ... ... 86 {edulevel,eyesight,hearing,shopping,meal,housewk,takemed,health, trouble,livealone,cough,tired,sneeze,hbp,heart,stroke,arthriti, eyetroub,eartroub,dental,chest,stomach,kidney,bladder,diabetes, feet,fracture,studyage,sex}

166

J. Li, P. Pattaraintakorn, and N. Cercone Table 3. Rule Importance for the Geriatric Care Data No. 0 1 2 3 4 5 ... 10 ... 216 217

Selected Rules Rule Importance SeriousHeartProblem → Death 100% SeriousChestProblem → Death 100% SeriousHearingProblem, HavingDiabetes → Death 100% SeriousEarTrouble → Death 100% SeriousEyeTrouble → Death 100% Sex Female → Death 100% ... ... Livealone, HavingDiabetes, NerveProblem → Death 95.35% ... ... SeriousHearingProblem, ProblemUsePhone → Death 1.16% TakeMedicineProblem, NerveProblem → Death 1.16%

Table 2 shows selected reduct sets among the 86 reducts generated by ROSETTA. All of these reducts contain the core attributes. For each reduct set, association rules are generated with support = 30%, conf idence = 80%. 3 218 unique rules are generated over these 86 reducts. These rules as well as their rule importance are shown in Table 3. Among these 218 rules, 87 rules have rule importance of no less than 50% , 8 of which have rule importance of 100%. All the rules with rule importance of 100% contain only core attributes. 4.5

Experiments on Generating Reduct Rules

The new decision table A8535×219 is constructed by using the 218 rules 4 as condition attributes, and the original decision attribute as the decision attribute. Note that after reconstructing the decision table, we must check for inconsistency again before generating reduct rules for this table. After removing the inconsistent data records, there are 5709 records left in the new decision table. The core rule set is empty. We use Johnson’s reduct generation algorithm on this table A5709×219 and the reduct rule set is {Rule0, Rule1, Rule3 , Rule5, Rule19 , Rule173}. We show these rules in Table 4. From Table 4 we can see that the reduct rule sets contain 6 rules. There are 4 rules judged to be the most important. The rule importance for Rule0 , Rule1, Rule3 and Rule5 are all 100%. The Rule19 has the importance of 82.56%, which is more important among the 218 rules. 3

4

Note that the value of support and conﬁdence can be adjusted to generate as many or as few rules as required. There are 1615 rules generated by apriori algorithm from the original data set with support = 30%, conf idence = 80%, after applying the rule template. We can circumvent problems inherent in considering all 1615 generated rules using the 218 unique rules that are derived from the 86 reducts obtained by ROSETTA’s genetic algorithm.

Rule Evaluations, Attributes, and Rough Sets: Extension and a Case Study

167

Table 4. Reduct Rules for the Geriatric Care Data No. in Table 3 0 1 3 5 19 173

5

Reduct Rules

Rule Importance SeriousHeartProblem → Death 100% SeriousChestProblem → Death 100% SeriousEarTrouble → Death 100% Sex Female → Death 100% Livealon, OftenSneeze, DentalProblems, HavingDiabetes → Death 82.56% ProblemHandleYourOwnMoney → Death 27.91%

A Case Study: Survival Analysis in HYRIS

This section provides a case study to illustrate how a rough sets based KDD system provides a useful mechanism for analyzing and distilling essential attributes and rules from survival data, and evaluates the generated rules in postprocessing for survival analysis. Use of medical survival data challenges researchers because of the size of data sets and vagaries of their structures. Among prognostic modeling techniques that induce models from medical data, survival analysis warrants special treatment in the type of data required and its modeling. Data required for medical analysis includes demographic, symptoms, laboratory tests and treatment information. Special features for survival data are the events of interest, censoring, followup time and survival time speciﬁc for each type of disease. Such data demands powerful analytical models for survival analysis. The studies applying rough sets to survival analysis related to our work are [24][25]. They illustrated rough sets contribution to a medical expert system for throat cancer patients successfully. Rough sets and decision trees have been used to study kidney dialysis patients’ survival [26]. HYRIS (HYbrid Rough sets Intelligent System) [27][28] is designed speciﬁcally to consider survival analysis with several statistical approaches. HYRIS uses the CDispro algorithm from a previous study [4]. HYRIS successively derives dispensable, probe attribute, reduct and probe reduct together with life time table and Kaplan-Meier survival curves [27]. In addition to survival analysis, HYRIS provides a general data analysis and decision rule generation and evaluation [27] as well. HYRIS Case Study I. HYRIS is able to analyze censor variable and survival time attributes that are a speciality for survival analysis. Given the survival data set, the system can identify the covariant levels of particular attributes according to rough sets and several statistical approaches. The Kaplan-Meier method, hazard function, hypothesis testing, log-rank, Brewslow, Tarone-Ware tests, p-value and CDispro [4]. incorporate the rough sets framework to generate core, dispensable attributes, probe attribute, reducts, and probe reducts which are the informative attributes. Consequently, the rules are derived and validated with ELEM2 [18].

168

J. Li, P. Pattaraintakorn, and N. Cercone

We demonstrate the utility of HYRIS by investigating a particular problem using both actual and benchmark medical data sets: geriatric data, melanoma data [29], pneumonia data [30] and primary biliary cirrhosis data (PBC) [31]. For the geriatric care data set, time lived in months are used as the survival time target function. Inconsistent records are removed. Data is discretized according to percentile groups. The age is used as probe attribute, the reducts and probe reducts are generated successfully. The rules generated for geriatric data care are the decision rules for predicting survival time. Note that in Section 4 the experimental results showing important rules are used to predict the survival status, not the survival time. Two sample survival prediction rules out of 1,600 rules of geriatric care data set (when considering the probe attribute {ExperienceDiabetes}) generated from HYRIS are provided as follows: Rule 1: UnHealthy, SevereHearingDamage, NerveProblem, FootProblem, SeriousHeartProblem, DentalDisease, StomachDisease, HighBloodPressure, ExperienceDiabetes → SurvivalTime = 7-18 months. Rule 2: FemalePatient, LowEducationLevel, EyesightProblemLowToSeriousType, HealthProblemFromLowToSeriousType, HearQuiteWell, DoNotHaveDiabetesExperience, EasilyTired, FootProblem, → SurvivalTime = 56-73 months. When comparing the accuracy of rules that were generated from original attributes and those generated from reducts, the accuracy of all data sets range between 83.7851%–90.5686%. Rule performance outcomes are improved signiﬁcantly as reported in [4]. HYRIS Case Study II. HYRIS can accomplish preprocessing, learning and model construction and broaden further to use in rule evaluation and postprocessing. We continue a series of studies in [28]. In this case study, we propose an alternative approach for decision rule learning with rough sets theory in the postprocessing step called ROSERULE - Rough Sets Rule Reducts Learning Algorithm. ROSERULE learned and analyzed from the rule set to generate rule reducts which can be used to reduce the number of the rules. Results imply a reduced number of rules that successfully preserve the original classiﬁcation. The rule numbers of geriatric data set reduced from 1,600 to 1,150, melanoma data set reduced from 16 to 15, pneumonia data set reduced from 606 to 42 and PBC data set reduced from 83 to 72. At the same time, the prediction accuracy is preserved for all data sets.

6

Conclusion

We study the work of rough sets based rule evaluations on knowledge discovery system. We propose solutions to the challenging problems brought by large real world data sets, such as the existence of missing values and analyzing huge

Rule Evaluations, Attributes, and Rough Sets: Extension and a Case Study

169

amount of generated rules manually. Three rough set based approaches to enhance the current KDD systems on rule evaluations are introduced. The ItemRSFit approach is used to predict missing attribute values using frequent itemset as a knowledge base. Complete data can be obtained using this approach. The rule importance measure provides a ranking of how important is a rule. Finally, the reduct rules are extracted using the concept of reduct by considering rules as condition attributes in a decision table. Experimental results on a real world geriatric care data set demonstrate the utilities of applying rough sets based rule evaluations to enhance current KDD systems. A case study of a recent knowledge discovery system shows the applications of approaches which have been incorporated into HYRIS with an emphasis on survival analysis.

Acknowledgements This research is supported by the Natural Sciences and Engineering Research Council of Canada (NSERC). The research of Puntip Pattraintakorn has also been supported by a grant from the King Mongkut’s Institute of Technology Ladkrabang (KMITL) research fund, Thailand. We would also like to thank Arnold Mitnitski from Dalhousie University for providing the geriatric care data set, and the anonymous reviewers for their helpful comments.

References 1. Pawlak, Z.: Rough Sets. In Theoretical Aspects of Reasoning about Data. Kluwer Academic Publishers (1991) 2. Li, J. and Cercone, N.: Introducing A Rule Importance Measure. Transactions on Rough Sets, Springer LNCS, vol 5 (2006) 3. Li, J., Cercone, N.: Discovering and Ranking Important Rules. In Proceedings of IEEE International Conference on Granular Computing, vol 2, Beijing China 25-27 July (2005) 506–511 4. Pattaraintakorn, P., Cercone, N., Naruedomkul, K.: Hybrid Intelligent Systems: Selecting Attributes for Soft-Computing Analysis. In Proc. of the 29th Annual International Computer Software and Applications Conference (COMPSAC), vol 2 (2005) 319–325 5. Kryszkiewicz, M., Rybinski, H.: Finding Reducts in Composed Information Systems, Rough Sets, Fuzzy Sets Knowldege Discovery. In W.P. Ziarko (Ed.), Proceedings of the International Workshop on Rough Sets, Knowledge Discovery, Heidelberg/Berlin: Springer-Verlag (1994) 261–273 6. Jan Bazan, Hung Son Nguyen, Sinh Hoa Nguyen, Piotr Synak, and Jakub Wroblewski.: Rough set algorithms in classiﬁcation problems. Rough Set Methods and Applications: New Developments in Knowledge Discovery in Information Systems, volume 56 of Studies in Fuzziness and Soft Computing, pages 49-88. Physica-Verlag, Heidelberg, Germany (2000). 7. Øhrn, A.: Discernibility and Rough Sets in Medicine: Tools and Applications. PhD Thesis, Department of Computer and Information Science, Norwegian University of Science and Technology, Trondheim, Norway. (1999) 8. RSES 2.2 User’s Guide. Warsaw University. http://logic.mimuw.edu.pl/∼rses/

170

J. Li, P. Pattaraintakorn, and N. Cercone

9. Predki, B., Wilk, Sz.: Rough Set Based Data Exploration Using ROSE System. In: Foundations of Intelligent Systems. Ras, Z. W., Skowron, A., Eds, LNAI 1609, Springer-Verlag, Berlin (1999) 172-180 10. Chouchoulas, A. and Shen, Q.: Rough Set-Aided Keyword Reduction For Text Categorization. Applied Artiﬁcial Intelligence, vol 15 (2001) 843–873 11. Hu, X., Lin, T., Han, J.: A New Rough Sets Model Based on Database Systems. Fundamenta Informaticae 59 no.2-3 (2004) 135–152 12. Freeman, R. L., Grzymala-Busse, J. W., Riﬀel, L. A., Schroeder, S. R.: Analyzing the Relation Between Heart Rate, Problem Behavior, and Environmental Events Using Data Mining System LERS. In 14th IEEE Symposium on Computer-Based Medical Systems (CBMS’01) (2001) 13. Ivo, D., Gunther, G.: The Rough Set Engine GROBIAN. In Proc. of the 15th IMACS World Congress, vol 4, Berlin, August (1997) 14. Hu, T., Shan, N., Cercone, N. and Ziarko, W.: DBROUGH: A Rough Set Based Knowledge Discovery System, Proc. of the 8th International Symposium on Methodologies for Intelligent System, LNAI 869, Spring Verlag (1994) 386–395 15. Hilderman, R. and Hamilton, H.: Knowledge discovery and interestingness measures: A survey. Technical Report 99-04, Department of Computer Science, University of Regina, October (1999) 16. Pang-Ning Tan and Vipin Kumar and Jaideep Srivastava: Selecting the right interestingness measure for association patterns. Processings of SIGKDD. (2002) 32–41 17. Bruha, Ivan: Quality of Decision Rules: Deﬁnitions and Classiﬁcation Schemes for Multiple Rules. In Machine Learning and Statistics, The Interface, Edited by G. Nakh aeizadeh and C. C. Taylor. John Wiley & Sons, Inc. (1997) 107–131 18. An, A. and Cercone, N.: ELEM2: A Learning System for More Accurate Classiﬁcations. In: Proceedings of Canadian Conference on AI (1998) 426–441 19. An, A. and Cercone, N.: Rule Quality Measures for Rule Induction Systems: Description and Evaluation. Computational Intelligence. 17-3 (2001) 409–424. 20. Li, J. and Cercone, N.: Assigning Missing Attribute Values Based on Rough Sets Theory. In Proceedings of IEEE Granular Computing, Atlanta, USA. (2006) 21. Li, J. and Cercone, N.: Predicting Missing Attribute Values based on Frequent Itemset and RSFit. Technical Report, CS-2006-13, School of Computer Science, University of Waterloo (2006) 22. Li, J. and Cercone, N.: Empirical Analysis on the Geriatric Care Data Set Using Rough Sets Theory. Technical Report, CS-2005-05, School of Computer Science, University of Waterloo (2005) 23. Borgelt, C.: Eﬃcient Implementations of Apriori and Eclat. Proceedings of the FIMI’03 Workshop on Frequent Itemset Mining Implementations. In: CEUR Workshop Proceedings (2003) 1613-0073 http://CEUR-WS.org/Vol-90/borgelt.pdf 24. Bazan, J., Osmolski, A., Skowron, A., Slezak, D., Szczuka, M., Wroblewski, J.: Rough Set Approach to the Survival Analysis. In Alpigini, J. J., et al. (Eds.): The Third International Conference on Rough Sets and Current Trends in Computing (RSCTC), Proceedings, LNAI 2475, Springer-Verlag Berlin Heidelberg (2002) 522-529 25. Bazan, J., Skowron, A., Slezak, D., Wroblewski, J.: Searching for the Complex Decision Reducts: The Case Study of the Survival Analysis, LNAI 2871, SpringerVerlag, Berlin Heidelberg (2003) 160-168 26. A. Kusiak, B. Dixon, S. Shah: Predicting Survival Time for kidney Dialysis Patients: A Data Mining Approach, Computers in Biology and Medicine 35 (2005) 311–327

Rule Evaluations, Attributes, and Rough Sets: Extension and a Case Study

171

27. Pattaraintakorn, P., Cercone, N., Naruedomkul, K.: Selecting Attributes for SoftComputing Analysis in Hybrid Intelligent Systems. In D. Slezak et al. (Eds.): Rough Sets, Fuzzy Sets, Data Mining, and Granular Computing 10th International Conference (RSFDGrC), Proceedings, Part II Series: Lecture Notes in Computer Science, Subseries: LNAI 3642, Springer-Verlag, Berlin, Heidelberg (2005) 698–708 28. Pattaraintakorn, P., Cercone, N., Naruedomkul, K.: Rule Analysis with Rough Sets Theory, The IEEE International Conference on Granular Computing, Atlanta, USA (2006) 29. Elisa, L.T., John, W.W.: Statistical methods for survival data analysis, 3rd edn. New York: John Wiley and Sons (2003) 30. Klein, J.P., Moeschberger, M.L.: Survival analysis: techniques for censored and truncated data, 2nd edn. Berlin: Springer (2003) 31. Newman, D.J., Hettich, S., Blake, C.L. and Merz, C.J.: UCI Repository of machine learning databases. University of California, Irvine, Department of Information and Computer Seiences (1998) http://www.ics.uci.edu/∼mlearn/MLRepository.html

The Impact of Rough Set Research in China: In Commemoration of Professor Zdzislaw Pawlak Qing Liu1,2 and Hui Sun1 1

Department of Computer Science & Technology Nanchang Institute of Technology, Nanchang 330099, China qliu− [email protected] 2 Department of Computer Science & Technology Nanchang University, Nanchang 330029, China

This article is dedicated to the creative genius Zdzislaw Pawlak for his contribution to the theoretical development of science and technology in China. His distinguished discovery of Rough Set Theory is a formal theory which is well suited for uncertainty computing to analyze imprecise, uncertain or incomplete information of data. Inspired by his work scientists and engineers in China has developed many theories and applications in various science and technology ﬁelds. For instance, J.H.Dai studied the theories of Rough Algebras and Axiom Problem of Rough 3-Valued Algebras [1, 2]. G.L.Liu studied the Rough Sets over Fuzzy Lattices [3, 4]. D.W.Pei studied the Generalized Model of Fuzzy Rough Sets [5]. W.Z.Wu Studied the On Random Rough Sets [6]. D.Q.Miao studied the Rough Group and Their Properties [7]. These are part of their recent research results related to rough set theory. As a matter of fact, there are still many researchers working in the ﬁeld of rough sets in China, who have proposed many creative results for last few years. These results are not listed one by one in this short commemorative article. We will try to review all the ”Rough Set” researchers and their research results in the appeared next article. In this article, we present only a recent partial research results of the authors. Based on Rough Logic and Decision Logic deﬁned by Pawlak [8, 9], ﬁrst author Liu has proposed a rough logic in a given information system [10]. Inﬂuenced by the concept of granular language proposed by Skowron [11], the granular logic deﬁned by Polkowski [12], and the work of Lin, Yao in [13, 14], we also have deﬁned a granular logic by applying the semantics of rough logical formulas in a given information system, and have created the deductive systems as well as have discussed many properties in [15, 16]. The proposed granular logic is a set which consists of granular formulas of form m(F ), where F is the rough logical formula in the given information system. It is used as individual variable of semantic function symbol m, so, we call it a paradigm of higher order logic. Truth values of granular formula of form m(F ) in the logic have two types. One is the function value, which is the meaning of rough logical formula, a subset in U; Another is the truth value of a degree, which is equal to a degree of meaning of the formula to close to universe U of objects. Pawlak introduced the concept of rough truth in 1987, assuming that a formula is roughly true in a J.F. Peters et al. (Eds.): Transactions on Rough Sets VI, LNCS 4374, pp. 172–175, 2007. c Springer-Verlag Berlin Heidelberg 2007

The Impact of Rough Set Research in China

173

given information system if and only if the upper approximation of meaning of the formula is equal to the whole universe. So, our approach extends Pawlak’s approach [9]. Skolem clause form, resolution principles, λ-resolution strategies and deductive reasoning of the granular logic are also discussed in the next article. These logic systems should be an extension of Rough Logic proposed by Pawlak in 1987 [9]. The practicability of the higher order logic will oﬀer the new idea for studying classical logic. It could also be a theoretical tool for studying granular computing. Based on reference [23], we further propose to use rough set theory to ﬁnd out minimal approximate solution set in the approximate solution space of diﬀerential equations and functional variation problems in mechanics. This could be a new studying style in rough set applications. The signiﬁcance and future development direction of the proposed Rough Sets are described. Any undeﬁnable subset on the universe of a given information system is constructed into precise deﬁnable lower and upper approximations via indiscernibility relation. Hence, complex and diﬃcult problems on undeﬁnable sets are resolved or transformed into precise deﬁnable lower and upper approximations [8, 17]. This is one of a great contribution of Pawlak’s Rough Sets Theory. Successful applications of rough sets in many ﬁelds oﬀer a lot of new idea of studying granular computing, which also promote the development of granular computing. Founder of predicate logic, G. Frege proposed the vague boundary in 1904, that is, how to compute the number of elements on vague boundary [18, 19]. Many mathematicians and computer scientists have made hard eﬀorts on the question. L. A. Zadeh proposed the Fuzzy Sets (FS) in 1965. He attempted to solve the computability of Frege’s vague boundary by Fuzzy Set concept [20]. Unfortunately, Fuzzy Sets are not mechanically computable, that is, the formula of exact describing for the fuzzy concept hasn’t been given. Therefore, the number of elements on the vague boundary could not be computed by exact formula. For example, the membership μ in Fuzzy Sets and fuzzy operator λ in operator fuzzy logic [21], could not be computed exactly. Z.Pawlak proposed the Rough Sets (RS) in 1982 for computing Frege’s vague boundary [8, 17], and the number of elements on the vague boundary could be exactly computed by it. Rough Set Theory is a new tool to deal with incomplete and uncertainty problems. In the ﬁeld of computer applications nowadays, this theory is no doubt a challenge to other uncertainty theories. Since rough set theory is one of the most important, newest theories and with the rapid development, it is also very important in artiﬁcial intelligence and cognitive science. Especially methods,which are based on rough set theory alone or are in combination with other approaches, have been used over a wide range of applications in many areas. More and more people in the region of China are attracted by them. In this article, we also present state of art of RS in China. Since May, 2001 Professor Pawlak being invited to China made his keynote speech in a conference and gave many invited talks among universities, the research of RS have been

174

Q. Liu and H. Sun

rapidly developed in China. Rough Set Theory has been used for important applications in human life, such as, data reduction, approximate classiﬁcation of data, management systems of business information, computing the average value and the standard deviation in quality test of products of statistics and so on. We developed successfully the Business Information Management Systems with rough set approach, which is a management for price of houses of some region [22]. We deﬁned an indiscernibility relation on numerical interval [a, b] by using ancient mathematical Golden Cut method and created rough sets by the deﬁned relation. Based on the rough set approach, we developed a ”Diagnosis Software of Blood Viscosity Syndrome on Hemorheology”, which is used to test Blood Viscosity chroma of patients. The systems has been applied in the clinic for many years. The medicine experts review that the diagnosis software is precursive, creative, scientiﬁc and practical [24]. In this article, we present the state of art of primary rough set research results and their applications in practice in China during last few years. Especially, in 2001, Pawlak was invited to China, his keynote speech had made a big inﬂuence on the development of science and technology of China. We would like to thank Pawlak for his fundamental and signiﬁcant contribution to the development of rough set research in China. We would like to thank the editor-in-chief Professor James F. Peters, Professor Andrej Skowron and Professor Ewa Orlowska for their kindness to let us publish article in this historical event to commemorate the great scientist Zdzislaw Pawlak for his contribution to the science and technology world. Thanks are also to the support of Natural Science Fund of China (NSFC-60173054). At last we would like to thank Dr. James Kuodo Huang (who is a IEEE member and have taught in the universities of USA for over 20 years) for his kind suggestions of English in this article. Still we would like to take the whole responsibility for any further errors made in this article.

References 1. Dai,J.H., Axiom Problem of Rough 3-Valued Algebras, The Proceedings of International Forum on Theory of GrC from Rough Set Perspective (IFTGrCRSP2006), Nanchang, China, Journal of Nanchang Institute of Technology, Vol.25, No.2, 2006, 48-51. 2. Dai,J.H., On Rough Algebras, Journal of software 16 (2005),1197-1204. 3. Liu,G.L., Rough Set Theory Over Fuzzy Lattices, The Proceedings of International Forum on Theory of GrC from Rough Set Perspective (IFTGrCRSP2006), Nanchang, China, Journal of Nanchang Institute of Technology, Vol.25, No.2, 2006, 51-51. 4. Liu,G.L., The Topological Structure of Rough Sets over Fuzzy Lattices, 2005 IEEE International Conference on Granualr Computing, Vol.I, Proceedings, Beijing, China, July 25-27,2005, 535-538. 5. Pei,D.W., A Generalized Model of Fuzzy Rough Sets, Int. J. General Systems 34 (5)2005, 603-613.

The Impact of Rough Set Research in China

175

6. Wu,W.Z., On Random Rough Sets, The Proceedings of International Forum on Theory of GrC from Rough Set Perspective (IFTGrCRSP2006), Nanchang, China, Journal of Nanchang Institute of Technology, Vol.25, No.2, 2006, 66-69. 7. Miao,D.Q., Rough Group, Rough Subgroup and their Properties, LNAI 3641, 10th International Conference, RSFDGrC2005, Regina, Canada, August/September 2005, Proceedings, Part I, 104-113. 8. Pawlak, Z., Rough Sets: Theoretical Aspects of Reasoning about Data, Kluwer Academic Publishers, Dordrecht , 1991. 9. Pawlak, Z., Rough Logic, Bulletin of the Polish Academy of Sciences, Technical Sciences, Vol.35, No.5-6, 1987, 253-259. 10. Liu, Q., Liu, S.H. and Zheng, F., Rough Logic and Applications in Data Reduction, Journal of Software, Vol.12, No.3, March 2001,415-419. (In Chinese). 11. Skowron, A., Toward Intelligent Systems: Calculi of Information Granules, Bulletin of International Rough Set Society, Vol.5, No.1/2, Japan, 2001,9-30. 12. Polkowski,L., A Calculus on Granules from Rough Inclusions in Information Systems, The Proceedings of International Forum on Theory of GrC from Rough Set Perspective (IFTGrCRSP2006), Nanchang, China, Journal of Nanchang Institute of Technology, Vol.25, No.2, 2006, 22-27. 13. Lin,T.Y., Liu,Q., First Order Rough Logic I: Approximate Reasoning via Rough Sets, Fundamenta Informaticae 2-3, 1996, 137-153. 14. Yao,Y.Y., Liu,Q., A Generalized Decision Logic in Interval-Set-Valued Information table, Lecture Notes in AI 1711, Springer-Verlag, Berlin, 1999, 285-294. 15. Liu,Q.,Sun,H., Theoretical Study of Granular Computing, LNAI 4062, The Proceedings of RSKT2006, by Springer, China, July 2006, 93-102. 16. Liu,Q. and Huang,Z.H., G-Logic and Its Resolution Reasoning, Chinese Journal of Computer, Vol.27,No.7, 2004, 865-873. (In Chinese). 17. Pawlak,Z., Rough Sets, Int. J. Inform. Comp. Sci., 11(1982), 341-356. 18. Frege,G., Grundgesetze der Arithmentic, In: Geach and Black (eds.) , Selection from the philosophical Writings of Gotlob Frege, Blackwei, Oxford 1970 19. Pawlak,Z., Rough Sets present State and Further Prospects, The Proceedings of Third International Workshop on Rough Sets and Soft Computing, Nov. 1012,1994, 72-76. 20. Zadeh,L.A., Fuzzy Sets, Information and Control, No.,8, 1965, 338-353. 21. Liu,X.H., Fuzzy Logic and Fuzzy Reasoning [M], Press. of Jilin University, Jilin, 1989,(In Chinese). 22. Liu,Q.,Sun H., Studying Direction of Granular Computing from Rough Set Perspective, Journal of Nanchang Institute of TechnologyVol. 25, No.3, 2006, 1-5. 23. Sun,H.,Liu,Q., The Research of Rough Sets in Normed Linear Space, LNAI 4259The Proceedings of RSCTC2006, by Springer, Japan, 8-11 Nov., 2006, 91-98. 24. Liu,Q., Jiang,F. and Deng,D.Y., Design and Implement for the Diagnosis Software of Blood Viscosity Syndrome Based on Hemorheology on GrC., Lecture Notes in Artiﬁcial Intelligence 2639, Springer-Verlag, 2003,413-420.

A Four-Valued Logic for Rough Set-Like Approximate Reasoning Jan Małuszy´nski1, Andrzej Szałas1,2 , and Aida Vit´oria3 1

Link¨oping University, Department of Computer and Information Science 581 83 Link¨oping, Sweden [email protected] 2 The University of Economics and Computer Science Olsztyn, Poland [email protected] 3 Dept. of Science and Technology, Link¨oping University S 601 74 Norrk¨oping, Sweden [email protected]

Abstract. This paper extends the basic rough set formalism introduced by Pawlak [1] to a rule-based knowledge representation language, called Rough Datalog, where rough sets are represented by predicates and described by finite sets of rules. The rules allow us to express background knowledge involving rough concepts and to reason in such a knowledge base. The semantics of the new language is based on a four-valued logic, where in addition to the usual values T RUE and FALSE, we also have the values B OUNDARY, representing uncertainty, and U NKNOWN corresponding to the lack of information. The semantics of our language is based on a truth ordering different from the one used in the well-known Belnap logic [2, 3] and we show why Belnap logic does not properly reflect natural intuitions related to our approach. The declarative semantics and operational semantics of the language are described. Finally, the paper outlines a query language for reasoning about rough concepts.

1 Introduction The seminal ideas of Pawlak [1, 4, 5, 6] on the treatment of imprecise and incomplete data opened a new area of research, where the notion of rough sets is used in theoretical studies as well as practical applications. Rough sets are constructed by means of approximations obtained by using elementary sets which partition a universe of considered objects. The assumption as to partitioning of the universe has been relaxed in many papers (see, e.g., [7, 8, 9, 10, 11, 12, 13, 14, 15]), however the Pawlak’s idea of approximations has remained the same. This paper extends the basic rough set formalism to a rule-based language, where rough sets are represented by predicates and are described by finite sets of rules. The rules allow one to express background knowledge concerning rough concepts and to reason in such a knowledge base. The new language is different from that proposed in [14, 15], where the rules described rough sets by combining their regions (lower approximation, upper approximation and boundary region). In contrast to the language J.F. Peters et al. (Eds.): Transactions on Rough Sets VI, LNCS 4374, pp. 176–190, 2007. c Springer-Verlag Berlin Heidelberg 2007

A Four-Valued Logic for Rough Set-Like Approximate Reasoning

177

described in this paper, the rules expressed in the language presented in [14, 15] refer explicitly to different regions of a rough set. Lifting the level of description makes the definitions easier to understand, also for the people not familiar with the technicalities of rough sets. The semantics of the new language is based on a four-valued logic, where in addition to the usual values T RUE and FALSE we have the values B OUNDARY representing uncertain/inconsistent information and U NKNOWN corresponding to the lack of information. As discussed in Section 3.2, the well-known four-valued Belnap logic [3, 2] does not properly reflect the natural intuitions related to our approach. We propose instead a slightly different truth ordering and use it, together with the standard knowledge ordering, for defining a declarative semantics of our language. By using the four-valued logic we propose, we are then able to deal with some important issues. First of all, we are able to provide a natural semantics for Datalog-like rules where negation can be used freely, both in the bodies and in the heads of rules. This, in previous approaches to various variants of negation, has always been problematic either due to the high computational complexity of queries or to a nonstandard semantics of negation, often leading to counterintuitive results (for an overview of different approaches to negation see, e.g., [16]). Our semantics reflects intuitions of fusing information from various independent sources. If all sources claim that a given fact is true (respectively, false) then we have an agreement and attach T RUE (respectively FALSE) to that fact. If information sources disagree in judgement of a fact, we attach to it the value B OUNDARY . If no source provides an information about a given fact, we then make it U NKNOWN. Second, we are able to import knowledge systems based on the classical logic without any changes and make them work directly within the rough framework. In such cases these systems would act as single information sources providing answers T RUE, FALSE, when queried about facts. Possible conflicting claims of different systems would then be solved by the same, uniform four-valued approach we propose. This might be useful in combining low level data sources, like classifiers as well as higher level expert systems. Third, one can import rough set-based systems, or systems supporting approximate reasoning, like for example, those described in [14, 15], or [17, 18]. In the latter threevalued logics are used (identifying B OUNDARY and U NKNOWN). The paper is structured as follows. First, in Section 2, we recall basic definitions related to rough sets and approximations. Next, in Section 3, we discuss our choice of four-valued logic. In Section 4 we introduce Rough Datalog and provide its semantics. Section 5 outlines a query language and discusses its implementation in logic programming. Finally, Section 6 concludes the paper.

2 Rough Sets According to Pawlak’s definition (see, e.g., [19]), a rough set S over a universe U is characterized by two subsets of U :

178

J. Małuszy´nski, A. Szałas, and A. Vit´oria Table 1. Test results considered in Example 1 car a a b c d

station safe s1 yes s2 no s2 no s1 yes s1 yes

– the set S, of all objects which can be certainly classified as belonging to S, called the lower approximation of S, and – the set S, of all objects which can be possibly classified as belonging to S, called the upper approximation of S. The set difference between the upper approximation and the lower approximation, denoted by S, is called the boundary region. In practice, in order to describe a given reality, one chooses a set of attributes and the elements of the underlying universe are described by tuples of attribute values. Rough sets are then defined by decision tables associating membership decisions with attribute values. The decisions are not exclusive: a given tuple of attribute values may be associated with the decision “yes”, with the decision “no”, with both or with none, if the tuple does not appear. Example 1. Consider a universe consisting of cars. If a car passed a test then it may be classified as safe (and as not safe, if it failed the test). Tests may be done independently at two test stations. The upper approximation of the rough set of safe cars would then include cars which passed at least one test. The lower approximation of the set would include the cars which passed all tests (and therefore, they did not fail at any test). The boundary region consists of the cars which passed one test and failed at one test. Notice that there are two other categories of cars, namely those which were not tested and those which failed all tests. As an example consider the situation described in Table 1, where the first column consists of cars, the second column consists of test stations and the third one contains test results. Denote by “Safe” the set of safe cars. Then: – the upper approximation of Safe consists of cars for which there is a decision “yes”, i.e., Safe = {a, c, d} – the lower approximation of Safe consists of cars for which all decisions are “yes”, i.e., Safe = {c, d} – the boundary region of Safe consists of cars for which there are both decisions “yes” and “no”, i.e., Safe = {a}. A decision table, representing a concept t, may be represented as a finite set of literals of the form t(y) or ¬t(x), where y ranges over the tuples of attribute values associated with the decision “yes” and x ranges over the tuples of attribute values associated with the decision “no”.

A Four-Valued Logic for Rough Set-Like Approximate Reasoning

179

Example 2. For the Example 1 with the universe of cars {a, b, c, d, e} and with two test stations, we may have the decision table, shown in Table 1, encoded as {safe(a), ¬safe(a), ¬safe(b), safe(c), safe(d)} . Notice that the literal safe(a) indicates that car a has passed a safety test in one of the stations while literal ¬safe(a) states that the same car as failed a safety test in another test station. In this case the rough set Safe has the approximations Safe = {a, c, d} and Safe = {c, d}. The rough set ¬Safe, describing those cars that have failed some test, has the approximations ¬Safe = {a, b} and ¬Safe = {b}. Note that it is totally unknown what is the status of car e. We notice that a decision table T of this kind defines two rough sets, T and ¬T , with a common boundary region which is the intersection of the upper approximations of both sets, i.e. T ∩ ¬T . As rough sets are usually defined by decision tables, we then adopt the following definition (used also in [20, 14, 15]). Definition 1. A rough set S over a universe U is a pair S, ¬S of subsets of U .

Intuitively, the rough set S describes those elements of U having certain property. The set S is the upper approximation of S, and consists of the elements of U for which there is an indication of having the given property. On the other hand, the set ¬S consists of the elements for which there is an indication of not having the property. In Example 2, Safe = {a, c, d} and ¬Safe = {a, b}. Remark 1 1. Observe that Definition 1 differs from the understanding of rough sets as defined by Pawlak. In fact, the definition of Pawlak requires the underlying elementary sets used in approximations to be based on equivalence relations, while Definition 1 relaxes this requirement. Such differences are examined and discussed in depth in [12]. 2. Since relations are sets of tuples, we further on also use the term rough relation to mean a rough set of tuples.

3 A Four-Valued Logic for Rough Sets 3.1 The Truth Values for Rough Membership Our objective is to define a logical language for rough set reasoning. The vocabulary of the language includes predicates to be interpreted as rough relations and constants to be used for representing attribute values. Consider an atomic formula of the form p(t1 , · · · , tn ), where p is a predicate, denoting a rough set P , and t1 , . . . , tn (with n > 0) are constants. We now want to define the truth value represented by an atom p(t1 , · · · , tn ). Let v = t1 , . . . , tn and “−” denote the set difference operation. Then, the following cases are possible:

180

J. Małuszy´nski, A. Szałas, and A. Vit´oria

– v ∈ P − ¬P : intuitively, we only have evidence that the element of the universe described by the attributes v has property P . Thus, the truth value of p(v) is defined to be T RUE . – v ∈ ¬P − P : intuitively, we only have evidence that the element of the universe described by the attributes v does not have property P . Thus, the truth value of p(v) is defined to be FALSE. – v ∈ P ∩ ¬P : in this case, we have contradictory evidences, i.e. an evidence that an element of the universe described by the attributes v has property P and an evidence that it does not have the property P . This is an uncertain information and we use the additional truth value B OUNDARY to denote it. – v ∈ P ∪ ¬P : in this case, we have no evidence whether the element of the universe described by the attributes v has property P . We then use another truth value called U NKNOWN. 3.2 Is Belnap Logic Suitable for Rough Reasoning? The truth values emerging from our discussion have been studied in the literature outside of the rough set context for defining four-valued logic. A standard reference is the well-known Belnap’s logic [2]. We now recall its basic principles and we discuss whether it is suitable for rough set reasoning. The Belnap logic is defined by considering a distributive bilattice of truth values and introducing logical connectives corresponding to the operations in the bilattice. Bilattices have been introduced in [21, 22]. They generalize the notion of Kripke structures (see, e.g., [23]). A bilattice is a structure B = U, ≤t , ≤k such that U is a non-empty set, ≤t and ≤k are partial orderings each making set U a lattice. Moreover, there is usually a useful connection between both orderings. We follow the usual convention that ∧t and ∨t stand respectively for the meet and join, with respect to ≤t . The symbols ∧k and ∨k stand respectively for the meet and join, with respect ≤k . Operations ∧t and ∨t are also called the conjunction and disjunction, and ∧k and ∨k are often designated as the consensus and accept all operators, respectively. The bilattice used in Belnap’s logic is shown in Fig 1. In the knowledge ordering, ≤k , U NKNOWN is the least value, reflecting total lack of knowledge. Each of the values T RUE and FALSE provide more information than U NKNOWN. Finally, the I NCONSISTENT value corresponds to the situation when there is evidence for both T RUE and FALSE.1 The truth ordering ≤t (see Fig 1) has T RUE as its largest element, and FALSE as its smallest element. Example 3. Assume that a family owns two cars: a and e. We want to check if the family has a safe car. This corresponds to the logical value of the expression safe(a) ∨t safe(e) .

1

(1)

Observe that I NCONSISTENT is replaced in our approach by B OUNDARY, which is closer to intuitions from rough set theory.

A Four-Valued Logic for Rough Set-Like Approximate Reasoning

181

≤k

6 I NCONSISTENT

FALSE

T RUE

U NKNOWN

- ≤t Fig. 1. The bilattice corresponding to Belnap’s logic

The truth values of safe(a) and safe(e) are determined by the results of the tests, as specified in Example 2. Thus safe(a) has the value B OUNDARY and safe(e) has the value U NKNOWN. If the join operation ∨t is defined by Belnap’s logic, as shown in Fig 1, then I NCONSISTENT ∨t U NKNOWN = T RUE . This contradicts our intuitions. We know that the safety of car a is unclear, since the results of both safety tests are contradictory, and we know nothing about safety of car e. Asking instead if all cars of the family are safe, safe(a) ∧t safe(e) ,

(2)

would in Belnap’s logic result in the answer FALSE. However, we really do not know whether both cars are safe because we do not have any information about the safety of car e. In contrast to the answer obtained in the Belnap’s logic, U NKNOWN seems to be a more intuitive answer in this case. The example above shows that the truth ordering of Fig 1, and consequently Belnap’s logic are not suitable for rough set-based reasoning. On the other hand, the knowledge ordering of Fig. 1 is adequate for our purposes. Indeed, the values T RUE and FALSE show that only one kind of evidence, either positive or negative, is known while the value B OUNDARY indicates existence of contradictory evidence, both positive and negative. 3.3 A Four-Valued Logic for Rough Set Reasoning We now define a four-valued logic suitable for rough set-based reasoning by modifying the bilattice of Fig.1. As discussed in Section 3.2, only the truth ordering is to be changed. We will use the new truth ordering to define conjunction (∧t ) as the greatest lower bound in this ordering. The ordering should preserve the usual meaning of conjunction for the truth values T RUE and FALSE. Intuitively, the value U NKNOWN represents the lack of information. Thus, the result of its conjunction with any other

182

J. Małuszy´nski, A. Szałas, and A. Vit´oria

truth value is accepted here to be U NKNOWN. A new information may arrive, replacing U NKNOWN by either T RUE, or FALSE, or B OUNDARY , providing in each case a different result. On the other hand, B OUNDARY represents existence of contradictory information. Its conjunction with T RUE would not remove this contradiction. Thus, we define the result of such a conjunction to be B OUNDARY . It also seems natural, that the conjunction of FALSE with T RUE or B OUNDARY gives FALSE. Consequently the truth ordering, ≤t , is redefined in our framework as U NKNOWN ≤t FALSE ≤t B OUNDARY ≤t T RUE .

(3)

The new structure R = U, ≤t , ≤k , where U is the universe of objects of interest, ≤t is the truth ordering defined in (3), and ≤k is the knowledge ordering as in the Belnap’s logic, gives the meaning of the logical connectives and is used in our approach. Example 4. Referring to Example 3, we then compute the logical values associated with the queries (1) and (2) by considering the new truth ordering above. The first query, (1) of Example 3, B OUNDARY ∨t U NKNOWN , returns the logical B OUNDARY which better corresponds to the intuitions. For the second query, (2) of Example 3, we have that B OUNDARY ∧t U NKNOWN = U NKNOWN . In contrast to Belnap’s logic, it is not excluded that some cars of the family of Example 3 are safe, but to be sure we need to obtain some information about the safety of car e. So, the answer U NKNOWN adequately reflects our intuitions. The proposition below shows that there is a connection between the knowledge ordering and the truth ordering. In this sense, the structure R can then be seen as a bilattice. Proposition 1. Consider the bilattice R = U, ≤t , ≤k and that x, y ∈ U . The operation ∧t is monotonic with respect to ≤k on both arguments, i.e. if x ≤k y then, for every z ∈ U , we have (z ∧t x) ≤k (z ∧t y) and (x ∧t z) ≤k (y ∧t z). Proof. Table 2 shows the result. Operation ∧t is obviously commutative.

We now define formally the logic underlying our work, called Rough Logic. Definition 2. Consider the following negation operation ¬. def

def

¬T RUE = FALSE, ¬FALSE = T RUE , def

def

¬B OUNDARY = B OUNDARY , ¬U NKNOWN = U NKNOWN. The propositional four-valued logic defined by the bilattice R together with negation ¬ is called the Rough Logic.

A Four-Valued Logic for Rough Set-Like Approximate Reasoning

183

Table 2. The table considered in the proof of Proposition 1 z B OUNDARY B OUNDARY B OUNDARY B OUNDARY B OUNDARY z FALSE FALSE FALSE FALSE FALSE z T RUE T RUE T RUE T RUE T RUE

x U NKNOWN U NKNOWN FALSE T RUE U NKNOWN

x U NKNOWN U NKNOWN FALSE T RUE U NKNOWN x U NKNOWN U NKNOWN FALSE T RUE U NKNOWN

z U NKNOWN U NKNOWN U NKNOWN U NKNOWN U NKNOWN

y T RUE FALSE B OUNDARY B OUNDARY B OUNDARY

y T RUE FALSE B OUNDARY B OUNDARY B OUNDARY y T RUE FALSE B OUNDARY B OUNDARY B OUNDARY

x U NKNOWN U NKNOWN FALSE T RUE U NKNOWN

z ∧t x U NKNOWN U NKNOWN FALSE B OUNDARY U NKNOWN

z ∧t x U NKNOWN U NKNOWN FALSE FALSE U NKNOWN z ∧t x U NKNOWN U NKNOWN FALSE T RUE U NKNOWN

y T RUE FALSE B OUNDARY B OUNDARY B OUNDARY

z ∧t y B OUNDARY FALSE B OUNDARY B OUNDARY B OUNDARY

z ∧t y FALSE FALSE FALSE FALSE FALSE z ∧t y T RUE FALSE B OUNDARY B OUNDARY B OUNDARY

z ∧t x U NKNOWN U NKNOWN U NKNOWN U NKNOWN U NKNOWN

z ∧t y U NKNOWN U NKNOWN U NKNOWN U NKNOWN U NKNOWN

4 Rough Datalog Language We now define a rule language, called Rough Datalog, such that its semantics is based on the Rough Logic. Intuitively, Rough Datalog corresponds to the usual logic programming language Datalog. While predicates in the latter denote crisp relations, in Rough Datalog a predicate p denotes a rough relation P . Thus, Rough Datalog caters for uncertainty in the knowledge. A rough literal is any expression of the form p(t1 , . . . , tn ) or ¬p(t1 , . . . , tn ). In Rough Datalog, knowledge is represented in the form of rough clauses, H:- B1 , . . . , Bn . where H and every Bi (0 ≤ i ≤ n) is a rough literal. A rough clause with the empty body (i.e. n = 0) is called a rough fact. A rough program P is a finite set of rough clauses. Rough clauses are used to specify rough relations as explained next. Intuitively, a rough clause is to be understood as the knowledge inequality ≤k stating that the truth value of the body is less than or equal to the truth value of the head. The comma

184

J. Małuszy´nski, A. Szałas, and A. Vit´oria

symbol “,” is interpreted as the meet in the truth ordering ≤t . Notice that the arguments of ≤k are the truth values U NKNOWN, B OUNDARY , T RUE, or FALSE but the logical value associated with a rough clause is either T RUE or FALSE. Information obtained from different rough clauses with heads referring to the same rough relation P (i.e. p or ¬p occurs in the head) is combined using the knowledge join operation ∨k . Example 5. The following rough clauses belong to an exemplary rough program P. (1) ¬useful(a) :- red(a), squared(a). “Object a is not useful if it is red and squared.” (2) squared(a) :- useful(a). —“Object a is squared if it is useful.” (3) ¬squared(a). —“Object a is not squared.”

4.1 Semantics of Rough Datalog Programs We now define notions of four-valued interpretation and model, extend the knowledge ordering to interpretations and show that each rough program has the least model in this ordering. Let P be a rough program and L be the set of all constant symbols occurring in P. Then, the Herbrand base HP is the set of all literals whose predicate symbols occur in P and whose arguments belong to L. A four-valued interpretation I of a rough program P associates with each atom a occurring in P a logical value I(a) ∈ {U NKNOWN, T RUE, FALSE, B OUNDARY } and ¬I(a) = I(¬a). The notion of interpretation extends naturally to conjunction (disjunction) of literals. Let l1 , . . . , ln , with n > 0, be rough literals. I(l1 ∧t · · · ∧t ln ) = I(l1 ) ∧t · · · ∧t I(ln ) . Definition 3. An interpretation I of a rough program P is any subset of the Herbrand base HP . Moreover, the rough relation I(p) is defined as I(p) = I(p), ¬I(p) = {t | p(t) ∈ I}, {t | ¬p(t) ∈ I} .

Intuitively, an interpretation associates each predicate p occurring in a program P with a rough set. Notice that ¬I(p) = I(¬p). Moreover, we have that – I(p(t)) = U NKNOWN, if t ∈ I(p) ∪ ¬I(p). – I(p(t)) = FALSE, if t ∈ ¬I(p). – I(p(t)) = T RUE, if t ∈ I(p). – I(p(t)) = B OUNDARY , if t ∈ I(p). Notice that we only consider variable-free rough programs. However, the results presented below can be also extended to rough programs with variables.

A Four-Valued Logic for Rough Set-Like Approximate Reasoning

185

An interpretation I of a rough program P satisfies a rough clause H:- B. ∈ P if I(B) ≤k I(H). A model M of P is any interpretation that satisfies every rough clause belonging to P. Notice also that the Herbrand base HP is a model of any rough program P. In this model the truth value of every literal is B OUNDARY . However, usually a program has more models. For comparing them we introduce a partial order on interpretations based on the knowledge ordering relation, ≤k . Definition 4. Let I1 ⊆ HP and I2 ⊆ HP be two interpretations. Then, I1 ≤k I2 , if and only if I1 (l) ≤k I2 (l), for every literal l ∈ HP . It can be checked that the knowledge ordering on interpretations corresponds to set inclusion. Proposition 2. I1 ≤k I2 if and only if I1 ⊆ I2 .

We show now that there is the least model for every rough program. Proposition 3. Let P be a rough program. Then, P has the least model with respect to ≤k . Proof. To prove that P has a least model with respect to ≤k , we show that the intersection of all models of P is also a model of P. n Let M = i Mi , where {M1 , . . . , Mn } (n ≥ 1) is the set of all models of P. Notice that, by Proposition 2, M ≤k Mi , with Mi ∈ {M1 , . . . , Mn }. We prove that M is a model of P. For this we have to show that, for any clause H :- B. ∈ P, we have M(H) ≥k M(B). We prove this by cases, considering possible truth values of the body of a clause. (a) If M(B) = U NKNOWN then M satisfies the rough clause, since U NKNOWN is the least element in the knowledge ordering. (b) If M(B) = T RUE then W (B) ≥t B OUNDARY , for every model W of P. Hence, W (H) ≥k T RUE, for every model W of P. Consequently, M(H) ≥k T RUE because the literal occurring in the head belongs to every model W . We conclude then that M satisfies the rough clause. (c) If M(B) = FALSE then B includes a literal l that is FALSE in some model of P and l is either FALSE or B OUNDARY in the other models. Obviously, no literal occurring in B can be U NKNOWN in any model. Consequently, M(H) ≥k FALSE because ¬H belongs to every model W . We conclude then that M satisfies the rough clause. (d) If M(B) = B OUNDARY then W (B) = B OUNDARY , for every model W of P. Notice that if I(B) = B OUNDARY , for some interpretation I of P, then we have that either I(l) = T RUE or I(l) = B OUNDARY , for every literal l in the body B. Hence, W (H) = B OUNDARY , for every model W of P. Consequently, M(H) = B OUNDARY because {H, ¬H} ⊆ W , for every model W . We conclude then that M satisfies the rough clause. The semantics of a rough program P is captured by its least model, with respect to ≤k .

186

J. Małuszy´nski, A. Szałas, and A. Vit´oria

Example 6. Consider again the rough program of Example 5. Its least model is M = {¬squared(a)}. Hence, useful(a) and red(a) are U NKNOWN, while squared(a) is FALSE. 4.2 A Fixpoint Characterization of the Least Model We now give a fixpoint characterization of the least model which makes it possible to compute the semantics of a program. We define an operator on interpretations, considered as sets of literals. We show that the operator is monotonic with respect to set inclusion. Thus, it has the least fixpoint (with respect to set inclusion) which can be obtained by iterations of the operator starting with the empty interpretation. We also show that the least fixpoint is a model. Taking into account Proposition 2, we can then conclude that the least fixpoint is also the least model of the program with respect to knowledge ordering. In the following definition if l is a negative literal of the form ¬a, then ¬l denotes a. Definition 5. Let P be a rough program. A total function TP mapping interpretations into interpretations is defined as follows: TP (I) = {l | l:- B. ∈ P and I(B) = T RUE} ∪ ∪ {¬l | l:- B. ∈ P and I(B) = FALSE} {l, ¬l | l:- B. ∈ P and I(B) = B OUNDARY } .

Thus, the set TP (I) consists of the heads of the rough clauses whose bodies are T RUE or B OUNDARY in I and, the negated heads of the rules whose bodies are FALSE or B OUNDARY in I. Such a way to gather heads of rules corresponds to defining the result by the disjunction of heads w.r.t. knowledge ordering ≤k . Proposition 4. Given a rough program P, the operator TP is monotonic with respect to set inclusion. Proof. The bodies of the program clauses are conjunctions of atoms. By Proposition 1 the conjunction is monotonic with respect to knowledge ordering. Hence by Proposition 2, it is also monotonic with respect to set inclusion of the interpretations. Thus, I ⊆ TP (I), for every interpretation I. The proposition above guarantees that TP has a least fixpoint (with respect to set inclusion), denoted as LFP(TP ). Proposition 5. Given a rough program P, the LFP(TP ) coincides with the least model of P. Proof. It is easy to see that the interpretation I = LFP(TP ) is a model of P. Assume the contrary. Then, there exists a clause H:- B. such that I(H) 0 , (18) μINCL(A,B) (x) = 0 otherwise . By assuming that μIncl(A,B) (x) = 0, for μA (x) = 0, we take into account only the support of the set A. For the sake of simplicity of the computational algorithm, it is not necessary to consider inclusion for all elements of the universe. Furthermore, we can require that the degree of inclusion with respect to x should be equal to 1, if the inequality μA (x) ≤ μB (x) for that x is satisﬁed I(μA (x), μB (x)) = 1,

if μA (x) ≤ μB (x) .

(19)

The requirement (19) is always satisﬁed by residual implicators. In order to deﬁne a suitable fuzzy counterpart of the rough inclusion function (12), we apply the notions of α-cut, power (cardinality) and support of a fuzzy set. Given a fuzzy subset A of the universe U , the α-cut of A, denoted by Aα , is a crisp set deﬁned as follows Aα = {x ∈ U : μA (x) ≥ α}

for

α ∈ [0, 1] .

(20)

For a ﬁnite fuzzy set A with n elements, power(A) and support(A) are given by power(A) =

n i=1

μA (xi ) ,

support(A) = {x : μA (xi ) > 0} .

(21)

198

A. Mieszkowicz-Rolka and L. Rolka

Using the above notions, we deﬁne the fuzzy rough inclusion function on the Cartesian product of the families F(U ) of all fuzzy subsets of the universe U να : F(U ) × F(U ) → [0, 1] .

(22)

Definition 5. The fuzzy rough α-inclusion function να (A, B) of any nonempty fuzzy set A in a fuzzy set B is defined as follows, for any α ∈ (0, 1] να (A, B) =

power(A ∩ INCL(A, B)α ) , power(A)

(23)

The value να (A, B) expresses how many elements of the nonempty fuzzy set A belong, at least to the degree α, to the fuzzy set B. First, we prove monotonicity of the proposed fuzzy rough inclusion function. Theorem 1. Implication-based fuzzy rough inclusion function να is monotonic with respect to the second parameter, for any α ∈ (0, 1] να (X, Y ) ≤ να (X, Z)

for any Y ⊆ Z,

where X, Y, Z ⊆ F(U ) .

Proof. According to the deﬁnition of a fuzzy subset [14], for Y ⊆ Z, we have μY (x) ≤ μZ (x), ∀ x ∈ U . Since every R-implicator, S-implicator and QL-implicator is right monotonic [25], it holds that: μI(X,Y ) (x) ≤ μI(X,Z) (x), ∀ x ∈ U . Thus, using the deﬁnition (18), we get μINCL(X,Y ) (x) ≤ μINCL(X,Z) (x),

∀x ∈ U .

Finally, for any α ∈ (0, 1], we can easy show that power(X ∩ INCL(X, Y )α ) power(X ∩ INCL(X, Z)α ) ≤ . power(X) power(X) Hence να (X, Y ) ≤ να (X, Z).

Furthermore, we can show that the rough inclusion function used in formulae (10) and (11) is a special case of the fuzzy rough inclusion function (23), when we use the implication-based inclusion set. Theorem 2. For any nonempty crisp set A, any crisp set B, and for α ∈ (0, 1], the implication-based inclusion function να (A, B) is equal to the inclusion degree incl(A, B). Proof. We show that for any crisp sets A and B, the inclusion set Incl(A, B) is equal to the crisp intersection A ∩ B. The membership function of any crisp set X is given by 1 for x ∈ X (24) μX (x) = 0 for x ∈ /X. Every implicator I satisﬁes the conditions: I(1, 1) = I(0, 1) = I(0, 0) = 1, and I(1, 0) = 0.

On Representation and Analysis of Crisp and Fuzzy Information Systems

Thus, applying the deﬁnition (18), we get 1 μIncl(A,B) (x) = μA∩B (x) = 0

if x ∈ A and x ∈ B otherwise .

199

(25)

For any ﬁnite crisp set X, and any α ∈ (0, 1], by formulae and (20), (21) and (24) we get: power(X) = card(X), and Xα = X. Consequently, applying (25), we ﬁnally have card(A ∩ B) power(A ∩ Incl(A, B)α ) = . power(A) card(A) Hence, we proved that να (A, B) = incl(A, B), for any α ∈ (0, 1].

We want to formulate the fuzzy rough approximation in a general way. Therefore, we introduce a function called res, deﬁned on the Cartesian product P(U )×F(U ), where P(U ) denotes the powerset of the universe U , and F(U ) the family of all fuzzy subsets of the universe U , respectively res : P(U ) × F(U ) → [0, 1] .

(26)

We require that res(∅, Y ) = 0 , res(X, Y ) ∈ {0, 1}, if Y is a crisp set , res(X, Y ) ≤ res(X, Z) for any Y ⊆ Z, where X ∈ P(U ), and Y, Z ∈ F(U ) . The form of the function res can be chosen depending on requirements of a considered application. For a given crisp set X and fuzzy set Y , the value of function res(X, Y ) should express the resulting membership degree in the set Y , taking into account not all elements of the universe, but only the elements of the set X. When we apply the limit-based approach, according to Dubois and Prade, we obtain the following form of the function res res(X, Y ) = inf μY (x) . x∈X

(27)

In the deﬁnition (27) of the function res, only one (limit) value of membership degree of elements in the set Y is taken into account. However, this means that we disregard the character (shape) of the membership function. Basing on a single value of membership degree is not always acceptable, especially in the case of large information systems. Hence, we can use the opportunity of giving another deﬁnitions of res, in which many values of membership degree are considered. Now, we introduce the notion of generalized fuzzy rough ε-approximation. Definition 6. For ε ∈ (0, 1], the ε-approximation Φε (A) of a fuzzy set A, by a fuzzy partition Φ = {F1 , F2 , . . . , Fn }, is a fuzzy set on the domain Φ with membership function expressed as μΦε (A) (Fi ) = res(Sε (Fi , A), INCL(Fi , A)) ,

(28)

200

A. Mieszkowicz-Rolka and L. Rolka

where Sε (Fi , A) = support(Fi ∩ INCL(Fi , A)αε ) , αε = sup{α ∈ [0, 1] : να (Fi , A) ≥ ε} . The set Sε (Fi , A) is equal to support of the intersection of the class Fi with the part of INCL(Fi , A), which contains those elements of the approximating class Fi which are included in A at least to the degree αε . The resulting membership μΦε (A) (Fi ) is determined using only the elements from Sε (Fi , A) instead of the whole class Fi . This is accomplished by applying the function res. It can be easy shown that applying the deﬁnition (27) of the function res leads to a simple form of the ε-approximation (28) μΦε (A) (Fi ) = sup{α ∈ [0, 1] : να (Fi , A) ≥ ε} .

(29)

In contrast to the approximations (16) and (17), which use two diﬀerent fuzzy connectives, we have a single uniﬁed deﬁnition of fuzzy rough approximation. In this way we obtain a consistent variable precision fuzzy rough set model. Thus, we are able to compare approximations determined for various values of the parameter ε. 3.3

Analysis of Fuzzy Decision Tables

In the analysis of fuzzy decision tables, two fuzzy partitions are generated with the help of a suitable similarity relation. The partition obtained with respect to condition attributes is used for approximation of fuzzy similarity classes obtained with respect to decision attributes. It is necessary to address the problem of comparing objects described by fuzzy sets. This issue has been widely studied in the literature, see, for example, [3,7,8]. In our considerations, we apply a symmetric T-transitive fuzzy similarity relation [3], which is deﬁned by means of the distance between the compared elements. In the following, we only give formulae for condition attributes. We apply the notation given in section 2. If we need to compare any two elements x and y of the universe U with respect to the condition attribute ci , i = 1, 2, . . . , n, then the similarity between x and y could be expressed using a T-similarity relation based on the Łukasiewicz T-norm [7]. (30) Sci (x, y) = 1 − max |μVik (x) − μVik (y)| . k=1,ni

In order to evaluate the similarity SC (x, y), with respect to all condition attributes C, we must aggregate the results obtained for particular attributes ci , i = 1, 2, . . . , n. This can be done by using the T-norm operator min as follows SC (x, y) = min Sci (x, y) = min (1 − max |μVik (x) − μVik (y)|) . i=1,n

i=1,n

k=1,ni

(31)

By calculating the similarity for all pairs of elements of the universe U , we obtain a symmetric similarity matrix. If the value of similarity between the elements x and y is equal to 1, they belong to the same similarity class. In that case two rows

On Representation and Analysis of Crisp and Fuzzy Information Systems

201

of the similarity matrix should be merged into one fuzzy set with membership degrees equal to 1 for x and y. In consequence, we get a family of fuzzy similarity = {C1 , C2 , . . . , Cn }, for condition attributes C and a family of fuzzy classes C = {D1 , D2 , . . . , Dm }, for decision attributes D. similarity classes D In the next step, we determine fuzzy rough approximations of elements of the by the family C, using the parameterized fuzzy rough set model. family D To determine the consistency of fuzzy decision tables and signiﬁcance of attributes, we apply a generalized measure of ε-approximation quality [17]. For = {C1 , C2 , . . . , Cn } the = {D1 , D2 , . . . , Dm } and the family C the family D ε-approximation quality of D by C is deﬁned as follows = γCε (D) where

= PosCε (D)

power(PosCε (D)) card(U )

,

(32)

ε (Dj )) ∩ Dj . ω(C

Dj ∈D

into the domain The fuzzy extension ω denotes a mapping from the domain C of the universe U , which is expressed for any fuzzy set X by μω(X) (x) = μX (Ci ),

if μCi (x) = 1 .

(33)

The deﬁnition (32) is based on the generalized notion of positive region. For any fuzzy set X and a similarity relation R, the positive region of X is deﬁned as follows (34) PosRε (X) = X ∩ ω(Rε (X)) . In the deﬁnition of the positive region (34), we take into account only those elements of the ε-approximation, for which there is no contradiction between the set X and the approximating similarity classes.

4

Fuzzy Flow Graphs

In addition to the VPFRS model, we want to introduce fuzzy ﬂow graphs as a second tool for analysis of fuzzy information systems. The idea of applying ﬂow graphs in the framework of crisp rough sets, for discovering the statistical properties of decision algorithms, was proposed by Pawlak [20,21,22]. We should start with recalling the basic notions of the crisp ﬂow graph approach. A ﬂow graph is given in the form of directed acyclic ﬁnal graph G = (N , B, ϕ), where N is a set of nodes, B ⊆ N × N is a set of directed branches, ϕ: B → R+ is a ﬂow function with values in the set of non-negative reals R+ . For any (X, Y ) ∈ B, X is an input of Y and Y is an output of X. The quantity ϕ(X, Y ) is called the throughﬂow from X to Y . I(X) and O(X) denote an input and an output of X, respectively. The input I(G) and output O(G) of a graph G are deﬁned by I(G) = {X ∈ N : I(X) = ∅} ,

O(G) = {X ∈ N : O(X) = ∅} .

(35)

202

A. Mieszkowicz-Rolka and L. Rolka

Every node X ∈ N of a ﬂow graph G is characterized by its inﬂow ϕ(Y, X) , ϕ+ (X) =

(36)

Y ∈I(X)

and by its outﬂow ϕ− (X) =

(37)

ϕ(X, Y ) .

Y ∈O(X)

For any internal node X, the equality ϕ+ (X) = ϕ− (X) = ϕ(X) is satisﬁed. The quantity ϕ(X) is called the ﬂow of the node X. The ﬂow for the whole graph G is deﬁned by ϕ− (X) = ϕ+ (X) . (38) ϕ(G) = x∈I(G)

x∈O(G)

By using the ﬂow ϕ(G), the normalized throughﬂow σ(X, Y ) and the normalized ﬂow σ(X) are determined as follows σ(X, Y ) =

ϕ(X, Y ) , ϕ(G)

σ(X) =

ϕ(X) . ϕ(G)

(39)

For every branch of a ﬂow graph G the certainty factor is deﬁned by cer(X, Y ) =

σ(X, Y ) . σ(X)

(40)

The coverage factor for every branch of a ﬂow graph G is deﬁned by cov(X, Y ) =

σ(X, Y ) . σ(Y )

The certainty and coverage factors satisfy the following properties cer(X, Y ) = 1 , cov(X, Y ) = 1 . Y ∈O(X)

(41)

(42)

X∈I(Y )

The measures of certainty (40) and coverage (41) are useful for analysis of decision algorithms [10]. Now, we consider the issue of applying ﬂow graphs to representation and analysis of fuzzy decision algorithms. We use decision tables with fuzzy values of attributes, presented in section 2. All possible decision rules, generated by the Cartesian product of sets of linguistic values of the attributes, will m be examined. n According to notation used in section 2, we obtain r = i=1 ni j=1 mj possible rules. The k-th decision rule, denoted by Rk , is expressed as follows Rk : IF c1 is V1k AND c2 is V2k . . . AND cn is Vnk k THEN d1 is W1k AND d2 is W2k . . . AND dm is Wm

(43)

where k = 1, 2, . . . , r , Vik ∈ Vi , i = 1, 2, . . . n , Wjk ∈ Wj , j = 1, 2, . . . , m.

On Representation and Analysis of Crisp and Fuzzy Information Systems

203

When we use the fuzzy Cartesian products C k = V1k × V2k × . . . × Vnk and k D = W1k × W2k × . . . × Wm , the k-th decision rule can be expressed in the form of a fuzzy implication, denoted here by C k → Dk . It is necessary to select a subset of decision rules which are relevant to the considered decision process. This can be done by determining to what degree any element x ∈ U , corresponding to a single row of the decision table, conﬁrms particular decision rules. We calculate the truth value of the decision rule’s antecedent and the truth value of the decision rule’s consequent, by determining the conjunction of the respective membership degrees of x in the linguistic values of attributes. If we take a decision table with crisp attributes, a decision rule can be conﬁrmed for some x, if the result of conjunction is equal to 1, both for the rule’s premise and the rule’s conclusion. Otherwise, the element x does not conﬁrm the considered decision rule. The set of elements x ∈ U , which conﬁrm a decision rule, is called the support of the decision rule. To determine the conﬁrmation degree of fuzzy decision rules, a T-norm operator need to be applied. By cd(x, k), we denote the conﬁrmation degree of the k-th decision rule by the element x ∈ U k

cd(x, k) = T(cda(x, k), cdc(x, k)) ,

(44)

where cda(x, k) denotes the conﬁrmation degree of the decision rule’s antecedent cda(x, k) = T(μV1k (x), μV2k (x), . . . , μVnk (x)) ,

(45)

and cdc(x, k) the conﬁrmation degree of the decision rule’s consequent cdc(x, k) = T(μW1k (x), μW2k (x), . . . , μWm k (x)) .

(46)

Through determining the conﬁrmation degrees (45), (46) and (44), we generate the following fuzzy sets on the domain U : the support of the decision rule’s antecedent support(cda(x, k)) = {cda(x1 , k)/x1 , cda(x2 , k)/x2 , . . . , cda(xN , k)/xN }, (47) the support of the decision rule’s consequent support(cdc(x, k)) = {cdc(x1 , k)/x1 , cdc(x2 , k)/x2 , . . . , cda(xN , k)/xN } , (48) and the support of the decision rule Rk , respectively support(Rk ) = {cd(x1 , k)/x1 , cd(x2 , k)/x2 , . . . , cd(xN , k)/xN } .

(49)

The introduced notions (47), (48) and (49) will be used for deﬁning strength, certainty, and coverage factors of a decision rule. Now, let us explain the way of constructing fuzzy ﬂow graphs on the basis of a decision table with fuzzy attributes.

204

A. Mieszkowicz-Rolka and L. Rolka

Every fuzzy attribute is represented by a layer of nodes. The nodes of a layer correspond to linguistic values of a given attribute. a fuzzy set on the universe U , which describes membership We denote by X degree of particular elements x ∈ U in the linguistic value represented by X. The can be found in a respective column of membership degrees of all x in the set X the considered decision table. Let us pick out such two attributes, which are represented by two consecutive layers of the ﬂow graph. We denote by X a linguistic value of the ﬁrst attribute, and by Y a linguistic value of the second attribute. In the case of crisp ﬂow graphs, the ﬂow between nodes X and Y is equal to the number of elements of the universe U , which are characterized by the combination of attribute values X and Y . In consequence, a particular element x ∈ U can only be assigned to a unique path in the ﬂow graph. In a fuzzy information system, however, every element of the universe can belong to several linguistic values, and it can be assigned to several paths in the ﬂow graph. It is possible to determine the ﬂow distribution in the crisp ﬂow graph by using the operations of set intersection and set cardinality. To obtain the ﬂow ϕ(X, Y ) for the branch (X, Y ) of a fuzzy ﬂow graph, we have to calculate power of the in and Y . Many deﬁnitions of fuzzy intersection (T-norm tersection of fuzzy sets X operator) are known. In order to satisfy the ﬂow conservation equations, it is necessary to use the T-norm operator prod for determining the intersection of sets. Furthermore, we should assume that the linguistic values of attributes satisfy the requirement (3). We conclude the above discussion with the following theorem. Theorem 3. Let S be a fuzzy information systems with the linguistic values of attributes satisfying the requirement (3), and let ∩ denote a fuzzy intersection operator based on the T-norm prod. The following properties are satisfied for the flow graph, which represents the information system S : (G1) the inflow for any output or internal layer node X is given by = ∩ Y ) , ϕ(Y, X) = power(X ϕ+ (X) = power(X) Y ∈I(X)

Y ∈I(X)

(G2) the outflow for any input or internal layer node X is given by = ∩ Y ) , ϕ− (X) = power(X) ϕ(X, Y ) = power(X Y ∈O(X)

(50)

(51)

Y ∈O(X)

(G3) for any internal layer node X, it holds that ϕ+ (X) = ϕ− (X) .

(52)

The properties (G1), (G2) and (G3) do not hold in general, if we use another Tnorm operator, e.g. min. In the special case of crisp decision tables, the formulae (50) and (51) become equivalent to (36) and (37). The layers corresponding to condition attributes can be merged into a single layer, which contains nodes representing all possible combinations of linguistic

On Representation and Analysis of Crisp and Fuzzy Information Systems

205

values of the condition attributes. We can also merge all the layers corresponding to decision attributes. Let us denote by X ∗ , a node of the resulting layer obtained for condition attributes and by Y ∗ , a node of the resulting layer obtained for decision attributes. The node X ∗ corresponds to antecedent of some decision rule Rk . Support of the antecedent of the decision rule Rk is determined with the help of formula (47). The decision rule Rk is represented by the branch (X ∗ , Y ∗ ). Power of the support of the rule Rk is equal to the ﬂow between the nodes X ∗ and Y ∗ , which is obtained using formula (49) ϕ(X ∗ , Y ∗ ) = power(support(Rk )) .

(53)

By applying the formulae (47), (48) and (49), we can determine, for every decision rule Rk , the certainty factor cer(X ∗ , Y ∗ ), the coverage factor cov(X ∗ , Y ∗ ), and the strength of the rule σ(X ∗ , Y ∗ ) cer(X ∗ , Y ∗ ) = cer(Rk ) =

power(support(Rk )) , power(support(cda(x, k)))

(54)

cov(X ∗ , Y ∗ ) = cov(Rk ) =

power(support(Rk )) , power(support(cdc(x, k)))

(55)

σ(X ∗ , Y ∗ ) = strength(Rk ) =

power(support(Rk )) . card(U )

(56)

It is possible to represent any decision rule by a sequence of nodes [X1 . . . Xn ], namely by a path from the 1-th to the n-th layer of the ﬂow graph G. For a given path [X1 . . . Xn ], the resulting certainty and strength can be deﬁned. In contrast to the deﬁnitions presented in [20,21,22], in which the statistical properties of ﬂow are taken into account, we propose a diﬀerent form of the path’s certainty and strength n−1 cer[X1 . . . Xn ] = cer(X1 . . . Xi , Xi+1 ) , (57) i=1

σ[X1 . . . Xn ] = σ(X1 ) cer[X1 . . . Xn ] , where cer(X1 . . . Xi , Xi+1 ) =

2 ∩ . . . ∩ X i+1 ) 1 ∩ X power(X . i ) power(X1 ∩ X2 ∩ . . . ∩ X

(58)

(59)

The resulting certainty (57) of the path [X1 . . . Xn ], expresses what part of the ﬂow of the starting node X1 reaches the ﬁnal node Xn , passing through all nodes of the path.

5

Examples

Let us analyze a fuzzy decision table (Table 1) with condition attributes c1 and c2 and one decision attribute d. All attributes have three linguistic values.

206

A. Mieszkowicz-Rolka and L. Rolka Table 1. Decision table with fuzzy attributes c1

x1 x2 x3 x4 x5 x6 x7 x8 x9 x10

c2

d

V11

V12

V13

V21

V22

V23

W11

W12

W13

0.1 0.8 0.0 0.1 0.0 0.8 0.1 0.0 0.0 0.1

0.9 0.2 0.2 0.9 0.8 0.2 0.9 0.1 0.2 0.9

0.0 0.0 0.8 0.0 0.2 0.0 0.0 0.9 0.8 0.0

0.0 1.0 0.0 0.0 0.8 0.0 0.0 0.8 0.0 0.1

0.9 0.0 0.2 0.9 0.2 0.2 0.9 0.2 0.2 0.9

0.1 0.0 0.8 0.1 0.0 0.8 0.1 0.0 0.8 0.0

0.0 0.0 0.9 0.0 0.0 1.0 0.1 0.0 0.9 0.0

1.0 0.1 0.1 1.0 0.1 0.0 0.9 0.0 0.1 0.9

0.0 0.9 0.0 0.0 0.9 0.0 0.0 1.0 0.0 0.1

First, we apply the variable precision fuzzy rough set approach. Using similarity relation in the form (31), we determine similarity matrices with respect to condition and decision attributes. By merging identical rows of the similarity matrix, we get 9 condition similarity classes and and 6 decision similarity classes. We calculate ε-approximation quality using the Łukasiewicz implication operator. The results are presented in table 2. Table 2. ε-approximation quality for diﬀerent values of parameter ε

Method

Ł-inf

γCε (D)

Removed attribute

ε=1

ε = 0.9

ε = 0.85

ε = 0.8

none c1 c2

0.830 0.820 0.250

0.900 0.880 0.250

0.900 0.880 0.410

0.910 0.910 0.450

We can state that the considered information system has a high consistency. The condition attribute c1 can be omitted from the decision table without a signiﬁcant decrease of the ε-approximation quality. In the next step, the ﬂow graph method will be applied. We use the same labels for both the linguistic values of the attributes and the corresponding nodes of the ﬂow graph. As stated in previous section, the T-norm operator prod should be used in our calculations. The obtained fuzzy ﬂow graph has a very simple form, because there is only one condition attribute c2 and one decision attribute d. Values of the normalized ﬂow between nodes of the condition layer and nodes of the decision layer are given in Table 3.

On Representation and Analysis of Crisp and Fuzzy Information Systems

207

Table 3. Normalized ﬂow between nodes of condition and decision layers σ(V2i , W1j )

V21 V22 V23 Σ

W11

W12

W13

Σ

0.000 0.065 0.225 0.290

0.027 0.348 0.045 0.420

0.243 0.047 0.000 0.290

0.270 0.460 0.270 1.000

We see that the ﬂow conservation equations (50) and (51), are satisﬁed, for example, 3 power(V21 ) σ− (V21 ) = = σ(V21 , W1i ) = 0.270 , card(U ) i=1 11 ) power(W = σ(V2i , W11 ) = 0.290 . card(U ) i=1 3

σ+ (W11 ) =

Let us determine the certainty and coverage factors for branches between the layers according to formulae (54), (55). The results are given in Tables 4 and 5. Table 4. Certainty factor for branches between condition and decision layers cer(V2i , W1j )

V21 V22 V23

W11

W12

W13

Σ

0.0000 0.1413 0.8333

0.1000 0.7565 0.1667

0.9000 0.1022 0.0000

1.0000 1.0000 1.0000

Table 5. Coverage factor for branches between condition and decision layers cov(V2i , W1j )

V21 V22 V23 Σ

W11

W12

W13

0.0000 0.2241 0.7759 1.0000

0.0643 0.8286 0.1071 1.0000

0.8379 0.1621 0.0000 1.0000

Fuzzy decision rules with the largest values of certainty factor (Table 6) can be included in the ﬁnal fuzzy inference system. The respective values of coverage factor are useful for explaining the selected decision rules. Only 3 decision rules

208

A. Mieszkowicz-Rolka and L. Rolka Table 6. Decision rules with the largest value of certainty factor decision rule

certainty

coverage

V21 → W13 V22 → W12 V23 → W11

0.9000 0.7565 0.8333

0.8379 0.8286 0.7759

strength [%] 24.30 34.80 22.50

could be generated from our decision table. Owing to the application of the VPFRS approach, we got a simple fuzzy ﬂow graph. Let us construct a ﬂow graph without a prior reduction of attributes. We merge the layers corresponding to condition attributes c1 and c2 to a resulting layer, which represents all possible linguistic values in the antecedences of decision rules. We determine the degrees of satisfaction of the rules’ antecedences for particular elements x ∈ U . For the antecedence represented by V12 V22 , we get: V 12 V22 = V12 ∩ V22 = { 0.81/x1 , 0.00/x2 , 0.04/x3 , 0.81/x4 , 0.16/x5 , 0.04/x6 , 0.81/x7 , 0.02/x8, 0.04/x9 , 0.81/x10 }, ϕ(V12 , V22 ) = power(V 12 V22 ) = 3.54, σ(V12 , V22 ) =

ϕ(V12 ,V22 ) cardU

= 0.354.

Table 7. Decision rules with the largest certainty factor (full information system) decision rule V11 V21 V11 V23 V12 V21 V12 V22 V13 V21 V13 V23

→ W13 → W11 → W13 → W12 → W13 → W11

certainty

coverage

0.8901 0.9567 0.8366 0.8763 0.9818 0.9000

0.2486 0.2210 0.2914 0.7386 0.2979 0.3972

strength [%] 7.21 6.41 8.45 31.02 8.64 11.52

Finally, we determine the normalized throughﬂow, certainty and coverage factors for branches between of the resulting condition and decision layers. Decision rules with the largest value of certainty factor are given in Table 7. We can observe that the attribute c1 is superﬂuous in the obtained decision rules.

6

Conclusions

Information systems with crisp and fuzzy attributes can be eﬀectively analyzed by a hybrid approach which combines the variable precision fuzzy rough set (VPFRS) model with fuzzy ﬂow graphs. The VPFRS model can be deﬁned in a uniﬁed way with the help of a single notion of ε-approximation. This allows to

On Representation and Analysis of Crisp and Fuzzy Information Systems

209

avoid the inconsistency of the VPFRS model caused by diﬀerent forms of fuzzy connectives. The proposed fuzzy ﬂow graph method is suitable for representing and analyzing decision tables with fuzzy attributes. Every fuzzy attribute can be represented by a layer of a ﬂow graph. All nodes of a layer correspond to linguistic values of an attribute. A fuzzy decision table can be reduced by applying the VPFRS approach prior to using the fuzzy ﬂow graph method for determining a system of fuzzy decision rules.

References 1. Bandler, W., Kohout, L.: Fuzzy Power Sets and Fuzzy Implication Operators. Fuzzy Sets and Systems 4 (1980) 13–30 2. Burillo, P., Frago, N., Fuentes, R.: Inclusion Grade and Fuzzy Implication Operators. Fuzzy Sets and Systems 114 (2000) 417–429 3. Chen, S.M., Yeh, M.S., Hsiao, P.Y.: A Comparison of Similarity Measures of Fuzzy Values. Fuzzy Sets and Systems 72 (1995) 79–89 4. Cornelis, C., Van der Donck, C., Kerre, E.: Sinha-Dougherty Approach to the Fuzziﬁcation of Set Inclusion Revisited. Fuzzy Sets and Systems 134 (2003) 283–295 5. De Baets, B., De Meyer, H., Naessens, H.: On Rational Cardinality-based Inclusion Measures. Fuzzy Sets and Systems 128 (2002) 169–183 6. Dubois, D., Prade, H.: Putting Rough Sets and Fuzzy Sets Together. [30] 203–232 7. Fernández Salido, J.M., Murakami, S.: Rough Set Analysis of a General Type of Fuzzy Data Using Transitive Aggregations of Fuzzy Similarity Relations. Fuzzy Sets and Systems 139 (2003) 635–660 8. Greco, S., Matarazzo, B., Słowiński, R.: Rough Set Processing of Vague Information Using Fuzzy Similarity Relations. In: Calude, C.S., Paun, G., (eds.): Finite Versus Inﬁnite — Contributions to an Eternal Dilemma. Springer-Verlag, Berlin Heidelberg New York (2000) 149–173 9. Greco, S., Pawlak, Z., Słowiński, R.: Generalized Decision Algorithms, Rough Inference Rules, and Flow Graphs. In: Alpigini, J., Peters, J.F., Skowron, A., Zhong, N., (eds.): Rough Sets and Current Trends in Computing. Lecture Notes in Artiﬁcial Intelligence, Vol. 2475. Springer-Verlag, Berlin Heidelberg New York (2002) 93–104 10. Greco, S., Pawlak, Z., Słowiński, R.: Bayesian Conﬁrmation Measures within Rough Set Approach. [31] 264–273 11. Greco, S., Matarazzo, B., Słowiński, R.: Rough Membership and Bayesian Conﬁrmation Measures for Parameterized Rough Sets. [28] 314–324 12. Inuiguchi, M.: Generalizations of Rough Sets: From Crisp to Fuzzy Cases. [31] 26–37 13. Katzberg, J.D., Ziarko, W.: Variable Precision Extension of Rough Sets. Fundamenta Informaticae 27 (1996) 155–168 14. Klir, G.J., Folger, T.A.: Fuzzy Sets, Uncertainty, and Information. Prentice Hall, Englewood, New Jersey (1988) 15. Lin, T.Y.: Coping with Imprecision Information — Fuzzy Logic. Downsizing Expo, Santa Clara Convention Center (1993) 16. Mieszkowicz-Rolka, A., Rolka, L.: Variable Precision Rough Sets: Evaluation of Human Operator’s Decision Model. In: Sołdek, J., Drobiazgiewicz, L., (eds.): Artiﬁcial Intelligence and Security in Computing Systems. Kluwer Academic Publishers, Boston Dordrecht London (2003) 33–40

210

A. Mieszkowicz-Rolka and L. Rolka

17. Mieszkowicz-Rolka, A., Rolka, L.: Variable Precision Fuzzy Rough Sets Model in the Analysis of Process Data. [28] 354–363 18. Nakamura, A.: Application of Fuzzy-Rough Classiﬁcations to Logics. [30] 233–250 19. Pawlak, Z.: Rough Sets: Theoretical Aspects of Reasoning about Data. Kluwer Academic Publishers, Boston Dordrecht London (1991) 20. Pawlak, Z.: Decision Algorithms, Bayes’ Theorem and Flow Graphs. In: Rutkowski, L., Kacprzyk, J., (eds.): Advances in Soft Computing. Physica-Verlag, Heidelberg (2003) 18–24 21. Pawlak, Z.: Flow Graphs and Data Mining. [23] 1–36 22. Pawlak, Z.: Rough Sets and Flow Graphs. [28] 1–11 23. Peters, J.F., et al., (eds.): Transactions on Rough Sets III. Lecture Notes in Computer Science (Journal Subline), Vol. 3400. Springer-Verlag, Berlin Heidelberg New York (2005) 24. Polkowski, L.: Toward Rough Set Foundations. Mereological Approach. [31] 8–25 25. Radzikowska, A.M., Kerre, E.E.: A Comparative Study of Fuzzy Rough Sets. Fuzzy Sets and Systems 126 (2002) 137–155 26. Skowron, A., Stepaniuk, J.: Tolerance Approximation Spaces. Fundamenta Informaticae 27 (1996) 245–253 27. Ślęzak, D., Ziarko, W.: Variable Precision Bayesian Rough Set Model. In: Wang, G., Liu, Q., Yao, Y., Skowron, A., (eds.): Rough Sets, Fuzzy Sets, Data Mining, and Granular Computing. Lecture Notes in Artiﬁcial Intelligence, Vol. 2639. SpringerVerlag, Berlin Heidelberg New York (2003) 312–315 28. Ślęzak, D., et al., (eds.): Rough Sets and Current Trends in Computing. Lecture Notes in Artiﬁcial Intelligence, Vol. 3641. Springer-Verlag, Berlin Heidelberg New York (2005) 29. Ślęzak, D.: Rough Sets and Bayes Factor. [23] 202–229 30. Słowiński, R., (ed.): Intelligent Decision Support: Handbook of Applications and Advances of the Rough Sets Theory. Kluwer Academic Publishers, Boston Dordrecht London (1992) 31. Tsumoto, S., et al., (eds.): Rough Sets and Current Trends in Computing. Lecture Notes in Artiﬁcial Intelligence, Vol. 3066. Springer-Verlag, Berlin Heidelberg New York (2004) 32. Yager, R.R., Filev, D.P.: Essentials of Fuzzy Modelling and Control. John Wiley & Sons, Inc., New York (1994) 33. Liu, W.N., Yao, J., Yao, Y.: Rough Approximations under Level Fuzzy Sets. [31] 78–83 34. Zadeh, L.: Fuzzy Sets. Information and Control 8 (1965) 338–353 35. Ziarko, W.: Variable Precision Rough Sets Model. Journal of Computer and System Sciences 46 (1993) 39–59 36. Ziarko, W.: Probabilistic Rough Sets. [28] 283–293

On Partial Covers, Reducts and Decision Rules with Weights Mikhail Ju. Moshkov1, Marcin Piliszczuk2 , and Beata Zielosko3 1

2

Institute of Computer Science, University of Silesia 39, B¸edzi´ nska St., Sosnowiec, 41-200, Poland [email protected] ´ aski S.A., 34, Sokolska St., Katowice, 40-086, Poland ING Bank Sl¸ [email protected] 3 Institute of Computer Science, University of Silesia 39, B¸edzi´ nska St., Sosnowiec, 41-200, Poland [email protected]

Abstract. In the paper the accuracy of greedy algorithms with weights for construction of partial covers, reducts and decision rules is considered. Bounds on minimal weight of partial covers, reducts and decision rules based on an information on greedy algorithm work are studied. Results of experiments with software implementation of greedy algorithms are described. Keywords: partial covers, partial reducts, partial decision rules, weights, greedy algorithms.

1

Introduction

The paper is devoted to consideration of partial decision-relative reducts (we will omit often words ”decision-relative”) and partial decision rules for decision tables on the basis of partial cover investigation. Rough set theory [11,17] often deals with decision tables containing noisy data. In this case exact reducts and rules can be ”overlearned” i.e. depend essentially on noise. If we see constructed reducts and rules as a way of knowledge representation [16] then instead of large exact reducts and rules it is more appropriate to work with relatively small partial ones. In [12] Zdzislaw Pawlak wrote that ”the idea of an approximate reduct can be useful in cases when a smaller number of condition attributes is preferred over accuracy of classiﬁcation”. Last years in rough set theory partial reducts, partial decision rules and partial covers are studied intensively [6,7,8,9,10,13,19,20,21,22,23,24,27]. Approximate reducts are investigated also in extensions of rough set model such as VPRS (variable precision rough sets) [26] and α-RST (alpha rough set theory) [14]. We study the case where each subset, used for covering, has its own weight, and we must minimize the total weight of subsets in partial cover. The same situation is with partial reducts and decision rules: each conditional attribute has its own weight, and we must minimize the total weight of attributes in partial J.F. Peters et al. (Eds.): Transactions on Rough Sets VI, LNCS 4374, pp. 211–246, 2007. c Springer-Verlag Berlin Heidelberg 2007

212

M.Ju. Moshkov, M. Piliszczuk, and B. Zielosko

reduct or decision rule. If weight of each attribute characterizes time complexity of attribute value computation then we try to minimize total time complexity of computation of attributes from partial reduct or partial decision rule. If weight characterizes a risk of attribute value computation (as in medical or technical diagnosis) then we try to minimize total risk, etc. In rough set theory various problems can be represented as set cover problems with weights: – problem of construction of a reduct [16] or partial reduct with minimal total weight of attributes for an information system; – problem of construction of a decision-relative reduct [16] or partial decisionrelative reduct with minimal total weight of attributes for a decision table; – problem of construction of a decision rule or partial decision rule with minimal total weight of attributes for a row of a decision table (note that this problem is closely connected with the problem of construction of a local reduct [16] or partial local reduct with minimal total weight of attributes); – problem of construction of a subsystem of a given system of decision rules which ”covers” the same set of rows and has minimal total weight of rules (in the capacity of a rule weight we can consider its length). So the study of covers and partial covers is of some interest for rough set theory. In this paper we list some known results on set cover problems which can be useful in applications and obtain certain new results. From results obtained in [20,22] it follows that the problem of construction of partial cover with minimal weight is N P -hard. Therefore we must consider polynomial approximate algorithms for minimization of weight of partial covers. In [18] a greedy algorithm with weights for partial cover construction was investigated. This algorithm is a generalization of well known greedy algorithm with weights for exact cover construction [2]. The algorithm from [18] is a greedy algorithm with one threshold which gives the exactness of constructed partial cover. Using results from [9] (based on results from [3,15] and technique created in [20,22]) on precision of polynomial approximate algorithms for construction of partial cover with minimal cardinality and results from [18] on precision of greedy algorithm with one threshold we show that under some natural assumptions on the class N P the greedy algorithm with one threshold is close to best polynomial approximate algorithms for construction of partial cover with minimal weight. However we can try to improve results of the work of greedy algorithm with one threshold for some part of set cover problems with weight. We generalize greedy algorithm with one threshold [18], and consider greedy algorithm with two thresholds. First threshold gives the exactness of constructed partial cover, and the second one is an interior parameter of the considered algorithm. We prove that for the most part of set cover problems there exist a weight function and values of thresholds such that the weight of partial cover constructed by greedy algorithm with two thresholds is less than the weight of partial cover constructed by greedy algorithm with one threshold.

On Partial Covers, Reducts and Decision Rules with Weights

213

We describe two polynomial algorithms which always construct partial covers that are not worse than the one constructed by greedy algorithm with one threshold, and for the most part of set cover problems there exists a weight function and a value of ﬁrst threshold such that the weight of partial covers constructed by the considered two algorithms is less than the weight of partial cover constructed by greedy algorithm with one threshold. Information on greedy algorithm work can be used for obtaining lower bounds on minimal cardinality of partial covers [9]. We ﬁx some kind of information on greedy algorithm work, and ﬁnd unimprovable lower bound on minimal weight of partial cover depending on this information. Obtained results show that this bound is not trivial and can be useful for investigation of set cover problems. There exist bounds on precision of greedy algorithm without weights for partial cover construction which do not depend on the cardinality of covered set [1,6,7,8]. We obtain similar bound for the case of weight. The most part of the results obtained for partial covers is generalized on the case of partial decision-relative reducts and partial decision rules for decision tables which, in general case, are inconsistent (a decision table is inconsistent if it has equal rows with diﬀerent decisions). In particular, we show that – Under some natural assumptions on the class N P greedy algorithms with weights are close to best polynomial approximate algorithms for minimization of total weight of attributes in partial reducts and partial decision rules. – Based on information receiving during greedy algorithm work it is possible to obtain nontrivial lower bounds on minimal total weight of attributes in partial reducts and partial decision rules. – There exist polynomial modiﬁcations of greedy algorithms which for a part of decision tables give better results than usual greedy algorithms. Obtained results will further to more wide use of greedy algorithms with weighs and their modiﬁcations in rough set theory and applications. This paper is, in some sense, an extension of [9] on the case of weights which are not equal to 1. However, problems considered in this paper (and proofs of results) are more complicated than the ones considered in [9]. Bounds obtained in this paper are sometimes more weak than the corresponding bounds from [9]. We must note also that even if all weights are equal to 1 then results of the work of greedy algorithms considered in this paper can be diﬀerent from the results of the work of greedy algorithms considered in [9]. For example, for case of reducts the number of chosen attributes is the same, but last attributes can diﬀer. The paper consists of ﬁve sections. In Sect. 2 partial covers are studied. In Sect. 3 partial tests (partial superreducts) and partial reducts are investigated. In Sect. 4 partial decision rules are considered. Sect. 5 contains short conclusions.

2 2.1

Partial Covers Main Notions

Let A = {a1 , . . . , an } be a nonempty ﬁnite set. Elements of A are enumerated by numbers 1, . . . , n (in fact we ﬁx a linear order on A). Let S = {Bi }i∈{1,...,m} =

214

M.Ju. Moshkov, M. Piliszczuk, and B. Zielosko

{B1 , . . . , Bm } be a family of subsets of A such that B1 ∪ . . . ∪ Bm = A. We will assume that S can contain equal subsets of A. The pair (A, S) will be called a set cover problem. Let w be a weight function which corresponds to each Bi ∈ S a natural number w(Bi ). The triple (A, S, w) will be called a set cover problem with weights. Note that in fact weight function w is given on the set of indexes {1, . . . , m}. But, for simplicity, we are writing w(Bi ) instead of w(i). Let I be a subset of {1, . . . , m}. The family P = {Bi }i∈I will be called a subfamily of S. The number |P | = |I| will be called the cardinality of P . Let P = {Bi }i∈I and Q = {Bi }i∈J be subfamilies of S. The notation P ⊆ Q will mean that I ⊆ J. Let us denote P ∪ Q = {Bi }i∈I∪J , P ∩ Q = {Bi }i∈I∩J , and P \ Q = {Bi }i∈I\J . A subfamily Q = {Bi1 , . . . , Bit } of the family S will be called a partial cover for (A, S). Let α be a real number such that 0 ≤ α < 1. The subfamily Q will be called an α-cover for (A, S) if |Bi1 ∪ . . . ∪ Bit | ≥ (1 − α)|A|. For example, 0.01-cover means that we must cover at least 99% of elements from A. Note that t a 0-cover is usual (exact) cover. The number w(Q) = j=1 w(Bij ) will be called the weight of the partial cover Q. Let us denote by Cmin (α) = Cmin (α, A, S, w) the minimal weight of α-cover for (A, S). Let α and γ be real numbers such that 0 ≤ γ ≤ α < 1. Let us describe a greedy algorithm with two thresholds α and γ. Let us denote N = |A|(1 − γ) and M = |A|(1 − α) . Let we make i ≥ 0 steps and choose subsets Bj1 , . . . , Bji . Let us describe the step number i + 1. Let us denote D = Bj1 ∪ . . . ∪ Bji (if i = 0 then D = ∅). If |D| ≥ M then we ﬁnish the work of the algorithm. The family {Bj1 , . . . , Bji } is the constructed α-cover. Let |D| < M . Then we choose a subset Bji+1 from S with minimal number ji+1 for which Bji+1 \ D = ∅ and the value w(Bji+1 ) min{|Bji+1 \ D|, N − |D|} is minimal. Pass to the step number i + 2. γ γ Let us denote by Cgreedy (α) = Cgreedy (α, A, S, w) the weight of α-cover constructed by the considered algorithm for the set cover problem with weights (A, S, w). Note that greedy algorithm with two thresholds α and α coincides with the greedy algorithm with one threshold α considered in [18]. 2.2

Some Known Results

In this subsection we assume that the weight function has values from the set of positive real numbers. For natural m denote H(m) = 1 + . . . + 1/m. It is known that ln m ≤ H(m) ≤ ln m + 1 .

On Partial Covers, Reducts and Decision Rules with Weights

215

Consider some results for the case of exact covers where α = 0. In this case γ = 0. First results belong to Chv´ atal. Theorem 1. (Chv´ atal [2]) For any set cover problem with weights (A, S, w) the 0 inequality Cgreedy (0) ≤ Cmin (0)H(|A|) holds. Theorem 2. (Chv´ atal [2]) For any set cover problem with weights (A, S, w) the 0 (0) ≤ Cmin (0)H (maxBi ∈S |Bi |) holds. inequality Cgreedy Chv´ atal proved in [2] that the bounds from Theorems 1 and 2 are almost unimprovable. Consider now some results for the case where α ≥ 0 and γ = α. First upper α (α) was obtained by Kearns. bound on Cgreedy Theorem 3. (Kearns [5]) For any set cover problem with weights (A, S, w) and α any α, 0 ≤ α < 1, the inequality Cgreedy (α) ≤ Cmin (α)(2H(|A|) + 3) holds. This bound was improved by Slav´ık. Theorem 4. (Slav´ık [18]) For any set cover problem with weights (A, S, w) and α (α) ≤ Cmin (α)H ((1 − α)|A| ) holds. any α, 0 ≤ α < 1, the inequality Cgreedy Theorem 5. (Slav´ık [18])) For any set cover problem with weights (A, S, w) and α any α, 0 ≤ α < 1, the inequality Cgreedy (α) ≤ Cmin (α)H (maxBi ∈S |Bi |) holds. Slav´ık proved in [18] that the bounds from Theorems 4 and 5 are unimprovable. 2.3

On Polynomial Approximate Algorithms for Minimization of Partial Cover Weight

In this subsection we consider three theorems which follow immediately from Theorems 13–15 [9]. Let 0 ≤ α < 1. Consider the following problem: for given set cover problem with weights (A, S, w) it is required to ﬁnd an α-cover for (A, S) with minimal weight. Theorem 6. Let 0 ≤ α < 1. Then the problem of construction of α-cover with minimal weight is N P -hard. From this theorem it follows that we must consider polynomial approximate algorithms for minimization of α-cover weight. Theorem 7. Let α ∈ IR and 0 ≤ α < 1. If N P ⊆ DT IM E(nO(log log n) ) then for any ε, 0 < ε < 1, there is no polynomial algorithm that for a given set cover problem with weights (A, S, w) constructs an α-cover for (A, S) which weight is at most (1 − ε)Cmin (α, A, S, w) ln |A|. Theorem 8. Let α be a real number such that 0 ≤ α < 1. If P = N P then there exists δ > 0 such that there is no polynomial algorithm that for a given set cover problem with weights (A, S, w) constructs an α-cover for (A, S) which weight is at most δCmin (α, A, S, w) ln |A|.

216

M.Ju. Moshkov, M. Piliszczuk, and B. Zielosko

α From Theorem 4 it follows that Cgreedy (α) ≤ Cmin (α)(1 + ln |A|). From this inequality and from Theorem 7 it follows that under the assumption N P ⊆ DT IM E(nO(log log n) ) greedy algorithm with two thresholds α and α (in fact greedy algorithm with one threshold α from [18]) is close to best polynomial approximate algorithms for minimization of partial cover weight. From the considered inequality and from Theorem 8 it follows that under the assumption P = N P greedy algorithm with two thresholds α and α is not far from best polynomial approximate algorithms for minimization of partial cover weight. However we can try to improve the results of the work of greedy algorithm with two thresholds α and α for some part of set cover problems with weights.

2.4

Comparison of Greedy Algorithms with One and Two Thresholds

The following example shows that if for greedy algorithm with two thresholds α and γ we will use γ such that γ < α we can obtain sometimes better results than in the case γ = α. Example 1. Consider a set cover problem (A, S, w) such that A = {1, 2, 3, 4, 5, 6}, S = {B1 , B2 }, B1 = {1}, B2 = {2, 3, 4, 5, 6}, w(B1 ) = 1 and w(B2 ) = 4. Let α = 0.5. It means that we must cover at least M = (1 − α)|A| = 3 elements from A. If γ = α = 0.5 then the result of the work of greedy algorithm with thresholds α and γ is the 0.5-cover {B1 , B2 } which weight is equal to 5. If γ = 0 < α then the result of the work of greedy algorithm with thresholds α and γ is the 0.5-cover {B2 } which weight is equal to 4. In this subsection we show that under some assumptions on |A| and |S| for the most part of set cover problems (A, S) there exist a weight function w γ (α, A, S, w) < and real numbers α, γ such that 0 ≤ γ < α < 1 and Cgreedy α Cgreedy (α, A, S, w). First, we consider criterion of existence of such w, α and γ (see Theorem 9). First part of the proof of this criterion is based on a construction similar to considered in Example 1. Let A be a ﬁnite nonempty set and S = {B1 , . . . , Bm } be a family of subsets of A. We will say that the family S is 1-uniform if there exists a natural number k such that |Bi | = k or |Bi | = k + 1 for any nonempty subset Bi from S. We will say that S is strongly 1-uniform if S is 1-uniform and for any subsets Bl1 , . . . , Blt from S the family {B1 \ U, . . . , Bm \ U } is 1-uniform where U = Bl1 ∪ . . . ∪ Blt . Theorem 9. Let (A, S) be a set cover problem. Then the following two statements are equivalent: 1. The family S is not strongly 1-uniform. 2. There exist a weight function w and real numbers α and γ such that 0 ≤ γ < γ α (α, A, S, w) < Cgreedy (α, A, S, w). α < 1 and Cgreedy Proof. Let S = {B1 , . . . , Bm }. Let the family S be not strongly 1-uniform. Let us choose minimal number of subsets Bl1 , . . . , Blt from the family S (it is possible

On Partial Covers, Reducts and Decision Rules with Weights

217

that t = 0) such that the family {B1 \ U, . . . , Bm \ U } is not 1-uniform where U = Bl1 ∪ . . . ∪ Blt (if t = 0 then U = ∅). Since {B1 \ U, . . . , Bm \ U } is not 1-uniform, there exist two subsets Bi and Bj from S such that |Bi \ U | > 0 and |Bj \ U | ≥ |Bi \ U | + 2. Let us choose real α and γ such that M = |A|(1 − α) = |U |+|Bi \U |+1 and N = |A|(1 − γ) = |U |+|Bi \U |+2. It is clear that 0 ≤ γ < α < 1. Let us deﬁne a weight function w as follows: w(Bl1 ) = . . . = w(Blt ) = 1, w(Bi ) = |A| · 2|Bi \ U |, w(Bj ) = |A|(2|Bi \ U | + 3) and w(Br ) = |A|(3|Bi \ U | + 6) for any Br from S such that r ∈ / {i, j, l1 , . . . , lt }. Let us consider the work of greedy algorithm with two thresholds α and α. One can show that during ﬁrst t steps the greedy algorithm will choose subsets Bl1 , . . . , Blt (may be in an another order). It is clear that |U | < M . Therefore the greedy algorithm must make the step number t + 1. During this step the greedy algorithm will choose a subset Bk from S with minimal number k for w(Bk ) k) which Bk \ U = ∅ and the value p(k) = min{|Bkw(B \U|,M−|U|} = min{|Bk \U|,|Bi \U|+1} is minimal. 1 It is clear that p(i) = 2|A|, p(j) = (2 + |Bi \U|+1 )|A| and p(k) > 3|A| for any subset Bk from S such that Bk \ U = ∅ and k ∈ / {i, j, l1 , . . . , lt }. Therefore during the step number t + 1 the greedy algorithm will choose the subset Bi . Since |U | + |Bi \ U | = M − 1, the greedy algorithm will make the step number t + 2 and will choose a subset from S which is diﬀerent from Bl1 , . . . , Blt , Bi . As α (α, A, S, w) ≥ t + |A| · 2|Bi \ U | + |A|(2|Bi \ U | + 3). the result we obtain Cgreedy Let us consider the work of greedy algorithm with two thresholds α and γ. One can show that during ﬁrst t steps the greedy algorithm will choose subsets Bl1 , . . . , Blt (may be in an another order). It is clear that |U | < M . Therefore the greedy algorithm must make the step number t + 1. During this step the greedy algorithm will choose a subset Bk from S with minimal number k for w(Bk ) k) which Bk \ U = ∅ and the value q(k) = min{|Bw(B = min{|Bk \U|,|B i \U|+2} k \U|,N −|U|} is minimal. 1 )|A| and q(k) ≥ 3|A| for It is clear that q(i) = 2|A|, q(j) = (2 − |Bi \U|+2 any subset Bk from S such that Bk \ U = ∅ and k ∈ / {i, j, l1 , . . . , lt }. Therefore during the step number t + 1 the greedy algorithm will choose the subset Bj . Since |U | + |Bj \ U | > M , the α-cover constructed by greedy algorithm will γ be equal to {Bl1 , . . . , Blt , Bj }. As the result we obtain Cgreedy (α, A, S, w) = α t+|A|(2|Bi \U |+3). Since Cgreedy (α, A, S, w) ≥ t+|A|·2|Bi \U |+|A|(2|Bi \U |+3) γ α (α, A, S, w) > Cgreedy (α, A, S, w). and |Bi \ U | > 0, we conclude that Cgreedy Let the family S be strongly 1-uniform. Consider arbitrary weight function w for S and real numbers α and γ such that 0 ≤ γ < α < 1. Let us show that γ α (α, A, S, w) ≥ Cgreedy (α, A, S, w). Let us denote M = |A|(1 − α) and Cgreedy γ α N = |A|(1 − γ) . If M = N then Cgreedy (α, A, S, w) = Cgreedy (α, A, S, w). Let N > M. Let us apply the greedy algorithm with thresholds α and α to the set cover problem with weights (A, S, w). Let during the construction of α-cover this algorithm choose sequentially subsets Bg1 , . . . , Bgt . Let us apply now the greedy algorithm with thresholds α and γ to the set cover problem with weights (A, S, w). If during the construction of α-cover this algorithm chooses sequentially subsets

218

M.Ju. Moshkov, M. Piliszczuk, and B. Zielosko

γ α Bg1 , . . . , Bgt then Cgreedy (α, A, S, w) = Cgreedy (α, A, S, w). Let there exist a nonnegative integer r, 0 ≤ r ≤ t − 1, such that during ﬁrst r steps the considered algorithm chooses subsets Bg1 , . . . , Bgr , but at the step number r + 1 the algorithm chooses a subset Bk such that k = gr+1 . Let us denote Bg0 = ∅, D = Bg0 ∪. . .∪Bgr and J = {i : i ∈ {1, . . . , m}, Bi \D = ∅}. It is clear that gr+1 , k ∈ J. w(Bi ) i) For any i ∈ J denote p(i) = min{|Biw(B \D|,M−|D|} , q(i) = min{|Bi \D|,N −|D|} . Since k = gr+1 , we conclude that there exists i ∈ J such that p(i) = q(i). Therefore |Bi \ D| > M − |D|. Since S is strongly 1-uniform family, we have |Bj \ D| ≥ M − |D| for any j ∈ J. From here it follows, in particular, that r + 1 = t, and {Bg1 , . . . , Bgt−1 , Bk } is an α-cover for (A, S). It is clear that p(gt ) ≤ p(k). Since |Bk \D| ≥ M −|D| and |Bgt \D| ≥ M −|D|, w(Bgt ) w(Bk ) we have p(k) = M−|D| , p(gt ) = M−|D| . Therefore w(Bgt ) ≤ w(Bk ). γ Taking into account that Cgreedy (α, A, S, w) = w(Bg1 ) + . . . + w(Bgt−1 ) + α w(Bk ) and Cgreedy (α, A, S, w) = w(Bg1 ) + . . . + w(Bgt−1 ) + w(Bgt ) we obtain γ α (α, A, S, w).

Cgreedy (α, A, S, w) ≥ Cgreedy

Let us show that under some assumptions on |A| and |S| the most part of set cover problems (A, S) is not 1-uniform, and therefore is not strongly 1-uniform. There is one-to-one correspondence between set cover problems and tables ﬁlled by numbers from {0, 1} and having no rows ﬁlled by 0 only. Let A = {a1 , . . . , an } and S = {B1 , . . . , Bm }. Then the problem (A, S) corresponds to the table with n rows and m columns which for i = 1, . . . , n and j = 1, . . . , m has 1 at the intersection of i-th row and j-th column if and only if ai ∈ Bj . A table ﬁlled by numbers from {0, 1} will be called SC-table if this table has no rows ﬁlled by 0 only. For completeness of the presentation we consider here a statement from [9] with proof. Lemma 1. The number of SC-tables with n rows and m columns is at least 2mn − 2mn−m+log2 n . Proof. Let i ∈ {1, . . . , n}. The number of tables in which the i-th row is ﬁlled by 0 only is equal to 2mn−m . Therefore the number of tables which are not SCtables is at most n2mn−m = 2mn−m+log2 n . Thus, the number of SC-tables is at least 2mn − 2mn−m+log2 n .

n/2

Lemma 2. Let n ∈ IN, n ≥ 4 and k ∈ {0, . . . , n}. Then Cnk ≤ Cn

q(li+1 ). Using inequality p(gi+1 ) ≥ q(gi+1 ) we obtain p(gi+1 ) > q(li+1 ) and p(li+1 ) > q(li+1 ). From last inequality it follows that |Bli+1 \ D| > M − |D|. Let gi+1 > li+1 . In this case we have p(gi+1 ) < p(li+1 ) and q(gi+1 ) ≥ q(li+1 ). Using inequality p(gi+1 ) ≥ q(gi+1 ) we obtain p(gi+1 ) ≥ q(li+1 ) and p(li+1 ) > q(li+1 ). From last inequality it follows that |Bli+1 \ D| > M − |D|. So in any case we have |Bli+1 \ D| > M − |D|. From this inequality it follows that after the step number i+1 the algorithm with thresholds α and γ must ﬁnish

the work. Thus, k = i + 1, k ≤ t, (l1 , . . . , lk−1 ) = (g1 , . . . , gk−1 ) and lk = gk .

2.6

Lower Bound on Cmin (α)

In this subsection we ﬁx some information about the work of greedy algorithm with two thresholds and ﬁnd the best lower bound on the value Cmin (α) depending on this information. Let (A, S, w) be a set cover problem with weights and α, γ be real numbers such that 0 ≤ γ ≤ α < 1. Let us apply the greedy algorithm with thresholds α and γ to the set cover problem with weights (A, S, w). Let during the construction of α-cover the greedy algorithm choose sequentially subsets Bg1 , . . . , Bgt . Let us denote Bg0 = ∅ and δ0 = 0. For i = 1, . . . , t denote δi = |Bgi \ (Bg0 ∪ . . . ∪ Bgi−1 )| and wi = w(Bgi ).

On Partial Covers, Reducts and Decision Rules with Weights

221

As information on the greedy algorithm work we will use numbers MC = MC (α, γ, A, S, w) = |A|(1 − α) and NC = NC (α, γ, A, S, w) = |A|(1 − γ) , and tuples ΔC = ΔC (α, γ, A, S, w) = (δ1 , . . . , δt ) and WC = WC (α, γ, A, S, w) = (w1 , . . . , wt ). For i = 0, . . . , t − 1 denote wi+1 (MC − (δ0 + . . . + δi )) ρi = . min{δi+1 , NC − (δ0 + . . . + δi )} Let us deﬁne parameter ρC (α, γ) = ρC (α, γ, A, S, w) as follows: ρC (α, γ) = max {ρi : i = 0, . . . , t − 1} . We will prove that ρC (α, γ) is the best lower bound on Cmin (α) depending on MC , NC , ΔC and WC . This lower bound is based on a generalization of the following simple reasoning: if we must cover M elementsand the maximal subsets. cardinality of a subset from S is δ then we must use at least M δ Theorem 11. For any set cover problem with weights (A, S, w) and any real numbers α, γ, 0 ≤ γ ≤ α < 1, the inequality Cmin (α, A, S, w) ≥ ρC (α, γ, A, S, w) holds, and there exists a set cover problem with weights (A , S , w ) such that MC (α, γ, A , S , w ) = MC (α, γ, A, S, w), NC (α, γ, A , S , w ) = NC (α, γ, A, S, w) ΔC (α, γ, A , S , w ) = ΔC (α, γ, A, S, w), WC (α, γ, A , S , w ) = WC (α, γ, A, S, w) ρC (α, γ, A , S , w ) = ρC (α, γ, A, S, w), Cmin (α, A , S , w ) = ρC (α, γ, A , S , w ) . Proof. Let (A, S, w) be a set cover problem with weights, S = {B1 , . . . , Bm }, and α, γ be real numbers such that 0 ≤ γ ≤ α < 1. Let us denote M = MC (α, γ, A, S, w) = |A|(1 − α) and N = NC (α, γ, A, S, w) = |A|(1 − γ) . Let {Bl1 , . . . , Blk } be an optimal α-cover for (A, S, w), i.e. w(Bl1 ) + . . . + w(Blk ) = Cmin (α, A, S, w) = Cmin (α) and |Bl1 ∪ . . . ∪ Blk | ≥ M . Let us apply the greedy algorithm with thresholds α and γ to (A, S, w). Let during the construction of α-cover the greedy algorithm choose sequentially subsets Bg1 , . . . , Bgt . Let us denote Bg0 = ∅. Let i ∈ {0, . . . , t − 1}. Let us denote D = Bg0 ∪ . . . ∪ Bgi . It is clear that after i steps of greedy algorithm work in the set Bl1 ∪ . . . ∪ Blk at least |Bl1 ∪ . . .∪ Blk | − |Bg0 ∪ . . . ∪ Bgi | ≥ M − |D| > 0 elements remained uncovered. After i-th step p1 = |Bl1 \ D| elements remained uncovered in the set Bl1 , ..., and pk = |Blk \ D| elements remained uncovered in the set Blk . We know that p1 + . . . + pk ≥ M − |D| > 0. Let, for the deﬁniteness, p1 > 0, . . . , pr > 0, pr+1 = . . . = pk = 0. For j = 1, . . . , r denote qj = min{pj , N −|D|}. It is clear that N −|D| ≥ M −|D|. w(B ) w(B ) Therefore q1 + . . . + qr ≥ M − |D|. Let us consider numbers q1l1 , . . . , qrlr . w(Bl1 )+...+w(Blr ) . q1 +...+qr w(Blr )qr +...+ > qr

Let us show that at least one of these numbers is at most β = w(B )q

1 l1 Assume the contrary. Then w(Bl1 ) + . . . + w(Blr ) = q1 (q1 + . . . + qr )β = w(Bl1 ) + . . . + w(Blr ) which is impossible.

222

M.Ju. Moshkov, M. Piliszczuk, and B. Zielosko

We know that q1 + . . . + qr ≥ M − |D| and w(Bl1 ) + . . . + w(Blr ) ≤ Cmin (α). min (α) , and there exists j ∈ {1, . . . , k} such that Blj \ D = ∅ Therefore β ≤ CM−|D| w(Blj ) w(Bgi+1 ) min{|Blj \D|,N −|D|} ≤ β. Hence min{|Bgi+1 \D|,N −|D|} w(Bg )(M−|D|) Cmin (α) ≥ min{|Bgi+1 \D|,N −|D|} . i+1

and

≤ β ≤

Cmin (α) M−|D|

and

into account

that Cmin (α) is a natural number we obtain Cmin (α) ≥ Taking w(Bgi+1 )(M−|D|) min{|Bg \D|,N −|D|} = ρi . Since last inequality holds for any i ∈ {0, . . . , t − 1} i+1

and ρC (α, γ) = ρC (α, γ, A, S, w) = max {ρi : i = 0, . . . , t − 1}, we conclude that Cmin (α) ≥ ρC (α, γ). Let us show that this bound is unimprovable depending on MC , NC , ΔC and WC . Let us consider a set cover problem with weights (A , S , w ) where A = A, S = {B1 , . . . , Bm , Bm+1 }, |Bm+1 | = M , Bg1 ∪ . . . ∪ Bgt−1 ⊆ Bm+1 ⊆ Bg1 ∪ . . . ∪ Bgt , w (B1 ) = w(B1 ), . . . , w (Bm ) = w(Bm ) and w (Bm+1 ) = ρC (α, γ). It is clear that MC (α, γ, A , S , w ) = MC (α, γ, A, S, w) = M and NC (α, γ, A , S , w ) = NC (α, γ, A, S, w) = N . We show ΔC (α, γ, A , S , w ) = ΔC (α, γ, A, S, w) and WC (α, γ, A , S , w ) = WC (α, γ, A, S, w). Let us show by induction on i ∈ {1, . . . , t} that for the set cover problem with weights (A , S , w ) at the step number i the greedy algorithm with two thresholds α and γ will choose the subset Bgi . Let us consider the ﬁrst step. (Bm+1 ) ρC (α,γ) Let us denote D = ∅. It is clear that min{|Bwm+1 \D|,N −|D|} = M−|D| . From the deﬁnition of ρC (α, γ) it follows that

w (Bg1 ) min{|Bg1 \D|,N −|D|}

=

w(Bg1 ) min{|Bg1 \D|,N −|D|}

≤

ρC (α,γ) M−|D| .

Using this fact and the inequality g1 < m + 1 it is not diﬃcult to prove that at the ﬁrst step greedy algorithm will choose the subset Bg1 . Let i ∈ {1, . . . , t − 1}. Let us assume that the greedy algorithm made i steps for (A , S , w ) and chose subsets Bg1 , . . . , Bgi . Let us show that at the step i + 1 the subset Bgi+1 will be chosen. Let us denote D = Bg1 ∪ . . . ∪ Bgi . Since Bg1 ∪ . . . ∪ Bgi ⊆ Bm+1 and |Bm+1 | = M , we have |Bm+1 \ D| = M − |D|. (Bm+1 ) ρC (α,γ) Therefore min{|Bwm+1 \D|,N −|D|} = M−|D| . From the deﬁnition of the parameter w (Bg

)

w(Bg

)

ρC (α,γ) i+1 i+1 ρC (α, γ) it follows that min{|Bg \D|,N −|D|} = min{|Bgi+1 \D|,N −|D|} ≤ M−|D| . i+1 Using this fact and the inequality gi+1 < m + 1 it is not diﬃcult to prove that at the step number i + 1 greedy algorithm will choose the subset Bgi+1 . Thus, ΔC (α, γ, A , S , w ) = ΔC (α, γ, A, S, w) and WC (α, γ, A , S , w ) = WC (α, γ, A, S, w). Therefore ρC (α, γ, A , S , w ) = ρC (α, γ, A, S, w) = ρC (α, γ). From been proven it follows that Cmin (α, A , S , w ) ≥ ρC (α, γ, A , S , w ). It is clear that {Bm+1 } is an α-cover for (A , S ) and the weight of {Bm+1 } is equal to ρC (α, γ, A , S , w ). Hence Cmin (α, A , S , w ) = ρC (α, γ, A , S , w ).

Let us consider a property of the parameter ρC (α, γ) which is important for practical use of the bound from Theorem 11. Proposition 2. Let (A, S, w) be a set cover problem with weights and α, γ be real numbers such that 0 ≤ γ ≤ α < 1. Then ρC (α, α, A, S, w) ≥ ρC (α, γ, A, S, w).

On Partial Covers, Reducts and Decision Rules with Weights

223

Proof. Let S = {B1 , . . . , Bm }, M = |A|(1 − α) , N = |A|(1 − γ) , ρC (α, α) = ρC (α, α, A, S, w) and ρC (α, γ) = ρC (α, γ, A, S, w). Let us apply the greedy algorithm with thresholds α and α to (A, S, w). Let during the construction of α-cover this algorithm choose sequentially subsets Bg1 , . . . , Bgt . Let us denote B g0 = ∅. For j = 0, . . . , t − 1 denote Dj = Bg0 ∪ . . . ∪ Bgj and ρC (α, α, j) =

w(Bgj+1 )(M−|Dj |) min{|Bgj+1 \Dj |,M−|Dj |}

. Then ρC (α, α) =

max{ρC (α, α, j) : j = 0, . . . , t − 1}. Apply the greedy algorithm with thresholds α and γ to (A, S, w). Let during the construction of α-cover this algorithm choose sequentially subsets Bl1 , . . . , Blk . From Proposition 1 it follows that either k = t and (l1 , . . . , lk ) = (g1 , . . . , gt ) or k ≤ t, (l1 , . . . , lk−1 ) = (g1 , . . . , gk−1 ) and lk = gk . Let us consider these two cases separately. Let k = t and (l1 , . . . , lk ) = (g1 , . . . , gt ). For j = 0, . . . , t − 1 denote ρC (α, γ, j) =

w(Bgj+1 )(M−|Dj |) min{|Bgj+1 \Dj |,N −|Dj |}

j = 0, . . . , t − 1}. Since N ≥ M , we 0, . . . , t − 1. Hence ρC (α, γ) ≤ ρC (α, α). and lk = gk . Let us denote ρC (α, γ, k

w(B )(M−|Dj |) g ρC (α, γ, j) = min{|Bg j+1\Dj |,N −|Dj |} j+1

. Then ρC (α, γ) = max{ρC (α, γ, j) :

have ρC (α, γ, j) ≤ ρC (α, α, j) for j = Let k ≤ t, (l1 , . . . , lk−1 ) = (g1 , . . . , gk−1 ) w(B k )(M−|Dk−1 |) and − 1) = min{|Bl l\D k−1 |,N −|Dk−1 |} k

for j = 0, . . . , k − 2. Then ρC (α, γ) =

max{ρC (α, γ, j) : j = 0, . . . , k − 1}. Since N ≥ M , we have ρC (α, γ, j) ≤ w(B ) ρC (α, α, j) for j = 0, . . . , k − 2. It is clear that min{|Bl \Dk−1lk|,N −|Dk−1 |} ≤ k

w(Bgk ) min{|Bgk \Dk−1 |,N −|Dk−1 |}

w(B

)

k ≤ min{|Bg \Dk−1g|,M−|D . Thus, ρC (α, γ, k − 1) ≤ k−1 |} k

ρC (α, α, k − 1) and ρC (α, γ) ≤ ρC (α, α).

2.7

γ (α) Upper Bounds on Cgreedy

In this subsection we study some properties of parameter ρC (α, γ) and obtain γ two upper bounds on the value Cgreedy (α) which do not depend directly on cardinality of the set A and cardinalities of subsets Bi from S. Theorem 12. Let (A, S, w) be a set cover problem with weights and α, γ be real numbers such that 0 ≤ γ < α < 1. Then 1−γ γ Cgreedy (α, A, S, w) < ρC (γ, γ, A, S, w) ln +1 . α−γ Proof. Let S = {B1 , . . . , Bm }. Let us denote M = |A|(1 − α) and N = |A|(1 − γ) . Let us apply the greedy algorithm with thresholds γ and γ to (A, S, w). Let during the construction of γ-cover the greedy algorithm choose sequentially subsets Bg1 , . . . , Bgt . Let us denote Bg0 = ∅, for i = 0, . . . , t − 1 denote Di = Bg0 ∪ . . . ∪ Bgi , and denote ρ = ρC (γ, γ, A, S, w). Immediately from the deﬁnition of the parameter ρ it follows that for i = 0, . . . , t − 1 w(Bgi+1 ) ρ ≤ . min{|Bgi+1 \ Di |, N − |Di |} N − |Di |

(1)

224

M.Ju. Moshkov, M. Piliszczuk, and B. Zielosko

Note that min{|Bgi+1 \ Di |, N − |Di |} = |Bgi+1 \ Di | for i = 0, . . . , t − 2 since {Bg0 , . . . , Bgi+1 } is not a γ-cover for (A, S). Therefore for i = 0, . . . , t − 2 we have

w(Bgi+1 ) |Bgi+1 \Di |

≤

ρ N −|Di |

and

N −|Di | ρ

≤

|Bgi+1 \Di | w(Bgi+1 ) .

Thus, for i = 1, . . . , t − 1

i−1 | during the step number i the greedy algorithm covers at least N −|D elements ρ on each unit of weight. From (1) it follows that that for i = 0, . . . , t − 1

w(Bgi+1 ) ≤

ρ min{|Bgi+1 \ Di |, N − |Di |} ≤ρ . N − |Di |

(2)

Assume that ρ = 1. Using (2) we obtain w(Bg1 ) = 1. From this equality and (1) it follows that |Bg1 | ≥ N . Therefore {Bg1 } is an α-cover for (A, S), and

γ 1−γ + 1 > 1. Therefore the statement of the (α) = 1. It is clear that ln α−γ Cgreedy theorem holds if ρ = 1. Assume now that ρ ≥ 2. Let |Bg1 | ≥ M . Then {B g1 } is an α-cover for (A, S). γ 1−γ Using (2) we obtain Cgreedy (α) ≤ ρ. Since ln α−γ +1 > 1, we conclude that the statement of the theorem holds if |Bg1 | ≥ M . Let |Bg1 | < M . Then there exists q ∈ {1, . . . , t − 1} such that |Bg1 ∪ . . . ∪ Bgq | < M and |Bg1 ∪ . . . ∪ Bgq+1 | ≥ M . Taking into account that for i = 1, . . . , q during the step number i the greedy i−1 | algorithm covers at least N −|D elements on each unit of weight we obtain ρ w(Bg1 )+...+w(Bgq ) . Let us denote k = w(Bg1 ) + N − |Bg1 ∪ . . . ∪ Bgq | ≤ N 1 − 1ρ k . . . + w(Bgq ). Then N − N 1 − 1ρ ≤ |Bg1 ∪ . . . ∪ Bgq | ≤ M − 1. Therefore k k |A|(1 − γ) − |A|(1 − γ) 1 − ρ1 < |A|(1 − α), 1 − γ − 1 + α < (1 − γ) ρ−1 , ρ k k ρ 1−γ 1−γ 1 . To obtain last inequality < α−γ , 1 + ρ−1 < 1−γ , and kρ < ln α−γ ρ−1

α−γ 1 1 we use known inequality ln 1 + r > r+1 which holds for any natural r. It is γ clear that Cgreedy (α) = k + w(Bq+1 ). Using (2) we conclude that w(Bq+1 ) ≤ ρ. γ 1−γ + ρ.

Therefore Cgreedy (α) < ρ ln α−γ

Corollary 1. Let ε be a real number, and 0 < ε < 1. Then for any α such that ε ≤ α < 1 the following inequalities hold: 1 α−ε ρC (α, α) ≤ Cmin (α) ≤ Cgreedy (α) < ρC (α − ε, α − ε) ln + 1 . ε For example, if ε = 0.01 and 0.01 ≤ α < 1 then ρC (α, α) ≤ Cmin (α) ≤ α−0.01 Cgreedy (α) < 5.61ρC (α − 0.01, α − 0.01), and if ε = 0.1 and 0.1 ≤ α < 1 α−0.1 then ρC (α, α) ≤ Cmin (α) ≤ Cgreedy (α) < 3.31ρC (α − 0.1, α − 0.1). The obtained results show that the lower bound Cmin (α) ≥ ρC (α, α) is nontrivial. Theorem 13. Let (A, S, w) be a set cover problem with weights and α, γ be real numbers such that 0 ≤ γ < α < 1. Then 1−γ γ Cgreedy (α, A, S, w) < Cmin (γ, A, S, w) ln +1 . α−γ

On Partial Covers, Reducts and Decision Rules with Weights

225

γ Proof. From Theorem 12 it follows that Cgreedy (α, A, S, w) < ρC (γ, γ, A, S, w) · 1−γ ln α−γ + 1 . The inequality ρC (γ, γ, A, S, w) ≤ Cmin (γ, A, S, w) follows from Theorem 11.

0 0.001 Corollary 2. Cgreedy (0.001) < 7.91Cmin(0), Cgreedy (0.01) < 5.71Cmin(0.001), 0.1 0.3 Cgreedy (0.2) < 3.20Cmin(0.1), Cgreedy (0.5) < 2.26Cmin (0.3).

0 Corollary 3. Let 0 < α < 1. Then Cgreedy (α) < Cmin (0) ln α1 + 1 . Corollary 4. Let ε be a real number, and 0 < ε < 1. Then for any α such that α−ε ε ≤ α < 1 the inequalities Cmin (α) ≤ Cgreedy (α) < Cmin (α − ε) ln 1ε + 1 hold.

3 3.1

Partial Tests and Reducts Main Notions

Let T be a table with n rows labeled by nonnegative integers (decisions) and m columns labeled by attributes (names of attributes) f1 , . . . , fm . This table is ﬁlled by nonnegative integers (values of attributes). The table T is called a decision table. Let w be a weight function for T which corresponds to each attribute fi a natural number w(fi ). Let us denote by P (T ) the set of unordered pairs of diﬀerent rows of T with diﬀerent decisions. We will say that an attribute fi separates a pair of rows (r1 , r2 ) ∈ P (T ) if rows r1 and r2 have diﬀerent numbers at the intersection with the column fi . For i = 1, . . . , m denote by P (T, fj ) the set of pairs from P (T ) which the attribute fi separates. Let α be a real number such that 0 ≤ α < 1. A set of attributes Q ⊆ {f1 , . . . , fm } will be called an α-test for T if attributes from Q separate at least (1 − α)|P (T )| pairs from the set P (T ). An α-test is called an α-reduct if each proper subset of the considered α-test is not α-test. If P (T ) = ∅ then each subset of {f1 , . . . , fm } is an α-test, and only empty set is an α-reduct. For example, 0.01-test means that we must separate at least 99% of pairs from P (T ). Note that 0-reduct is usual (exact) reduct. It must be noted also that each α-test contains at least one α-reduct as a subset. The number w(Q) = fi ∈Q w(fi ) will be called the weight of the set Q. If Q = ∅ then w(Q) = 0. Let us denote by Rmin (α) = Rmin (α, T, w) the minimal weight of α-reduct for T . It is clear that Rmin (α, T, w) coincides with the minimal weight of α-test for T . Let α, γ be real numbers such that 0 ≤ γ ≤ α < 1. Let us describe a greedy algorithm with thresholds α and γ which constructs an α-test for given decision table T and weight function w. If P (T ) = ∅ then the constructed α-test is empty set. Let P (T ) = ∅. Let us denote M = |P (T )|(1 − α) and N = |P (T )|(1 − γ) . Let we make i ≥ 0

226

M.Ju. Moshkov, M. Piliszczuk, and B. Zielosko

steps and construct a set Q containing i attributes (if i = 0 then Q = ∅). Let us describe the step number i + 1. Let us denote by D the set of pairs from P (T ) separated by attributes from Q (if i = 0 then D = ∅). If |D| ≥ M then we ﬁnish the work of the algorithm. The set of attributes Q is the constructed α-test. Let |D| < M . Then we choose an attribute fj with minimal number j for which P (T, fj ) \ D = ∅ and the value w(fj ) min{|P (T, fj ) \ D|, N − |D|} is minimal. Add the attribute fj to the set Q. Pass to the step number i + 2. γ γ Let us denote by Rgreedy (α) = Rgreedy (α, T, w) the weight of α-test constructed by greedy algorithm with thresholds α and γ for given decision table T and weight function w. 3.2

Relationships Between Partial Covers and Partial Tests

Let (A, S, w) be a set cover problem with weights and α, γ be real numbers such that 0 ≤ γ ≤ α < 1. Let us apply the greedy algorithm with thresholds α and γ to (A, S, w). Let during the construction of α-cover the greedy algorithm choose sequentially subsets Bj1 , . . . , Bjt from the family S. Let us denote OC (α, γ, A, S, w) = (j1 , . . . , jt ). Let T be a decision table with m columns labeled by attributes f1 , . . . , fm , and with a nonempty set P (T ). Let w be a weight function for T . We correspond a set cover problem with weights (A(T ), S(T ), uw ) to the considered decision table T and weight function w in the following way: A(T ) = P (T ), S(T ) = {B1 (T ), . . . , Bm (T )} where B1 (T ) = P (T, f1 ), . . . , Bm (T ) = P (T, fm ), uw (B1 (T )) = w(f1 ), . . . , uw (Bm (T )) = w(fm ). Let α, γ be real numbers such that 0 ≤ γ ≤ α < 1. Let us apply the greedy algorithm with thresholds α and γ to decision table T and weight function w. Let during the construction of α-test the greedy algorithm choose sequentially attributes fj1 , . . . , fjt . Let us denote OR (α, γ, T, w) = (j1 , . . . , jt ). Let us denote P (T, fj0 ) = ∅. For i = 1, . . . , t denote wi = w(fji ) and δi = |P (T, fji ) \ (P (T, fj0 ) ∪ . . . ∪ P (T, fji−1 ))| . Let us denote MR (α, γ, T, w) = |P (T )|(1 − α) , NR (α, γ, T, w) = |P (T )|(1 − γ) , ΔR (α, γ, T, w) = (δ1 , . . . , δt ) and WR (α, γ, T, w) = (w1 , . . . , wt ). It is not diﬃcult to prove the following statement. Proposition 3. Let T be a decision table with m columns labeled by attributes f1 , . . . , fm , P (T ) = ∅, w be a weight function for T , and α, γ be real numbers such that 0 ≤ γ ≤ α < 1. Then |P (T )| = |A(T )| , |P (T, fi )| = |Bi (T )|, i = 1, . . . , m , OR (α, γ, T, w) = OC (α, γ, A(T ), S(T ), uw ) ,

On Partial Covers, Reducts and Decision Rules with Weights

227

MR (α, γ, T, w) = MC (α, γ, A(T ), S(T ), uw ) , NR (α, γ, T, w) = NC (α, γ, A(T ), S(T ), uw ) , ΔR (α, γ, T, w) = ΔC (α, γ, A(T ), S(T ), uw ) , WR (α, γ, T, w) = WC (α, γ, A(T ), S(T ), uw ) , Rmin (α, T, w) = Cmin (α, A(T ), S(T ), uw ) , γ γ (α, T, w) = Cgreedy (α, A(T ), S(T ), uw ) . Rgreedy Let (A, S, w) be a set cover problem with weights where A = {a1 , . . . , an } and S = {B1 , . . . , Bm }. We correspond a decision table T (A, S) and a weight function vw for T (A, S) to the set cover problem with weights (A, S, w) in the following way. The table T (A, S) contains m columns labeled by attributes f1 , . . . , fm and n+ 1 rows ﬁlled by numbers from {0, 1}. For i = 1, . . . , n and j = 1, . . . , m at the intersection of i-th row and j-th column the number 1 stays if and only if ai ∈ Bj . The row number n + 1 is ﬁlled by 0. First n rows are labeled by the decision 0. Last row is labeled by the decision 1. Let vw (f1 ) = w(B1 ), . . . , vw (fm ) = w(Bm ). For i = {1, . . . , n + 1} denote by ri the i-th row. It is not diﬃcult to see that P (T (A, S)) = {(r1 , rn+1 ), . . . , (rn , rn+1 )}. Let i ∈ {1, . . . , n} and j ∈ {1, . . . , m}. One can show that the attribute fj separates the pair (ri , rn+1 ) if and only if ai ∈ Bj . It is not diﬃcult to prove the following statement. Proposition 4. Let (A, S, w) be a set cover problem with weights and α, γ be real numbers such that 0 ≤ γ ≤ α < 1. Then |P (T (A, S))| = |A| , OR (α, γ, T (A, S), vw ) = OC (α, γ, A, S, w) , MR (α, γ, T (A, S), vw ) = MC (α, γ, A, S, w) , NR (α, γ, T (A, S), vw ) = NC (α, γ, A, S, w) , ΔR (α, γ, T (A, S), vw ) = ΔC (α, γ, A, S, w) , WR (α, γ, T (A, S), vw ) = WC (α, γ, A, S, w) , Rmin (α, T (A, S), vw ) = Cmin (α, A, S, w) , γ γ Rgreedy (α, T (A, S), vw ) = Cgreedy (α, A, S, w) .

3.3

On Precision of Greedy Algorithm with Thresholds α and α

The following two statements are simple corollaries of results of Slav´ık (see Theorems 4 and 5) and Proposition 3. Theorem 14. Let T be a decision table, P (T ) = ∅, w be a weight function for α T , α ∈ IR and 0 ≤ α < 1. Then Rgreedy (α) ≤ Rmin (α)H ((1 − α)|P (T )| ). Theorem 15. Let T be a decision table with m columns labeled by attributes f1 , . . . , fm , P (T ) = ∅, w be a weight function for T , and α be a real number α such that 0 ≤ α < 1. Then Rgreedy (α) ≤ Rmin (α)H maxi∈{1,...,m} |P (T, fi )| .

228

M.Ju. Moshkov, M. Piliszczuk, and B. Zielosko

3.4

On Polynomial Approximate Algorithms

In this subsection we consider three theorems which follows immediately from Theorems 26–28 [9]. Let 0 ≤ α < 1. Let us consider the following problem: for given decision table T and weight function w for T it is required to ﬁnd an α-test (α-reduct) for T with minimal weight. Theorem 16. Let 0 ≤ α < 1. Then the problem of construction of α-test (αreduct) with minimal weight is N P -hard. So we must consider polynomial approximate algorithms for minimization of α-test (α-reduct) weight. Theorem 17. Let α ∈ IR and 0 ≤ α < 1. If N P ⊆ DT IM E(nO(log log n) ) then for any ε, 0 < ε < 1, there is no polynomial algorithm that for given decision table T with P (T ) = ∅ and weight function w for T constructs an α-test for T which weight is at most (1 − ε)Rmin (α, T, w) ln |P (T )|. Theorem 18. Let α be a real number such that 0 ≤ α < 1. If P = N P then there exists δ > 0 such that there is no polynomial algorithm that for given decision table T with P (T ) = ∅ and weight function w for T constructs an αtest for T which weight is at most δRmin (α, T, w) ln |P (T )|. α (α) ≤ Rmin (α)(1 + ln |P (T )|). From this From Theorem 14 it follows that Rgreedy inequality and from Theorem 17 it follows that under the assumption N P ⊆ DT IM E(nO(log log n) ) greedy algorithm with two thresholds α and α is close to best polynomial approximate algorithms for minimization of partial test weight. From the considered inequality and from Theorem 18 it follows that under the assumption P = N P greedy algorithm with two thresholds α and α is not far from best polynomial approximate algorithms for minimization of partial test weight. However we can try to improve the results of the work of greedy algorithm with two thresholds α and α for some part of decision tables.

3.5

Two Modiﬁcations of Greedy Algorithm

First, we consider binary diagnostic decision tables and prove that under some assumptions on the number of attributes and rows for the most part of tables there exist weight function w and numbers α, γ such that the weight of α-test constructed by greedy algorithm with thresholds α and γ is less than the weight of α-test constructed by greedy algorithm with thresholds α and α. Binary means that the table is ﬁlled by numbers from the set {0, 1} (all attributes have values from {0, 1}). Diagnostic means that rows of the table are labeled by pairwise diﬀerent numbers (decisions). Let T be a binary diagnostic decision table with m columns labeled by attributes f1 , . . . , fm and with n rows. We will assume that rows of T with numbers 1, . . . , n are labeled by decisions 1, . . . , n respectively. Therefore the number of considered tables is equal to 2mn . Decision table will be called simple if it has no equal rows.

On Partial Covers, Reducts and Decision Rules with Weights

229

Theorem 19. Let us consider binary diagnostic decision tables with m columns labeled by attributes f1 , . . . , fm and n ≥ 4 rows labeled by decisions 1, . . . , n. The fraction of decision tables T for each of which there exist a weight function w and γ α (α, T, w) < Rgreedy (α, T, w) is numbers α, γ such that 0 ≤ γ < α < 1 and Rgreedy at least 1 −

3m n

m −1 2

−

n2 2m .

Proof. We will say that a decision table T is not 1-uniform if there exist two attributes fi and fj of T such that |P (T, fi )| > 0 and |P (T, fj )| ≥ |P (T, fi )| + 2. Otherwise, we will say that T is 1-uniform. Using Theorem 9 and Proposition 3 we conclude that if T is not 1-uniform then there exist a weight function w and γ α (α, T, w) < Rgreedy (α, T, w). numbers α, γ such that 0 ≤ γ < α < 1 and Rgreedy We evaluate the number of simple decision tables which are 1-uniform. Let us consider a simple decision table T which is 1-uniform. Let fi be an attribute of T . It is clear that |P (T, fi )| = 0 if and only if the number of units in the column fi is equal to 0 or n. Let k, l be natural numbers such that k, k + l ∈ {1, . . . , n − 1}, and i, j ∈ {1, . . . , m}, i = j. Let the decision table T have k units in the column fi and k + l units in the column fj . Then |P (T, fi )| = k(n − k) = kn − k 2 and |P (T, fj )| = (k + l)(n − k − l) = kn − k 2 + l(n − 2k − l). Since T is 1-uniform, we have l(n − 2k − l) ∈ {0, 1, −1}. Let l(n − 2k − l) = 0. Then n − 2k − l = 0 and l = n − 2k. Since l is a natural number, we have k < n/2. Let l(n − 2k − l) = 1. Since l, n and k are natural numbers, we have l = 1 and n − 2k − 1 = 1. Therefore k = n2 − 1. Since k is a natural number, we have n is even. Let l(n − 2k − l) = −1. Since l, n and k are natural numbers, we have l = 1 and n − 2k − 1 = −1. Therefore k = n2 . Since k is a natural number, we have n is even. Let n be odd. Then there exists natural k such that 1 ≤ k < n2 and the number of units in each column of T belongs to the set {0, n, k, n − k}. Therefore the n/2 number of considered tables is at most k=1 (Cnk + Cnn−k + 2)m . Since n ≥ 4, n/2 we have 2 ≤ Cn . Using Lemma 2 we concludethat the number of 1-uniform m n/2 n/2 m n √ simple tables is at most k=1 3Cn < n 3·2 . n Let n be even. Then there exists natural k such that 1 ≤ k < n2 − 1 and the number of units in each column of T belongs to the set {0, n, k, n − k}, or the number of units in each column belongs to the set {0, n, n2 − 1, n2 , n2 + 1}. n/2−2 k (Cn + Cnn−k + Therefore the number of considered tables is at most k=1 n/2−1 n/2 n/2+1 m m 2) + (Cn + Cn + Cn + 2) . It is well known (see, for example, [25], n/2 r page 178) that Cn < Cn for anyr ∈ {1,. . . , n} \ {n/2}. Therefore the number n/2

of 1-uniform tables is at most n 3Cn

m

. Using Lemma 2 we conclude that

(as of odd n) the number of 1-uniform simple tables is less than m case mn inn the 3·2 2 3m n √n = m2 −1 . The number of tables which are not simple is at most n

n2 2mn−m . Therefore the number of tables which are not 1-uniform is at least

230

M.Ju. Moshkov, M. Piliszczuk, and B. Zielosko mn m 3 m −1 2

2mn − 2

n

− n2 2mn−m . Thus, the fraction, considered in the statement of the

theorem, is at least 1 −

3m n

m −1 2

−

n2 2m .

m

So if m ≥ 4 and n, 2n2 are large enough then for the most part of binary diagnostic decision tables there exist weight function w and numbers α, γ such that the weight of α-test constructed by greedy algorithm with thresholds α and γ is less than the weight of α-test constructed by greedy algorithm with thresholds α and α. The obtained results show that the greedy algorithm with two thresholds α and γ is of some interest. Now we consider two polynomial modiﬁcations of greedy algorithm which allow to use advantages of greedy algorithm with two thresholds α and γ. Let T be a decision table, P (T ) = ∅, w be a weight function for T and α be a real number such that 0 ≤ α < 1. 1. It is impossible to consider eﬀectively all γ such that 0 ≤ γ ≤ α. Instead of this we can consider all natural N such that M ≤ N ≤ |P (T )| where M = |P (T )|(1 − α) (see the description of greedy algorithm with two thresholds). For each N ∈ {M, . . . , |P (T )|} we apply greedy algorithm with parameters M and N to T and w and after that choose an α-test with minimal weight among constructed α-tests. 2. There exists also an another way to construct an α-test which is not worse than the one obtained under consideration of all N such that M ≤ N ≤ |P (T )|. Let us apply greedy algorithm with thresholds α and α to T and w. Let the algorithm choose sequentially attributes fj1 , . . . , fjt . For each i ∈ {0, . . . , t − 1} we ﬁnd (if it is possible) an attribute fli of T with minimal weight w(fli ) such that the set {fj1 , . . . , fji , fli } is an α-test for T (if i = 0 then it will be the set {fl0 }). After that among constructed α-tests {fj1 , . . . , fjt }, ..., {fj1 , . . . , fji , fli }, ... we choose an α-test with minimal weight. From Proposition 5 it follows that the constructed α-test is not worse than the one constructed under consideration of all γ, 0 ≤ γ ≤ α, or (which is the same) all N , M ≤ N ≤ |P (T )|. Next statement follows immediately from Propositions 1 and 3. Proposition 5. Let T be a decision table, P (T ) = ∅, w be a weight function for T and α, γ be real numbers such that 0 ≤ γ < α < 1. Let the greedy algorithm with two thresholds α and α, which is applied to T and w, choose sequentially attributes fg1 , . . . , fgt . Let the greedy algorithm with two thresholds α and γ, which is applied to T and w, choose sequentially attributes fl1 , . . . , flk . Then either k = t and (l1 , . . . , lk ) = (g1 , . . . , gt ) or k ≤ t, (l1 , . . . , lk−1 ) = (g1 , . . . , gk−1 ) and lk = gk . 3.6

Bounds on Rmin (α) and Rγgreedy (α)

First, we ﬁx some information about the work of greedy algorithm with two thresholds and ﬁnd the best lower bound on the value Rmin (α) depending on this information.

On Partial Covers, Reducts and Decision Rules with Weights

231

Let T be a decision table such that P (T ) = ∅, w be a weight function for T , and α, γ be real numbers such that 0 ≤ γ ≤ α < 1. Let us apply the greedy algorithm with thresholds α and γ to the decision table T and the weight function w. Let during the construction of α-test the greedy algorithm choose sequentially attributes fg1 , . . . , fgt . Let us denote P (T, fg0 ) = ∅ and δ0 = 0. For i = 1, . . . , t denote δi = |P (T, fgi ) \ (P (T, fg0 ) ∪ . . . ∪ P (T, fgi−1 ))| and wi = w(fgi ). As information on the greedy algorithm work we will use numbers MR = MR (α, γ, T, w) = |P (T )|(1 − α) and NR = NR (α, γ, T, w) = |P (T )|(1 − γ) , and tuples ΔR = ΔR (α, γ, T, w) = (δ1 , . . . , δt ) and WR = WR (α, γ, T, w) = (w1 , . . . , wt ). For i = 0, . . . , t − 1 denote wi+1 (MR − (δ0 + . . . + δi )) ρi = . min{δi+1 , NR − (δ0 + . . . + δi )} Let us deﬁne parameter ρR (α, γ) = ρR (α, γ, T, w) as follows: ρR (α, γ) = max {ρi : i = 0, . . . , t − 1} . We will show that ρR (α, γ) is the best lower bound on Rmin (α) depending on MR , NR , ΔR and WR . Next statement follows from Theorem 11 and Propositions 3 and 4. Theorem 20. For any decision table T with P (T ) = ∅, any weight function w for T , and any real numbers α, γ, 0 ≤ γ ≤ α < 1, the inequality Rmin (α, T, w) ≥ ρR (α, γ, T, w) holds, and there exist a decision table T and a weight function w for T such that MR (α, γ, T , w ) = MR (α, γ, T, w), NR (α, γ, T , w ) = NR (α, γ, T, w) , ΔR (α, γ, T , w ) = ΔR (α, γ, T, w), WR (α, γ, T , w ) = WR (α, γ, T, w) , ρR (α, γ, T , w ) = ρR (α, γ, T, w), Rmin (α, T , w ) = ρR (α, γ, T , w ) . Let us consider a property of the parameter ρR (α, γ) which is important for practical use of the bound from Theorem 20. Next statement follows from Propositions 2 and 3. Proposition 6. Let T be a decision table with P (T ) = ∅, w be a weight function for T , α, γ ∈ IR and 0 ≤ γ ≤ α < 1. Then ρR (α, α, T, w) ≥ ρR (α, γ, T, w). Now we study some properties of parameter ρR (α, γ) and obtain two upper γ bounds on the value Rgreedy (α) which do not depend directly on cardinality of the set P (T ) and cardinalities of subsets P (T, fi ). Next statement follows from Theorem 12 and Proposition 3. Theorem 21. Let T be a decision table with P (T ) = ∅, w be a weight function for T and α, γ be real numbers such that 0 ≤ γ < α < 1. Then 1−γ γ Rgreedy (α, T, w) < ρR (γ, γ, T, w) ln +1 . α−γ

232

M.Ju. Moshkov, M. Piliszczuk, and B. Zielosko

Corollary 5. Let ε ∈ IR and 0 < ε < 1. Then for any α, ε ≤ α < 1, the α−ε inequalities ρC (α, α) ≤ Rmin (α) ≤ Rgreedy (α) < ρR (α − ε, α − ε) ln 1ε + 1 hold.

1 1 For example, ln 0.01 + 1 < 5.61 and ln 0.1 + 1 < 3.31. The obtained results show that the lower bound Rmin (α) ≥ ρR (α, α) is nontrivial. Next statement follows from Theorem 13 and Proposition 3. Theorem 22. Let T be a decision table with P (T ) = ∅, w be a weight function for T and α, γ be real numbers such that 0 ≤ γ < α < 1. Then 1−γ γ Rgreedy (α, T, w) < Rmin (γ, T, w) ln +1 . α−γ 0 0.001 (0.001) < 7.91Rmin(0), Rgreedy (0.01) < 5.71Rmin(0.001), Corollary 6. Rgreedy 0.1 0.3 Rgreedy (0.2) < 3.20Cmin(0.1), Rgreedy (0.5) < 2.26Rmin(0.3).

0 Corollary 7. Let 0 < α < 1. Then Rgreedy (α) < Rmin (0) ln α1 + 1 .

Corollary 8. Let ε be a real number, and 0 < ε < 1. Then for any α such that α−ε (α) < Rmin (α − ε) ln 1ε + 1 hold. ε ≤ α < 1 the inequalities Rmin (α) ≤ Rgreedy 3.7

Results of Experiments for α-Tests and α-Reducts

In this subsection we will consider only binary decision tables with binary decision attributes. First Group of Experiments. First group of experiments is connected with study of quality of greedy algorithm with one threshold (where γ = α or, which is the same, N = M ), and comparison of quality of greedy algorithm with one threshold and ﬁrst modiﬁcation of greedy algorithm (where for each N ∈ {M, . . . , |P (T )|} we apply greedy algorithm with parameters M and N to decision table and weight function and after that choose an α-test with minimal weight among constructed α-tests). We generate randomly 1000 decision tables T and weight functions w such that T contains 10 rows and 10 conditional attributes f1 , . . . , f10 , and 1 ≤ w(fi ) ≤ 1000 for i = 1, . . . , 10. For each α ∈ {0.0, 0.1, . . . , 0.9} we ﬁnd the number of pairs (T, w) for which greedy algorithm with one threshold constructs an α-test with minimal weight α (α, T, w) = Rmin (α, T, w). This number is con(an optimal α-test), i.e. Rgreedy tained in the row of Table 1 labeled by ”Opt”. We ﬁnd the number of pairs (T, w) for which ﬁrst modiﬁcation of greedy algorithm constructs an α-test which weight is less than the weight of α-test constructed by greedy algorithm with one threshold, i.e. there exists γ such that γ α (α, T, w) < Rgreedy (α, T, w). This number is contained in 0 ≤ γ < α and Rgreedy the row of Table 1 labeled by ”Impr”. Also we ﬁnd the number of pairs (T, w) for which ﬁrst modiﬁcation of greedy algorithm constructs an optimal α-test which weight is less than the weight of

On Partial Covers, Reducts and Decision Rules with Weights

233

α-test constructed by greedy algorithm with one threshold, i.e. there exists γ γ α such that 0 ≤ γ < α and Rgreedy (α, T, w) = Rmin (α, T, w) < Rgreedy (α, T, w). This number is contained in the row of Table 1 labeled by ”Opt+”. Table 1. Results of ﬁrst group of experiments with α-tests α 0.0 Opt 409 Impr 0 Opt+ 0

0.1 575 42 22

0.2 625 47 28

0.3 826 33 24

0.4 808 24 22

0.5 818 8 5

0.6 950 6 6

0.7 981 5 5

0.8 0.9 992 1000 2 0 2 0

The obtained results show that the percentage of pairs for which greedy algorithm with one threshold ﬁnds an optimal α-test grows almost monotonically (with local minimum near to 0.4–0.5) from 40.9% up to 100%. The percentage of problems for which ﬁrst modiﬁcation of greedy algorithm can improve the result of the work of greedy algorithm with one threshold is less than 5%. However, sometimes (for example, if α = 0.3 or α = 0.7) the considered improvement is noticeable. Second Group of Experiments. Second group of experiments is connected with comparison of quality of greedy algorithm with one threshold and ﬁrst modiﬁcation of greedy algorithm. We make 25 experiments (row ”Nr” in Table 2 contains the number of experiment). Each experiment includes the work with three randomly generated families of pairs (T, w) (1000 pairs in each family) such that T contains n rows and m conditional attributes, and w has values from the set {1, . . . , v}. If the column ”n” contains one number, for example ”40”, it means that n = 40. If this row contains two numbers, for example ”30–120”, it means that for each of 1000 pairs we choose the number n randomly from the set {30, . . . , 120}. The same situation is for the column ”m”. If the column ”α” contains one number, for example ”0.1”, it means that α = 0.1. If this column contains two numbers, for example ”0.2–0.4”, it means that we choose randomly the value of α such that 0.2 ≤ α ≤ 0.4. For each of the considered pairs (T, w) and number α we apply greedy algorithm with one threshold and ﬁrst modiﬁcation of greedy algorithm. Column ”#i”, i = 1, 2, 3, contains the number of pairs (T, w) from the family number i for each of which the weight of α-test, constructed by ﬁrst modiﬁcation of greedy algorithm, is less than the weight of α-test constructed by greedy algorithm with one threshold. In other words, in column ”#i” we have the number of pairs (T, w) from the family number i such that there exists γ for which γ α 0 ≤ γ < α and Rgreedy (α, T, w) < Rgreedy (α, T, w). The column ”avg” contains #1+#2+#3 . the number 3 In experiments 1–3 we consider the case where the parameter v increases. In experiments 4–8 the parameter α increases. In experiments 9–12 the parameter m increases. In experiments 13–16 the parameter n increases. In experiments

234

M.Ju. Moshkov, M. Piliszczuk, and B. Zielosko Table 2. Results of second group of experiments with α-tests Nr n 1 1–50 2 1–50 3 1–50 4 1–50 5 1–50 6 1–50 7 1–50 8 1–50 9 50 10 50 11 50 12 50 13 1–20 14 20–40 15 40–60 16 60–80 17 10 18 10 19 10 20 10 21 10 22 10 23 10 24 10 25 10

m 1–50 1–50 1–50 1–50 1–50 1–50 1–50 1–50 1–20 20–40 40–60 60–80 30 30 30 30 10 10 10 10 10 10 10 10 10

v 1–10 1–100 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000

α 0–1 0–1 0–1 0–0.2 0.2–0.4 0.4–0.6 0.6–0.8 0.8–1 0–0.2 0–0.2 0–0.2 0–0.2 0–0.2 0–0.2 0–0.2 0–0.2 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9

#1 1 5 10 16 23 7 3 1 19 26 21 13 27 34 22 19 36 33 43 30 10 12 3 5 0

#2 2 6 8 20 8 6 5 0 11 24 18 18 26 37 26 14 42 53 25 18 10 13 13 2 0

#3 3 13 11 32 12 5 3 0 22 24 23 22 39 35 23 14 50 46 45 19 13 7 6 6 0

avg 2.00 8.00 9.67 22.67 14.33 6.00 3.67 0.33 17.33 24.67 20.67 17.67 30.67 35.33 23.67 15.67 42.67 44.00 37.67 22.33 11.00 10.67 7.33 4.33 0

17–25 the parameter α increases. The results of experiments show that the value of #i can change from 0 to 53. It means that the percentage of pairs for which ﬁrst modiﬁcation of greedy algorithm is better than the greedy algorithm with one threshold can change from 0% to 5.3%. Third Group of Experiments. Third group of experiments is connected with investigation of quality of lower bound Rmin (α) ≥ ρR (α, α). We choose natural n, m, v and real α, 0 ≤ α < 1. For each chosen tuple (n, m, v, α) we generate randomly 30 pairs (T, w) such that T contains n rows and m conditional attributes, and w has values from the set {1, ..., v}. After α (α, T, w) and ρR (α, α, T, w) for each of generated that we ﬁnd values of Rgreedy α 30 pairs. Note that ρR (α, α, T, w) ≤ Rmin (α, T, w) ≤ Rgreedy (α, T, w). Finally, α we ﬁnd mean values of Rgreedy (α, T, w) and ρR (α, α, T, w) for generated 30 pairs. Results of experiments can be found in Figs. 1 and 2. In these ﬁgures mean values of ρR (α, α, T, w) are called ”average lower bound” and mean values of α Rgreedy (α, T, w) are called ”average upper bound”.

On Partial Covers, Reducts and Decision Rules with Weights

235

In Fig. 1 (left-hand side) one can see the case when n ∈ {1000, 2000, . . . , 5000}, m = 30, v = 1000 and α = 0.01. In Fig. 1 (right-hand side) one can see the case when n = 1000, m ∈ {10, 20, . . . , 100}, v = 1000 and α = 0.01. 1000

2500 average upper bound average lower bound

average upper bound average lower bound

900 2000 800

700

1500

600 1000

500

400 500 300

200 1000

0 1500

2000

2500

3000 number of rows

3500

4000

4500

5000

10

20

30

40

50 60 number of attributes

70

80

90

100

Fig. 1. Results of third group of experiments with α-tests (n and m are changing)

1000

6000 average upper bound average lower bound

average upper bound average lower bound

900 5000 800 700 4000 600 500

3000

400 2000 300 200 1000 100 0 100

0 200

300

400

500 600 range of weight

700

800

900

1000

0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

alpha

Fig. 2. Results of third group of experiments with α-tests (v and α are changing)

In Fig. 2 (left-hand side) one can see the case when n = 1000, m = 30, v ∈ {100, 200, . . . , 1000} and α = 0.01. In Fig. 2 (right-hand side) one can see the case when n = 1000, m = 30, v = 1000 and α ∈ {0.0, 0.1, . . . , 0.9}. Results of experiments show that the considered lower bound is nontrivial and can be useful in investigations.

4

Partial Decision Rules

In this section we omit reasoning on relationships between partial covers and partial decision rules including reductions of one problem to another (description of such reductions can be found in [9]) and two propositions similar to Propositions 3 and 4.

236

4.1

M.Ju. Moshkov, M. Piliszczuk, and B. Zielosko

Main Notions

Let T be a table with n rows labeled by nonnegative integers (decisions) and m columns labeled by attributes (names of attributes) f1 , . . . , fm . This table is ﬁlled by nonnegative integers (values of attributes). The table T is called a decision table. Let w be a weight function for T which corresponds to each attribute fi a natural number w(fi ). Let r = (b1 , . . . , bm ) be a row of T labeled by a decision d. Let us denote by U (T, r) the set of rows from T which are diﬀerent from r and are labeled by decisions diﬀerent from d. We will say that an attribute fi separates rows r and r ∈ U (T, r) if rows r and r have diﬀerent numbers at the intersection with the column fi . For i = 1, . . . , m denote by U (T, r, fi ) the set of rows from U (T, r) which attribute fi separates from the row r. Let α be a real number such that 0 ≤ α < 1. A decision rule fi1 = bi1 ∧ . . . ∧ fit = bit → d

(3)

. , fit separate from r is called an α-decision rule for T and r if attributes fi1 , . . at least (1 − α)|U (T, r)| rows from U (T, r). The number tj=1 w(fij ) is called the weight of the considered decision rule. If U (T, r) = ∅ then for any fi1 , . . . , fit ∈ {f1 , . . . , fm } the rule (3) is an αdecision rule for T and r. Also, the rule (3) with empty left-hand side (when t = 0) is an α-decision rule for T and r. The weight of this rule is equal to 0. For example, 0.01-decision rule means that we must separate from r at least 99% of rows from U (T, r). Note that 0-rule is usual (exact) rule. Let us denote by Lmin (α) = Lmin (α, T, r, w) the minimal weight of α-decision rule for T and r. Let α, γ be real numbers such that 0 ≤ γ ≤ α < 0. Let us describe a greedy algorithm with thresholds α and γ which constructs an α-decision rule for given T , r and weight function w. Let r = (b1 , . . . , bm ), and r be labeled by the decision d. The right-hand side of constructed α-decision rule is equal to d. If U (T, r) = ∅ then the left-hand side of constructed α-decision rule is empty. Let U (T, r) = ∅. Let us denote M = |U (T, r)|(1 − α) and N = |U (T, r)|(1 − γ) . Let we make i ≥ 0 steps and construct a decision rule R with i conditions (if i = 0 then the left-hand side of R is empty). Let us describe the step number i + 1. Let us denote by D the set of rows from U (T, r) separated from r by attributes belonging to R (if i = 0 then D = ∅). If |D| ≥ M then we ﬁnish the work of the algorithm, and R is the constructed α-decision rule. Let |D| < M . Then we choose an attribute fj with minimal number j for which U (T, r, fj ) \ D = ∅ and the value w(fj ) min{|U (T, r, fj ) \ D|, N − |D|} is minimal. Add the condition fj = bj to R. Pass to the step number i + 2. Let us denote by Lγgreedy (α) = Lγgreedy (α, T, r, w) the weight of α-decision rule constructed by the considered algorithm for given table T , row r and weight function w.

On Partial Covers, Reducts and Decision Rules with Weights

4.2

237

On Precision of Greedy Algorithm with Thresholds α and α

The following two statements are simple corollaries of results of Slav´ık (see Theorems 4 and 5). Theorem 23. Let T be a decision table, r be a row of T , U (T, r) = ∅, w be a weight function for T , and α be a real number such that 0 ≤ α < 1. Then Lα greedy (α) ≤ Lmin (α)H ((1 − α)|U (T, r)| ). Theorem 24. Let T be a decision table with m columns labeled by attributes f1 , . . . , fm , r be a row of T , U (T, r) = ∅, w be

a weight function for T , α ∈ IR and 0 ≤ α < 1. Then Lα greedy (α) ≤ Lmin (α)H maxi∈{1,...,m} |U (T, r, fi )| . 4.3

On Polynomial Approximate Algorithms

In this subsection we consider three theorems which follow immediately from Theorems 39–41 [9]. Let 0 ≤ α < 1. Let us consider the following problem: for given decision table T , row r of T and weight function w for T it is required to ﬁnd an α-decision rule for T and r with minimal weight. Theorem 25. Let 0 ≤ α < 1. Then the problem of construction of α-decision rule with minimal weight is N P -hard. So we must consider polynomial approximate algorithms for minimization of α-decision rule weight. Theorem 26. Let α ∈ IR and 0 ≤ α < 1. If N P ⊆ DT IM E(nO(log log n) ) then for any ε, 0 < ε < 1, there is no polynomial algorithm that for given decision table T , row r of T with U (T, r) = ∅ and weight function w for T constructs α-decision rule for T and r which weight is at most (1−ε)Lmin(α, T, r, w) ln |U (T, r)|. Theorem 27. Let α be a real number such that 0 ≤ α < 1. If P = N P then there exists δ > 0 such that there is no polynomial algorithm that for given decision table T , row r of T with U (T, r) = ∅ and weight function w for T constructs α-decision rule for T and r which weight is at most δLmin (α, T, r, w) ln |U (T, r)|. From Theorem 23 it follows that Lα greedy (α) ≤ Lmin (α)(1 + ln |U (T, r)|). From this inequality and from Theorem 26 it follows that under the assumption N P ⊆ DT IM E(nO(log log n) ) greedy algorithm with two thresholds α and α is close to best polynomial approximate algorithms for minimization of partial decision rule weight. From the considered inequality and from Theorem 27 it follows that under the assumption P = N P greedy algorithm with two thresholds α and α is not far from best polynomial approximate algorithms for minimization of partial decision rule weight.

238

M.Ju. Moshkov, M. Piliszczuk, and B. Zielosko

However we can try to improve the results of the work of greedy algorithm with two thresholds α and α for some part of decision tables. 4.4

Two Modiﬁcations of Greedy Algorithm

First, we consider binary diagnostic decision tables and prove that under some assumptions on the number of attributes and rows for the most part of tables for each row there exist weight function w and numbers α, γ such that the weight of α-decision rule constructed by greedy algorithm with thresholds α and γ is less than the weight of α-decision rule constructed by greedy algorithm with thresholds α and α. Binary means that the table is ﬁlled by numbers from the set {0, 1} (all attributes have values from {0, 1}). Diagnostic means that rows of the table are labeled by pairwise diﬀerent numbers (decisions). Let T be a binary diagnostic decision table with m columns labeled by attributes f1 , . . . , fm and with n rows. We will assume that rows of T with numbers 1, . . . , n are labeled by decisions 1, . . . , n respectively. Therefore the number of considered tables is equal to 2mn . Decision table will be called simple if it has no equal rows. Theorem 28. Let us consider binary diagnostic decision tables with m columns labeled by attributes f1 , . . . , fm and n ≥ 5 rows labeled by decisions 1, . . . , n. The fraction of decision tables T for each of which for each row r of T there exist a weight function w and numbers α, γ such that 0 ≤ γ < α < 1 and Lγgreedy (α, T, r, w) < Lα greedy (α, T, r, w) is at least 1−

n2 n3m − . m 2m (n − 1) 2 −1

Proof. Let T be a decision table and r be a row of T with number s ∈ {1, . . . , n}. We will say that a decision table T is 1-uniform relatively r if there exists natural p such that for any attribute fi of T if |U (T, r, fi )| > 0 then |U (T, r, fi )| ∈ {p, p+1}. Using reasoning similar to the proof of Theorem 9 one can show that if T is not 1-uniform relatively r then there exist a weight function w and numbers α, γ such that 0 ≤ γ < α < 1 and Lγgreedy (α, T, r, w) < Lα greedy (α, T, r, w). We evaluate the number of decision tables which are not 1-uniform relatively each row. Let (δ1 , . . . , δm ) ∈ {0, 1}m. First, we evaluate the number of simple decision tables for which r = (δ1 , . . . , δm ) and which are 1-uniform relatively r. Let us consider such a decision table T . It is clear that there exists p ∈ {1, . . . , n − 2} such that for i = 1, . . . , m the column fi contains exactly 0 or p or p + 1 numbers ¬δi . Thereforethe number of considered decision tables m n−2 p p+1 is at most p=1 Cn−1 + Cn−1 + 1 . Using Lemma 2 we conclude that this m n−1 m mn−m m (n−1)/2 3 √ < (n − 1) 3·2 = 2 number is at most (n − 2) 3Cn−1 m −1 . n−1 2 (n−1)

There are 2m variants for the choice of the tuple (δ1 , . . . , δm ) and n variants for the choice of the number s of row r. Therefore the number of simple decision mn−m m 3 tables which are 1-uniform relatively at least one row is at most n2m 2 m −1 = 2 (n−1)

On Partial Covers, Reducts and Decision Rules with Weights n2mn 3m . m (n−1) 2 −1

239

The number of tables which are not simple is at most n2 2mn−m .

Hence the number of tables which are not 1-uniform for each row is at least mn m 2mn − n2 m23 −1 − n2 2mn−m . Thus, the fraction, considered in the statement of (n−1)

the theorem, is at least 1 −

n3m m (n−1) 2 −1

−

n2 2m .

m

So if m ≥ 6 and n, 2n2 are large enough then for the most part of binary diagnostic decision tables for each row there exist weight function w and numbers α, γ such that the weight of α-decision rule constructed by greedy algorithm with thresholds α and γ is less than the weight of α-decision rule constructed by greedy algorithm with thresholds α and α. The obtained results show that the greedy algorithm with two thresholds α and γ is of some interest. Now we consider two polynomial modiﬁcations of greedy algorithm which allow to use advantages of greedy algorithm with two thresholds α and γ. Let T be a decision table with m columns labeled by attributes f1 , . . . , fm , r = (b1 , . . . , bm ) be a row of T labeled by decision d, U (T, r) = ∅, w be a weight function for T and α be a real number such that 0 ≤ α < 1. 1. It is impossible to consider eﬀectively all γ such that 0 ≤ γ ≤ α. Instead of this we can consider all natural N such that M ≤ N ≤ |U (T, r)| where M = |U (T, r)|(1 − α) (see the description of greedy algorithm with two thresholds). For each N ∈ {M, . . . , |U (T, r)|} we apply greedy algorithm with parameters M and N to T , r and w and after that choose an α-decision rule with minimal weight among constructed α-decision rules. 2. There exists also an another way to construct an α-decision rule which is not worse than the one obtained under consideration of all N such that M ≤ N ≤ |U (T, r)|. Let us apply greedy algorithm with thresholds α and α to T , r and w. Let the algorithm choose sequentially attributes fj1 , . . . , fjt . For each i ∈ {0, . . . , t − 1} we ﬁnd (if it is possible) an attribute fli of T with minimal weight w(fli ) such that the rule fj1 = bj1 ∧ . . . ∧ fji = bji ∧ fli = bli → d is an α-decision rule for T and r (if i = 0 then it will be the rule fl0 = bl0 → d). After that among constructed α-decision rules fj1 = bj1 ∧ . . . ∧ fjt = bjt → d, ..., fj1 = bj1 ∧ . . . ∧ fji = bji ∧ fli = bli → d, ... we choose an α-decision rule with minimal weight. From Proposition 7 it follows that the constructed α-decision rule is not worse than the one constructed under consideration of all γ, 0 ≤ γ ≤ α, or (which is the same) all N , M ≤ N ≤ |U (T, r)|. Using Propositions 1 one can prove the following statement. Proposition 7. Let T be a decision table, r be a row of T , U (T, r) = ∅, w be a weight function for T and α, γ be real numbers such that 0 ≤ γ < α < 1. Let the greedy algorithm with two thresholds α and α, which is applied to T , r and w, choose sequentially attributes fg1 , . . . , fgt . Let the greedy algorithm with two thresholds α and γ, which is applied to T , r and w, choose sequentially attributes fl1 , . . . , flk . Then either k = t and (l1 , . . . , lk ) = (g1 , . . . , gt ) or k ≤ t, (l1 , . . . , lk−1 ) = (g1 , . . . , gk−1 ) and lk = gk .

240

4.5

M.Ju. Moshkov, M. Piliszczuk, and B. Zielosko

Bounds on Lmin (α) and Lγgreedy (α)

First, we ﬁx some information about the work of greedy algorithm with two thresholds and ﬁnd the best lower bound on the value Lmin (α) depending on this information. Let T be a decision table, r be a row of T such that U (T, r) = ∅, w be a weight function for T , and α, γ be real numbers such that 0 ≤ γ ≤ α < 1. Let us apply the greedy algorithm with thresholds α and γ to the decision table T , row r and the weight function w. Let during the construction of α-decision rule the greedy algorithm choose sequentially attributes fg1 , . . . , fgt . Let us denote U (T, r, fg0 ) = ∅ and δ0 = 0. For i = 1, . . . , t denote δi = |U (T, r, fgi ) \ (U (T, r, fg0 ) ∪ . . . ∪ U (T, r, fgi−1 ))| and wi = w(fgi ). As information on the greedy algorithm work we will use numbers ML = ML (α, γ, T, r, w) = |U (T, r)|(1 − α) , NL = NL (α, γ, T, r, w) = |U (T, r)|(1 − γ) and tuples ΔL = ΔL (α, γ, T, r, w) = (δ1 , . . . , δt ), WL = WL (α, γ, T, r, w) = (w1 , . . . , wt ). For i = 0, . . . , t − 1 denote wi+1 (ML − (δ0 + . . . + δi )) ρi = . min{δi+1 , NL − (δ0 + . . . + δi )} Let us deﬁne parameter ρL (α, γ) = ρL (α, γ, T, r, w) as follows: ρL (α, γ) = max {ρi : i = 0, . . . , t − 1} . We will show that ρL (α, γ) is the best lower bound on Lmin (α) depending on ML , NL , ΔL and WL . Using Theorem 11 one can prove the following statement. Theorem 29. For any decision table T , any row r of T with U (T, r) = ∅, any weight function w for T , and any real numbers α, γ, 0 ≤ γ ≤ α < 1, the inequality Lmin (α, T, r, w) ≥ ρL (α, γ, T, r, w) holds, and there exist a decision table T , a row r of T and a weight function w for T such that ML (α, γ, T , r , w ) = ML (α, γ, T, r, w), NL (α, γ, T , r , w ) = NL (α, γ, T, r, w) , ΔL (α, γ, T , r , w ) = ΔL (α, γ, T, r, w), WL (α, γ, T , r , w ) = WL (α, γ, T, r, w) , ρL (α, γ, T , r , w ) = ρL (α, γ, T, r, w), Lmin (α, T , r , w ) = ρL (α, γ, T , r , w ) . Let us consider a property of the parameter ρL (α, γ) which is important for practical use of the bound from Theorem 29. Using Proposition 2 one can prove the following statement. Proposition 8. Let T be a decision table, r be a row of T with U (T, r) = ∅, w be a weight function for T , and α, γ be real numbers such that 0 ≤ γ ≤ α < 1. Then ρL (α, α, T, r, w) ≥ ρL (α, γ, T, r, w). Now we study some properties of parameter ρL (α, γ) and obtain two upper bounds on the value Lγgreedy (α) which do not depend directly on cardinality of the set U (T, r) and cardinalities of subsets U (T, r, fi ). Using Theorem 12 one can prove the following statement.

On Partial Covers, Reducts and Decision Rules with Weights

241

Theorem 30. Let T be a decision table, r be a row of T with U (T, r) = ∅, w be a weight function for T , α, γ ∈ IR and 0 ≤ γ < α < 1. Then Lγgreedy (α, T, r, w) < 1−γ +1 . ρL (γ, γ, T, r, w) ln α−γ Corollary 9. Let ε ∈ IR and 0 < ε < 1. Then for any α, ε ≤ α < 1, the 1 inequalities ρL (α, α) ≤ Lmin (α) ≤ Lα−ε greedy (α) < ρL (α − ε, α − ε) ln ε + 1 hold.

1

1 + 1 < 5.61 and ln 0.1 + 1 < 3.31. The obtained results For example, ln 0.01 show that the lower bound Lmin (α) ≥ ρL (α, α) is nontrivial. Using Theorem 13 one can prove the following statement. Theorem 31. Let T be a decision table, r be a row of T with U (T, r) = ∅, w be a weight function for T , α, γ ∈ IR and 0 ≤ γ < α < 1. Then Lγgreedy (α, T, r, w) < 1−γ +1 . Lmin (γ, T, r, w) ln α−γ Corollary 10. L0greedy (0.001) < 7.91Lmin(0), L0.001 greedy (0.01) < 5.71Lmin (0.001), 0.3 (0.2) < 3.20L (0.1), L (0.5) < 2.26L L0.1 min min(0.3). greedy greedy

Corollary 11. Let 0 < α < 1. Then L0greedy (α) < Lmin(0) ln α1 + 1 . Corollary 12. Let ε be a real number, and 0 < ε < 1. Then for any α such that 1 (α) < L (α − ε) ln + 1 hold. ε ≤ α < 1 the inequalities Lmin (α) ≤ Lα−ε min greedy ε 4.6

Results of Experiments for α-Decision Rules

In this subsection we will consider only binary decision tables T with binary decision attributes. First Group of Experiments. First group of experiments is connected with study of quality of greedy algorithm with one threshold (where γ = α or, which is the same, N = M ), and comparison of quality of greedy algorithm with one threshold and ﬁrst modiﬁcation of greedy algorithm (where for each N ∈ {M, . . . , |U (T, r)|} we apply greedy algorithm with parameters M and N to decision table, row and weight function and after that choose an α-decision rule with minimal weight among constructed α-decision rules). We generate randomly 1000 decision tables T , rows r and weight functions w such that T contains 40 rows and 10 conditional attributes f1 , . . . , f10 , r is the ﬁrst row of T , and 1 ≤ w(fi ) ≤ 1000 for i = 1, . . . , 10. For each α ∈ {0.1, . . . , 0.9} we ﬁnd the number of triples (T, r, w) for which greedy algorithm with one threshold constructs an α-decision rule with minimal weight (an optimal α-decision rule), i.e. Lα greedy (α, T, r, w) = Lmin (α, T, r, w). This number is contained in the row of Table 3 labeled by ”Opt”. We ﬁnd the number of triples (T, r, w) for which ﬁrst modiﬁcation of greedy algorithm constructs an α-decision rule which weight is less than the weight of α-decision rule constructed by greedy algorithm with one threshold, i.e. there

242

M.Ju. Moshkov, M. Piliszczuk, and B. Zielosko

exists γ such that 0 ≤ γ < α and Lγgreedy (α, T, r, w) < Lα greedy (α, T, r, w). This number is contained in the row of Table 3 labeled by ”Impr”. Also we ﬁnd the number of triples (T, r, w) for which ﬁrst modiﬁcation of greedy algorithm constructs an optimal α-decision rule which weight is less than the weight of α-decision rule constructed by greedy algorithm with one threshold, i.e. there exists γ such that 0 ≤ γ < α and Lγgreedy (α, T, r, w) = Lmin(α, T, r, w) < Lα greedy (α, T, r, w). This number is contained in the row of Table 3 labeled by ”Opt+”. Table 3. Results of ﬁrst group of experiments with α-decision rules α 0.0 Opt 434 Impr 0 Opt+ 0

0.1 559 31 16

0.2 672 51 35

0.3 800 36 28

0.4 751 22 17

0.5 733 27 26

0.6 866 30 25

0.7 966 17 13

0.8 0.9 998 1000 1 0 1 0

The obtained results show that the percentage of triples for which greedy algorithm with one threshold ﬁnds an optimal α-decision rule grows almost monotonically (with local minimum near to 0.4–0.5) from 43.4% up to 100%. The percentage of problems for which ﬁrst modiﬁcation of greedy algorithm can improve the result of the work of greedy algorithm with one threshold is less than 6%. However, sometimes (for example, if α = 0.3, α = 0.6 or α = 0.7) the considered improvement is noticeable. Second Group of Experiments. Second group of experiments is connected with comparison of quality of greedy algorithm with one threshold and ﬁrst modiﬁcation of greedy algorithm. We make 25 experiments (row ”Nr” in Table 4 contains the number of experiment). Each experiment includes the work with three randomly generated families of triples (T, r, w) (1000 triples in each family) such that T contains n rows and m conditional attributes, r is the ﬁrst row of T , and w has values from the set {1, . . . , v}. If the column ”n” contains one number, for example ”40”, it means that n = 40. If this row contains two numbers, for example ”30–120”, it means that for each of 1000 triples we choose the number n randomly from the set {30, . . . , 120}. The same situation is for the column ”m”. If the column ”α” contains one number, for example ”0.1”, it means that α = 0.1. If this column contains two numbers, for example ”0.2–0.4”, it means that we choose randomly the value of α such that 0.2 ≤ α ≤ 0.4. For each of the considered triples (T, r, w) and number α we apply greedy algorithm with one threshold and ﬁrst modiﬁcation of greedy algorithm. Column ”#i”, i = 1, 2, 3, contains the number of triples (T, r, w) from the family number i for each of which the weight of α-decision rule, constructed by ﬁrst modiﬁcation of greedy algorithm, is less than the weight of α-decision rule constructed by

On Partial Covers, Reducts and Decision Rules with Weights

243

Table 4. Results of second group of experiments with α-decision rules Nr 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25

n 1–100 1–100 1–100 1–100 1–100 1–100 1–100 1–100 100 100 100 100 1–30 30–60 60–90 90–120 40 40 40 40 40 40 40 40 40

m 1–100 1–100 1–100 1–100 1–100 1–100 1–100 1–100 1–30 30–60 60–90 90–120 30 30 30 30 10 10 10 10 10 10 10 10 10

v 1–10 1–100 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000 1–1000

α 0–1 0–1 0–1 0–0.2 0.2–0.4 0.4–0.6 0.6–0.8 0.8–1 0–0.2 0–0.2 0–0.2 0–0.2 0–0.2 0–0.2 0–0.2 0–0.2 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9

#1 4 7 19 20 28 22 7 0 35 47 45 37 11 20 30 40 31 37 35 27 32 28 10 0 0

#2 2 14 13 39 29 23 6 1 38 43 51 40 8 22 33 48 39 39 30 20 32 26 12 2 0

#3 4 13 15 22 28 34 4 0 28 31 36 55 9 35 34 38 34 47 37 27 36 24 10 0 0

avg 3.33 11.33 15.67 27.00 28.33 26.33 5.67 0.33 33.67 40.33 44.00 44.00 9.33 25.67 32.33 42.00 34.67 41.00 34.00 24.67 33.33 26.00 10.67 0.67 0

greedy algorithm with one threshold. In other words, in column ”#i” we have the number of triples (T, r, w) from the family number i such that there exists γ for which 0 ≤ γ < α and Lγgreedy (α, T, r, w) < Lα greedy (α, T, r, w). The column #1+#2+#3 . ”avg” contains the number 3 In experiments 1–3 we consider the case where the parameter v increases. In experiments 4–8 the parameter α increases. In experiments 9–12 the parameter m increases. In experiments 13–16 the parameter n increases. In experiments 17–25 the parameter α increases. The results of experiments show that the value of #i can change from 0 to 55. It means that the percentage of triples for which the ﬁrst modiﬁcation of greedy algorithm is better than the greedy algorithm with one threshold can change from 0% to 5.5%. Third Group of Experiments. Third group of experiments is connected with investigation of quality of lower bound Lmin (α) ≥ ρL (α, α). We choose natural n, m, v and real α, 0 ≤ α < 1. For each chosen tuple (n, m, v, α) we generate randomly 30 triples (T, r, w) such that T contains

244

M.Ju. Moshkov, M. Piliszczuk, and B. Zielosko

1000

2500 average upper bound average lower bound

average upper bound average lower bound

900 2000 800

700

1500

600 1000

500

400 500 300

200 1000

0 1500

2000

2500

3000 number of rows

3500

4000

4500

5000

10

20

30

40

50 60 number of attributes

70

80

90

100

Fig. 3. Results of third group of experiments with rules (n and m are changing)

900

1600 average upper bound average lower bound

average upper bound average lower bound

800

1400

700

1200

600 1000 500 800 400 600 300 400

200

200

100

0 100

0 200

300

400

500 600 range of weight

700

800

900

1000

0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

alpha

Fig. 4. Results of third group of experiments with rules (v and α are changing)

n rows and m conditional attributes, r is the ﬁrst row of T , and w has values from the set {1, ..., v}. After that we ﬁnd values of Lα greedy (α, T, r, w) and ρL (α, α, T, r, w) for each of generated 30 triples. Note that ρL (α, α, T, r, w) ≤ Lmin (α, T, r, w) ≤ Lα greedy (α, T, r, w). Finally, for generated 30 triples we ﬁnd mean values of Lα greedy (α, T, r, w) and ρL (α, α, T, r, w). Results of experiments can be found in Figs. 3 and 4. In these ﬁgures mean values of ρL (α, α, T, r, w) are called ”average lower bound” and mean values of Lα greedy (α, T, r, w) are called ”average upper bound”. In Fig. 3 (left-hand side) one can see the case when n ∈ {1000, 2000, . . . , 5000}, m = 30, v = 1000 and α = 0.01. In Fig. 3 (right-hand side) one can see the case when n = 1000, m ∈ {10, 20, . . . , 100}, v = 1000 and α = 0.01. In Fig. 4 (left-hand side) one can see the case when n = 1000, m = 30, v ∈ {100, 200, . . . , 1000} and α = 0.01. In Fig. 4 (right-hand side) one can see the case when n = 1000, m = 30, v = 1000 and α ∈ {0.0, 0.1, . . . , 0.9}. Results of experiments show that the considered lower bound is nontrivial and can be useful in investigations.

On Partial Covers, Reducts and Decision Rules with Weights

5

245

Conclusions

The paper is devoted (mainly) to theoretical and experimental analysis of greedy algorithms with weights and their modiﬁcations for partial cover, reduct and decision rule construction. Obtained results will further to more wide use of such algorithms in rough set theory and its applications. In the further investigations we are planning to generalize the obtained results to the case of decision tables which can contain missing values, continuous attributes, and discrete attributes with large number of values.

Acknowledgments The authors are greatly indebted to Andrzej Skowron for stimulating discussions and to anonymous reviewers for helpful remarks and suggestions.

References 1. Cheriyan, J., Ravi, R.: Lecture Notes on Approximation Algorithms for Network Problems (1998) http://www.math.uwaterloo.ca/˜jcheriya/lecnotes.html 2. Chv´ atal, V.: A greedy heuristic for the set-covering problem. Mathematics of Operations Research 4 (1979) 233–235. 3. Feige, U.: A threshold of ln n for approximating set cover (Preliminary version). Proceedings of 28th Annual ACM Symposium on the Theory of Computing (1996) 314–318. 4. Gavrilov G.P., Sapozhenko, A.A.: Problems and Exercises in Discrete Mathematics (third edition). Fizmatlit, Moscow, 2004 (in Russian). 5. Kearns, M.J.: The Computational Complexity of Machine Learning. MIT Press, Cambridge, Massachussetts, 1990. 6. Moshkov, M.Ju.: Greedy algorithm for set cover in context of knowledge discovery problems. In: Proceedings of the International Workshop on Rough Sets in Knowledge Discovery and Soft Computing (ETAPS 2003 Satellite Event). Warsaw, Poland. Electronic Notes in Theoretical Computer Science 82 (2003). 7. Moshkov, M.Ju.: On greedy algorithm for partial cover construction. In: Proceedings of the Fourteenth International Workshop Design and Complexity of Control Systems. Nizhny Novgorod, Russia (2003) 57 (in Russian). 8. Moshkov, M.Ju., Piliszczuk, M., Zielosko, B.: Greedy algorithm for construction of partial covers. In: Proceedings of the Fourteenth International Conference Problems of Theoretical Cybernetics. Penza, Russia (2005) 103 (in Russian). 9. Moshkov, M.Ju., Piliszczuk, M., Zielosko, B.: On partial covers, reducts and decision rules. LNCS Transactions on Rough Sets, Springer-Verlag (submitted). ´ ezak, D.: Approximate reducts and association rules - correspon10. Nguyen, H.S., Sl¸ dence and complexity results. In: Proceedings of the Seventh International Workshop on Rough Sets, Fuzzy Sets, Data Mining, and Granular-Soft Computing. Yamaguchi, Japan. Lecture Notes in Artiﬁcial Intelligence 1711, Springer-Verlag (1999) 137–145. 11. Pawlak, Z.: Rough Sets – Theoretical Aspects of Reasoning about Data. Kluwer Academic Publishers, Dordrecht, Boston, London, 1991.

246

M.Ju. Moshkov, M. Piliszczuk, and B. Zielosko

12. Pawlak, Z.: Rough set elements. In: Polkowski, L., Skowron, A. (Eds.), Rough Sets in Knowledge Discovery 1. Methodology and Applications (Studies in Fuzziness and Soft Computing 18). Phisica-Verlag. A Springer-Verlag Company (1998) 10–30. 13. Piliszczuk, M.: On greedy algorithm for partial reduct construction. In: Proceedings of Concurrency, Speciﬁcation and Programming Workshop 2. Ruciane-Nida, Poland (2005) 400–411. 14. Quafafou, M.: α-RST: a generalization of rough set theory. Information Sciences 124 (2000) 301–316. 15. Raz, R., Safra, S.: A sub-constant error-probability low-degree test, and subconstant error-probability PCP characterization of NP. In: Proceedings of 29th Annual ACM Symposium on the Theory of Computing (1997) 475–484. 16. Skowron, A.: Rough sets in KDD. In: Proceedings of the 16-th World Computer Congress (IFIP’2000). Beijing, China (2000) 1–14. 17. Skowron, A., Rauszer, C.: The discernibility matrices and functions in information systems. In: Slowinski, R. (Ed.), Intelligent Decision Support. Handbook of Applications and Advances of the Rough Set Theory. Kluwer Academic Publishers, Dordrecht, Boston, London (1992) 331–362. 18. Slav´ık, P.: Approximation algorithms for set cover and related problems. Ph.D. thesis. University of New York at Buﬀalo (1998). ´ ezak, D.: Approximate reducts in decision tables. In: Proceedings of the Congress 19. Sl¸ Information Processing and Management of Uncertainty in Knowledge-based Systems 3. Granada, Spain (1996) 1159–1164. ´ ezak, D.: Normalized decision functions and measures for inconsistent decision 20. Sl¸ tables analysis. Fundamenta Informaticae 44 (2000) 291–319. ´ ezak, D.: Approximate decision reducts. Ph.D. thesis. Warsaw University (2001) 21. Sl¸ (in Polish). ´ ezak, D.: Approximate entropy reducts. Fundamenta Informaticae 53 (2002) 22. Sl¸ 365–390. ´ ezak, D., Wr´ 23. Sl¸ oblewski, J.: Order-based genetic algorithms for the search of approximate entropy reducts. In: Proceedings of the International Conference Rough Sets, Fuzzy Sets, Data Mining, and Granular Computing. Chongqing, China. Lecture Notes in Artiﬁcial Intelligence 2639, Springer-Verlag (2003) 308–311. 24. Wr´ oblewski, J.: Ensembles of classiﬁers based on approximate reducts. Fundamenta Informaticae 47 (2001) 351–360. 25. Yablonskii, S.V.: Introduction into Discrete Mathematics (fourth edition). Vishaya Shkola, Moscow, 2003 (in Russian). 26. Ziarko, W.: Analysis of uncertain information in the framework of variable precision rough sets. Foundations of Computing and Decision Sciences 18 (1993) 381–396. 27. Zielosko, B.: On partial decision rules. In: Proceedings of Concurrency, Speciﬁcation and Programming Workshop 2. Ruciane-Nida, Poland (2005) 598–609.

A Personal View on AI, Rough Set Theory and Professor Pawlak Toshinori Munakata Computer and Information Science Department, Cleveland State University Cleveland, Ohio 44115, U.S.A. t.munakata@csuohio.edu

It is an honor to contribute my short article to this special issue commemorating the life and work of Professor Zdzislaw Pawlak. In this article I would like to discuss my encounters with the ﬁeld of artiﬁcial intelligence (AI) in general, and how I see rough set theory and Professor Zdzislaw Pawlak in this context. I have been fortunate to know some of the greatest scholars in the AI ﬁeld. There are many of them, but if I had to choose the three I admire most, they are: Professors Zdzislaw Pawlak, Lotﬁ Zadeh and Herbert A. Simon. There are common characteristics among all of them. Although they are the most prominent of scholars, all are frank and easy and pleasant to talk with. All are professionally active at ages where ordinary people would have long since retired. I became interested in the ﬁeld of AI in the mid 70s. I have observed many ups and downs of the ﬁeld in terms of the popularity since then - a common phenomena in any ﬁeld. The AAAI (American Association for Artiﬁcial Intelligence) was inaugurated and the ﬁrst issue of the AI Magazine was published in the spring 1980. The timing of the birth of rough set theory was soon after this event. At this time many people in the world were becoming interested in the ﬁeld of AI, while there were only a handful researchers when the ﬁeld started in the 1950s. In the spring of 1986, the ﬁrst issue of the IEEE Expert (now IEEE Intelligent Systems) was inaugurated. I served as an Associate Editor of this magazine for two terms from 1987 to 1991. In terms of public popularity AI was ﬂourishing in this eras. During many years of the 70s and 80s, I observed that despite media hype and claims for break-thorough technologies, most AI techniques were not practical. Here “practical” means “having real-world commercial and industrial applications on an everyday basis.” For example, I could not ﬁnd cases where machine learning techniques discussed in textbooks such as “learning from examples” and “learning from analogy” were actually employed at industrial plants or commercial banks. The same were true for other AI techniques such as blackboard and neural networks. After Minsky’s pessimistic view on the ﬁeld, the U.S. government funding ceased, and only a handful researchers remained active in the ﬁeld. The ﬁeld revived in the mid to late 80s, and became quite popular. However, I could not ﬁnd a single case where neural networks were actually used every day for commercial and industrial applications. For all of these observations I could be wrong because there could have been exceptions I was not aware of, but I was certain that these exceptions were few, if any. J.F. Peters et al. (Eds.): Transactions on Rough Sets VI, LNCS 4374, pp. 247–252, 2007. c Springer-Verlag Berlin Heidelberg 2007

248

T. Munakata

This situation of impracticality of most AI techniques appeared to start to change around 1990. That is, many AI techniques were becoming truly practical and their application domains much broader. Of course, there were practical AI techniques before 1990. Robotics was one. The ﬁrst major industrial fuzzy control was implemented in Denmark in 1982, followed by the famous Hitachi’s Sendai subway control in 1986. There had been practical expert systems in the 80s. However, the repertories of AI techniques and their application domains were becoming much more extensive around 1990. With this background, in March 1993 I made a proposal to the Communications of the ACM (CACM ) for a Special Issue entitled “Commercial and Industrial AI.” The CACM was a primary computer science magazine subscribed by some 85,000 professionals worldwide at that time. Its readers went far beyond the computer science discipline including ﬁelds such as engineering, social science and education. The proposal was accepted and I selected the most practical or promising AI areas with the help of many experts. The Special Issue was published in March 1994 [3] and was divided into four major sections with 11 articles. They are: I. “Knowledge Engineering Systems” with two articles – an article on general expert systems and an article on case-based reasoning. II. “Perception, Understanding, and Action” with three articles on vision, speech and robotics. III. “Fuzzy Systems” with two articles – an overview and a soft computing article by Professor Zadeh. IV. “Models of the Brain and Evolution” with four articles - two articles by Rumelhart, Widrow, et al., an article on neural networks in Japan, and an article on genetic algorithms. There were many behind-the-scene stories and one of them was that my original plan was to have only one article by Rumelhart, et al. After much delay, they had an article twice as long as originally planned, and I suggested splitting the manuscript into two parts. In the Guest Editors Introduction, I wrote: The practical application of artiﬁcial intelligence (AI) has been the center of controversy for many years. Certainly, if we mean AI to be a realization of real human intelligence in the machine, its current state may be considered primitive. In this sense, the name artiﬁcial ”intelligence” can be misleading. However, when AI is looked at as ”advanced computing,” it can be seen as much more. In the past few years, the repertory of AI techniques has evolved and expanded, and applications have been made in everyday commercial and industrial domains. AI applications today span the realms of manufacturing, consumer products, ﬁnance, management, and medicine. Implementation of the correct AI technique in an application is often a must to stay competitive. Truly proﬁtable AI techniques are even kept secret. Many of the statements I wrote here are still basically true today. This Special Issue turned out to be a big hit. The ACM printed 1,000 extra copies of this issue for back orders, but they sold out less than a month. A person from a Japanese company wanted to purchase a box of ﬁfty copies of this issue, but it was too late. The issue became one of the most cited CACM, for not only within

A Personal View on AI, Rough Set Theory and Professor Pawlak

249

computer science but also some unexpected places such as Scientiﬁc American, the Berkeley Law School, the Stanford Philosophy Department, etc. The US Air Force has issued interesting predictions in the past. Around 1947, the Air Force predicted likely technologies for the next 50 years. They included jet rather than propeller powered, and supersonic airplanes. They became the reality by 1997. In 1997, they issued predictions for the next 50 years, i.e., by 2047. Between these major predictions, they published future perspectives for shorter time ranges. My Special Issue was cited in a document within a report entitled “Air Force 2025” that describes 30-year predictions by the US Air Force [1]. When I was preparing my ﬁrst CACM Special Issue, I knew there were other AI areas that were not covered. As soon as the ﬁrst issue was nearly complete, I started working on a follow-up Special Issue entitled “New Horizons in Commercial and Industrial AI.” In the “Editorial Pointers” in the ﬁrst issue, Executive Editor Diane Crawford wrote: “He has assembled some of the foremost minds in AI to author and/or review the 11 articles presented here. If that weren’t enough, he’s already digging into another proposed issue for Communications to appear early next year, where he hopes to address new horizons and applications in other AI-related ﬁelds.” For the ﬁrst Special Issue I received many responses. One of them was a letter from Professor Herbert A. Simon of Carnegie-Mellon University, a prominent scholar in AI with a Turing Award and a Nobel Prize in economics. Basically, he stated: “The ﬁrst issue was well done, although if I were the Guest Editor I would have had less emphasis on neural networks and included an article on machine learning.” He suggested placing an article on machine learning in the second issue. I replied to him saying I had already planned that and asked him to write one, and subsequently he co-authored an article. I was lucky to be able to have close contact with Professor Simon. When IBM’s Deep Blue defeated the human chess champion Garry Kasparov in 1997, he and I co-authored a commentary article on the signiﬁcance of this event on AI [14]. He was a pleasant person to talk with. He was a fulltime professor and active until two weeks before his death in 2001 at age 84. For the second Special Issue I had planned to include certain topics from a very early stage. They included symbolic machine learning, natural language processing (e.g., machine translation) and logic programming. Also, I wanted to include articles addressing the commonsense problem, although I did not expect that this area would have many commercial or industrial applications in the near term. At a later stage of preparation of the second issue, I searched for additional areas appropriate for the issue, and found rough set theory. I was not familiar with this area, but from what I found I thought it was a promising technique, appropriate for the second issue. Perhaps it could complement other AI techniques. I contacted Professor Pawlak and asked him whether he was interested in contributing an article to such a CACM Special Issue. These were my ﬁrst encounters with rough set theory and Professor Pawlak. He kindly accepted my invitation and contributed an article co-authored with Professors Jerzy GrzymalaBusse, Roman Slowinski and Wojciech Ziarko [8]. This was my

250

T. Munakata

ﬁrst acquaintance with the rough set community. As said earlier, CACM has a large number of audience worldwide and its impact is high. I don’t know how much the appearance of this article has inﬂuenced the promotion of this theory, but I think at least it helped to introduce the term “rough sets” worldwide. Incidentally, when I studied practically successful symbolic machine learning techniques for the ﬁrst time, such as ID3, I was a bit disappointed. From the term “learning,” I expected some elements of human-like learning. For example, given a speciﬁc experience the machine would abstract it, generalize it and be able to use it for similar circumstances in the future. I did not see such human-like learning in ID3. Rather, it simply classiﬁes data based on entropy in information theory. The characteristics of the target data seemed to be too simple. Perhaps the term “learning” was misleading, and probably I expected too much on what we could do from this kind of data. Both ID3 and rough sets can learn from data, but probably ID3 had attracted more attention than rough sets in the scientiﬁc community, at least during the 80s and 90s. Why? One reason might be that ID3 appears to have been in the main stream in the machine learning community, and had received more support from its early introduction. Professor Simon was one of it’s supporters, and he was a great scientist as well as a good salesman to promote his beliefs. For example, he called a software system he developed a “general problem solver,” which implied, with a bit of exaggeration, the system would solve every problem on earth. He was also an optimist. In the late 1950s he predicted that a computer would defeat a human chess champion within 10 years. We waited 10 years, another 10 years, and so forth for the next 40 years. In contrast, Professor Pawlak was a humble and modest scientist and perhaps not particularly a good salesman. In my opinion, rough set theory was not as widely recognized in the AI and CS ﬁelds as it should have been. After my ﬁrst encounter with the rough set community through my CACM second special issue, I have been fortunate to be able to work in this ﬁeld together with these people. I attended several rough set related conferences after my ﬁrst encounter [4, 5, 6]. To promote rough sets, I could think of two among many possibilities. One was to have promotional articles in journals of large audience like the CACM. Another area was to have a rough set application with a high social impact. For the latter, rough control might be a good candidate, I thought. Fuzzy set theory became a hot topic after Hitachi successfully applied fuzzy logic to Sendai subway control. I tried to push rough control, and I was Chair of the rough control interest group. The basic idea of rough control is to employ rough sets to automatically generate input-to-output control rules [7, 9]. The idea was not particularly new, but breakthrough applications would place rough set theory in the spotlight. A long time rough set activist Professor T.Y. Lin ﬁnancially supported me for this endeavor. Although we have not observed a major breakthrough yet, I think possibilities are still there. In the communication with Professor Pawlak, he suggested presenting a co-authored conference paper [13]. When I published an AI book from Springer, I included a chapter for rough sets [10]. When I served as Guest Editor for third time for CACM Special Section on knowledge discovery [11], I asked Professor Ziarko to contribute an article.

A Personal View on AI, Rough Set Theory and Professor Pawlak

251

When a young colleague approached me to work on a data mining article, I suggested employing rough sets [2]. I am currently working on another article on a rough set application with a young assistant professor. Although we all saddened by the recent death of Professor Pawlak, I think he was fortunate to observe that his theory has been widely recognized in the scientiﬁc community worldwide. This was not necessarily the case for many great scholars in the past. During my sabbatical in the fall of 2002, I traveled to Poland, visiting Professors Slowinski, Skowron and Pawlak, and received a warm welcome. This was the last time I saw Professor Pawlak. What are the future prospects of rough sets? No one knows, but the following is my speculation. Despite it’s founder’s death, the community will grow – there will be more researchers worldwide and more theoretical and application developments. But, growth in the ﬁeld may level out eventually, unless we achieve major breakthroughs. As in the case of other machine learning techniques and AI in general, we don’t know what, when or if such breakthroughs may come. Targeting to extremely large volumes of data (e.g., terabytes) and/or massively parallel computing alone do not look very promising, as we have observed similar attempts such as the Cyc and the Connection Machine. For knowledge discovery techniques such as rough sets, there may be a limit when we deal only with decision tables. Perhaps we should also look at other formats of data as well as other types of data, for example, non-text, comprehensive types of information, such as symbolic, visual, audio, etc. Also, the use of huge background knowledge, in a manner similar to human thought, would be necessary and eﬀective. Human-computer interactions would also enhance the discovery processes. Other totally diﬀerent domains are non-silicon based new computing paradigms. I am currently working on my fourth Special Section for the Communications of the ACM as a guest editor on this subject [12]. These approaches may lead to a new dimension of information processing in a wide range of application domains including rough sets. As with other scientiﬁc developments in history, such as alchemy and the ﬁrst airplane, a breakthrough may come in a totally unexpected form.

References 1. Clarence E. Carter, et al. The Man In The Chair: Cornerstone Of Global Battlespace Dominance, Air Force 2025, 1996. 2. Brenda Mak and T. Munakata. ”Rule extraction from expert heuristics: A comparative study of rough sets with neural networks and ID3,” European Journal of Operational Research, 136(1), pp. 212-229, 2002. 3. Toshinori Munakata. ”Guest Editor’s Introduction,” in the Special Issue ”Commercial and Industrial AI Applications,” Communications of the ACM, 37(3), pp. 2325, 1994. 4. Toshinori Munakata. ”Commercial and Industrial AI: Where it is now, and where it will be,” an invited talk at the Third International Workshop on Rough Sets and Soft Computing (RSSC’94), San Jose, CA, Nov., 1012, pp. 5155, 1994.

252

T. Munakata

5. Toshinori Munakata. ”Rough Control: A Perspective,” Workshop on Rough Sets and Data Mining, the 1995 ACM Computer Science Conference, Nashville, TN, March 2, 1995. 6. Toshinori Munakata. ”Commercial and Industrial AI and a Future Perspective on Rough Sets,” in Soft Computing: the Third International Workshop on Rough Sets and Soft Computing (RSSC94), T.Y. Lin and A.M. Wildberger (Eds.), the Society of Computer Simulation, pp. 219222, 1995. 7. Toshinori Munakata. ”Rough Control: Basic Ideas and Applications,” Workshop on Rough Set Theory, the Second Annual Joint Conference on Information Sciences, Wrightsville Beach, NC, Sept. 28 Oct. 1, 1995. 8. Toshinori Munakata. ”Guest Editor’s Introduction,” in the Special Issue ”New Horizons for Commercial and Industrial AI,” Communications of the ACM, 38(11), pp. 28-31, 1995. 9. Toshinori Munakata. ”Rough Control: A Perspective,” in T.Y. Lin and N. Cercone (Eds.), Rough Sets and Data Mining: Analysis of Imprecise Data, Kluwer Academic, pp. 7788, 1997. 10. Toshinori Munakata. Fundamentals of the New Artiﬁcial Intelligence: Beyond Traditional Paradigms, Springer-Verlag, 1998, 2007. 11. Toshinori Munakata. ”Guest Editor’s Introduction,” in the Special Section “Knowledge Discovery,” Communications of the ACM, 42(11), pp. 26-29, 1999. 12. Toshinori Munakata. ”Guest Editor’s Introduction,” in the Special Section “Beyond Silicon: New Computing Paradigms,” Communications of the ACM, to appear in 2007. 13. Zdzislaw Pawlak and Toshinori Munakata. ”Rough Control: Application of Rough Set Theory to Control,” EUFIT ’96 the Fourth European Congress on Intelligent Techniques and Soft Computing, Aachen, Germany, Sept. 25, 1996, pp. 209-217. 14. Herbert A. Simon and Toshinori Munakata. ”Kasparov vs. Deep Blue: The Aftermath - AI Lessons,” Communications of the ACM, 40(8), pp. 23-25, 1997.

Formal Topology and Information Systems Piero Pagliani1 and Mihir K. Chakraborty2 1

Research Group on Knowledge and Communication Models Via Imperia, 6. 00161 Roma, Italy p.pagliani@agora.stm.it 2 Department of Pure Mathematics, University of Calcutta 35, Ballygunge Circular Road, Calcutta-700019, India mihirc99@vsnl.com

Abstract. Rough Set Theory may be considered as a formal interpretation of observation of phenomena. On one side we have objects and on the other side we have properties. This is what we call a Property System. Observing is then the act of perceiving and then interpreting the binary relation (of satisfaction) between the two sides. Of course, the set of properties can be given a particular structure. However, from a pure ”phenomenological” point of view, a structure is given by the satisfaction relation we observe. So it is a result and not a precondition. Phenomena, in general, do not give rise to topological systems but to pre-topological systems. In particular, ”interior” and ”closure” operators are not continuous with respect to joins, so that they can ”miss” information. To obtain continuous operators we have to lift the abstraction level of Property Systems by synthesizing relations between objects and properties into systems of relations between objects and objects. Such relations are based on the notion of a minimal amount of information that is carried by an item. This way we can also account for Attribute Systems, that is, systems in which we have attributes instead of properties and items are evaluated by means of attribute values. But in order to apply our mathematical machinery to Attribute Systems we have to transform them into Property Systems in an appropriate manner. Keywords: approximation spaces, formal topology, Galois adjunctions, rough sets, information quanta, information systems, pointless topology, pretopology.

1

Introduction

Rough Sets arise from information systems in which items are evaluated against a set of attributes or properties. In Computer Science properties are often interpreted as ”open subsets” of some topological space. M. Smyth pioneered this interpretation in 1983 when he observed that semi-decidable properties are analogous to open sets in a topological space (cf. [28]). This intuition was developed by distinguished scholars such as D. Scott who introduced Domain Theory and the so-called Scott Topology to study continuous approximating maps between structures of information called domains. J.F. Peters et al. (Eds.): Transactions on Rough Sets VI, LNCS 4374, pp. 253–297, 2007. c Springer-Verlag Berlin Heidelberg 2007

254

P. Pagliani and M.K. Chakraborty

This research is paralleled, in a sense, by logical studies such as Cohen’s forcing and Kripke models, where the notion of approximable sets of properties (or approximable information) is the core of the construction. W. Lawvere showed that these constructions can be synthesized into the notion of a topos as the abstract form of continuously variable sets. S. Vickers combined the logical and the Computer Science approach. In [33], a prominent role is played by topological systems where just the formal properties of open sets are considered, without mentioning points (pointless topology). Indeed, this approach originates in Stone’s and Birkhoﬀ’s representation theorems where the notion of an abstract point is de facto, introduced. And an abstract point is nothing else but a bunch of properties (once we interpret the elements of a lattice as properties). Inﬂuenced by P. Martin-L¨ of’s Intuitionistic Type Theory, G. Sambin undertook his own way to deal with pointless topology, and speciﬁcally pointless pretopology, as related to Logic (namely Linear Logic), which led to the notion of a Formal Topology (see [25]) which later on has been presented as a result of a construction arising from binary relations between the concrete side (points) and the abstract side (properties) of an observational system called a Basic Pair (cf. [26]). As a matter of fact, the interrelations between concrete points and abstract properties is considered by Vickers, too. However in Formal Topology one does not impose any pre-established structure on the set of properties, not even that suggested by ”Observation Logic” in Vicker’s approach, which makes a system of observations into a frame1 . In [18] it was noted that the properties of the operators of Formal Topology may be deduced from the fact that they are based on constructors which enjoy adjointness relations. The pretopological approach was applied to account for approximation operators arising from families of Property Systems in [19] and, ﬁnally, to generalize approximation operators arising from single Property Systems ([21] and [22]). Moreover, this machinery was applied to Attribute Systems too, by transforming them into Property Systems (cf. [20]). The latter researches were, in turn, inﬂuenced by A. Skowron, J. Stepaniuk and T. Y. Lin’s pioneering investigations which have shown that neighborhood systems may account for natural organizations of information systems (for this topic and its applications the reader is referred to [11] and [29]). Moreover, it must be noticed that neighborhood systems give rise to pretopological operators which turn into topological operators just under particular conditions. Therefore, we claim that pre-topology is a most natural setting for approximation operators as induced by information systems. Also, this claim ﬁts with recent suggestions on the role of ”true borders” (hence non topological) to account for a more dynamic approach to data analysis and granular computing (see, for instance [19] and for ”rough topologies” see [24]). Speciﬁcally, the present investigation is induced by the observation that, from a very general point of view, Rough Set Theory arises from a sort of ”phenomenological” approach to data analysis with two peculiar characteristics: 1

A frame is a lattice with ﬁnite meets distributing over arbitrary joins.

Formal Topology and Information Systems

255

– data is analyzed statically at a given point in time of a possibly evolving observation activity; – as a consequence, the analyzed data provides us only with an approximated picture of the domain of interest. We shall see that the status of an observation system at a certain point in time is essentially a triple P = G, M, , that we call a Property system were G is a set of objects (also called ”points”), M a set of properties (also called ”formal neighborhoods”) and ⊆ G × M is intended as a fulﬁllment relation2 . From the concept of an ”observation” we shall deﬁne a family of basic ”perception constructors” mapping sets of objects into sets of properties, called intensional constructors, and sets of properties into sets of objects, called extensional constructors. We show that some pairs of constructors from opposite sides, fulﬁll adjunction properties. That is, one behaves in a particular way with respect to properties if and only if the other behaves in a mirror way with respect to objects. Hence, adjunction properties state a sort of ”dialectic” relationship, or mutual relationship, between perception constructors, which is exactly what we require in view of a ”phenomenological” approach. Adjunction properties make some combinations of these basic constructors into generalized closure and generalized interior operators. Particularly, some combinations happen to be pre-topological operators in the sense of Sambin’s Formal Topology. Actually, we shall see that they are generalizations of the approximation operators provided by Rough Set Theory. However, for they are pretopological and not topological, these approximation operators are not continuous, that is, they exhibit ”jumps” in the presence of set-theoretical operations. Therefore we synthesize the structuring properties of a Property system, P, into a second level informational structure G, G, RP , called an Information Quantum Relational System - IQRS, where RP is a relation between objects - hence no longer between objects and properties - embedding the relevant informational patterns of P. In IQRSs adjointness makes second level approximation operators fulﬁll nice properties. Also, this way we shall be able to account for Attribute systems after appropriately transforming them into Property Systems3 . This study aims at presenting the state-of-the-art of a conception of Rough Set Theory as a convergence of diﬀerent approaches and diﬀerent techniques, such as Formal Topology, duality and representation theory, Quantum Logic, adjoint functors and so on, as sketched in the following ﬁgure: 2

3

Property Systems may be also regarded as ”Chu Spaces” ([37]). However Chu Spaces have additional features, namely a pair of covariant and contravariant functors, which links spaces together. For Chu Spaces, see the WWW site edited by V. Pratt, http://boole.stanford.edu/chuguide.html. The term ”Quantum” refers to the fact that a basic information grains is given by the minimal amount of information which is organised by RP around an item what, technically, is linked to the notion of a quantum of information at a location once one substitute ”item” for ”location” - cf. [4].

256

P. Pagliani and M.K. Chakraborty

Fig. 1. A Rough Set connection

More precisely, in the present paper we shall deal with the boxed topics along the ﬂow displayed by bold lines. Dotted arrows display the relationships between logical systems and some topics connected to Rough Sets Theory. Dotted lines show some interesting links between some techniques used in the present paper and topics connected with data and information analysis. Bold arrows display some well-established links between logico-algebraic systems and Rough Set systems, while the dotted arrow marked with ”?” suggests links between the modal-style approach applied in the paper and logico-algebraic interpretations to be explored4 .

2

Formal Relationships Between ”Objects” and ”Observables”

Observation is a dynamic process aimed at getting more and more information about a domain. The larger the information, the ﬁner the picture that we have about the elements of the domain. Using topological terms, an observation process makes it possible to move from a trivial topology on the domain, in 4

We have proved that Rough Set Systems are semi-simple Nelson algebras (or equivalently, three-valued L ukasiewicz algebras) (cf. [14] and [15]). What algebraic systems arise from this generalisation has to be studied yet. Brouwer-Zadeh Lattices, Bilattices and Semi-Post algebras may provide some hints.

Formal Topology and Information Systems

257

which everything is indistinguishable to a topology in which any single element is sharply separable from all the other elements (say a discrete topology or a Hausdorﬀ space). In this case we can ”name” each single element of the domain by means of its characteristic properties. However, in this respect, observation is an asymptotic process. What usually happens is that at a certain point in time we stop our observation process, at least temporarily, and analyse the stock of pieces of information we have collected so far. In a sense we consider a ”ﬂat slice” of the observation process.

Fig. 2. A process of diﬀerentiation via observations

Therefore, our slice is basically composed by: (a) a set G of ’objects’; (b) a set M of ’observable properties’; (c) a relation between G and M , denoted with the symbol . Given g ∈ G and m ∈ M we shall say that if g m, then g enjoys property m, or that g induces the observable property m. We have also to assume that is deﬁned for all the elements of G and M because we consider immaterial any property which cannot help making any distinction among objects and, symmetrically, if an object g does not manifest any property, then it is a ”non-object” from a phenomenological point of view5 . We encode everything in the following deﬁnition: Deﬁnition 1. A triple G, M, where G and M are ﬁnite sets, ⊆ G × M is a relation such that for all g ∈ G there is m ∈ M such that g m, and for all m ∈ M there is g ∈ G such that g m, is called a property system or a P-system. Among P-systems we distinguish: a) Functional systems, or FP-systems, where is functional in the sense that for any element g ∈ G, g m and g m implies m = m . 5

The symbols ”G” and ”M” are after the German terms ”Gegenst¨ ande” (”objects”) and, respectively, ”Merkmale (”properties”). A ”Gegenstand” is what stays in front of a subject, while the German term ”Object” means an interpreted ”Gegenstand”. These are the terms used in Formal Concept Analysis and we adopt them for their philosophical meaning.

258

P. Pagliani and M.K. Chakraborty

b) Dichotomic systems or DP-systems, if for all p ∈ M there is p ∈ M such that for all g ∈ G, g p if and only if g p. Functional and dichotomic systems enjoy particular classiﬁcation properties6 . Moreover we shall also consider Deterministic attribute systems: Deﬁnition 2. A structure of the form G, At, {Va }a∈At , , where G, At and Va are sets (of objects, attributes and, respectively, attribute-values) and for each a ∈ At, a : G −→ Ata , is called a deterministic Attribute System or an Asystem7 . From now on we assume that P always denotes a Property System G, M, and that A denotes an Attribute System G, At, {Va }a∈At , . Moreover, we shall use the following notation: If f : A −→ B and g : B −→ C are functions, then: (a) with (f ◦g)(x) or, equivalently, g(f (x)) we denote the composition of g after f ; (b) f → : ℘(A) −→ ℘(B); f → (X) = {f (a) : a ∈ X} - denotes the image of X via f ; (c) f ← : ℘(B) −→ ℘(A); f ← (Y ) = {a : f (a) ∈ Y } - denotes the pre-image of Y via f ; (d) the set f → (A) is denoted with Imf ; 1A , denotes the identity function on A; (e) the map f o : A −→ Imf ; f o (a) = f (a) denotes the corestriction of f to Imf and the map fo : Imf −→ B; fo (b) = b denotes the inclusion of Imf into B; (f) the equivalence relation kf = {a, a : f (a) = f (a )} is called the kernel of f or the ﬁbred product A ×B A obtained by pulling back f along itself. 2.1

Ideal Observation Situations

If G, M, is an FP-system, we are in a privileged position for classifying objects, for the reasons we are going to explain. The ”best” case is when is an injective function. Indeed in this case the converse relation (or, also, −1 ) is a function, too, and we are able to distinguish, sharply, each object. In mathematical words we can compute the retraction of . Deﬁnition 3. Let f : A −→ B be a function. Then a morphism r : B −→ A is called a retraction of f , if f ◦ r = 1A . But, ﬁrst of all, this is an unusual situation, from a practical point of view. Further, ”observation” and ”interpretation” is a modeling activity. Thus, from an epistemological point of view we may wonder if the best model of a horse is really a horse. Actually, ”modeling” means ”abstracting” and ”abstracting” means ”losing something”, some quality or characteristic. Thus the situation depicted in Figures 3 and 4 cannot be but the result of some reduction process. 6 7

Indeed FP-systems and DP-systems are closely linked together, as we shall see. The traditional term in Rough Set Theory is ”Information System”.

Formal Topology and Information Systems

259

Fig. 3. An ideal situation: the existence of retractions

Fig. 4. Is a horse the best model of a horse?

A second optimal situation is when is a surjective function (what always happens of FP-systems). Indeed, by reading back we obtain an equivalence relation E , so that any element of G will belong to one and just one equivalence class modulo E , without ambiguity and borderline situations, what is a perfect case of a classiﬁcation. Indeed, E is the kernel of and it induces a classiﬁcation of the elements of G through properties in M . This is tantamount to the construction of stalk spaces, or espace etal´e, through ﬁbers (or stalks, sorts). This means that has a section or a co-retraction. Deﬁnition 4. Let f : A −→ B be a function. Then a morphism s : B −→ A is called a section or co-retraction of f , if s ◦ f = 1B . We can interpret f as a way to list or parametrise (some of) the elements of B, through the elements of A. In turn the notions of a section and a retraction are special cases of a more fundamental concept: a divisor.

260

P. Pagliani and M.K. Chakraborty

Deﬁnition 5. Let f : A −→ B, g : B −→ C and h : A −→ C be three functions. Then, g is called a right divisor of h by f and f is called a left divisor of h by g if h = f ◦ g, that is, if the following diagram commutes: B f

A

@ @ g @ @ @ R - C

h From the above deﬁnitions, we immediately deduce that if s is a section of f , then f is a retraction of s; and vice-versa. Moreover, it is not diﬃcult to verify that f does not have any section if it is not surjective on B (otherwise, how would it be possible to obtain 1B ?). Intuitively if there is a b ∈ B that is not f -image of any a ∈ A, then b would be associated with a void sort. Vice-versa, a function f does not have any retraction if f is not injective in B. In fact, if f (a) = f (a ) = b, for a = a , then any morphism from B to A either maps b onto a and forgets a , or it maps b onto a and forgets a, because of unicity of the image, and we could not obtain 1A (thus, for any function f : A −→ A, fo is a section with retraction f o ).

Fig. 5. An almost ideal situation: the existence of co-retractions

If r : A −→ B is a retraction of a function h : B −→ A, then r◦h is an idempotent endomorphism of A: (r ◦ h) ◦ (r ◦ h) = r ◦ (h ◦ r) ◦ h = r ◦ 1B ◦ h = r ◦ h. It follows that if s : B −→ A is a section of f : A −→ B, then f ◦ s is an idempotent endomorphism in A, provided s is onto, because f is a retraction of s (see above). Clearly, if a = s(b), then s(f (a)) = s(f (s(b))) = s(1B (b)) = s(b) = a. Hence, any image of the section s is a ﬁxed point of the endomorphism f ◦ s.

Formal Topology and Information Systems

261

Thus sections, retractions and kernels of a function f make it possible to organise the domain of f into sharp classiﬁcation categories (groups) and to single out representatives of such categories.

3

Categorizing Through Relational P-systems

On the contrary, if we deal with generic relational P-systems, not necessarily functional, it hardly happens to obtain sharp classiﬁcations, at least without a proper manipulation of the given P-system that, in turn, may be or may not be an appropriate maneuver. It follows that the identity relation in the deﬁnition of left and right divisors must be weakened to an inequality relation ”≥” or ”≤”. Therefore, to deal with generic cases we need a more subtle mathematical machinery. Such a machinery is based on the notion of an ”approximation”. However, this notion depends on another one. Indeed, we cannot speak of ”approximation” without comparing a result with a goal and this comparison depends on the granularity of the target and of the instruments to get it. For instance, in a document search system, in general we face a situation in which queries refer to a set of possible answers and not to single objects. Otherwise we would not have ”queries” but ”selections” (the realm of sections and retractions). In other words, objects constitute, in principle, a ﬁner domain than those obtained by any modeling or interpretation activity. So we can distinguish an extensional granulation, related to objects, and an intensional granulation, related to properties, and assume that the extensional granulation is ﬁner than the intensional one. Thus, when we have to determine a point on the extensional scale by means of the intensional ruler, we hardly will be able to get a precise determination. We can approximate it. But in order to be able to have ”best approximations” the intensional granulation and its relationships with the extensional granulation must fulﬁll non trivial properties. First of all we need an order. Suppose X is a set of candidate results of a query. Then we generally do not have a selection criterion to single out elements of X. But if the elements of X are ordered in some way, we can use this order to choose, for instance, the least or the largest element in X, if any.

Fig. 6. A usual P-system needs a scale

262

P. Pagliani and M.K. Chakraborty

But this is not enough. In fact dealing with functions (either surjective or injective) is, in a sense, a lucky event in this business which happens only if we are able to completely reduce a structure to a simpler one. This means that generally we cannot have sections and retractions, so that we cannot directly manipulate pre-images of single elements of the codomain or images of single elements of the domain of some ”connecting” function. On the contrary we have to manipulate some kinds of subsets of the domain and co-domain which, we hope, embed enough ordering features to compute approximations. Having this picture in mind, we underline what follows. From an observational point of view the only relationships between objects are induced by the fulﬁllment relation and they are grouping relationships so that we can compare subsets of objects (or properties) but not, directly, objects (or properties). In other words in this paper we assume that there is no relation (hence any order) either between objects or between properties. Hence the result of an approximation activity is, generally, a ”type” not a ”token”8. It follows that we shall move from the level of pure P-systems G, M, to that of Perception systems ℘(G), ℘(M ), {φi }i∈I where φi is a map from ℘(G) to ℘(M ) or from ℘(M ) to ℘(G).

Fig. 7. Approximation deals with types, not with tokens

4

Concrete and Formal Observation Spaces

Given a P-system, the ﬁrst, and obliged, step we have to do is ”observing”, in the elementary sense of ”perceiving” the manifested properties. Thus if P is a P-system let us deﬁne an ’observation function’ obs : G → ℘(M ), by setting m ∈ obs(g) ⇔ g m.

(1)

Technically, obs is what is called a constructor for it builds-up a set from a point. Indeed, for each point g, obs(g) = {m ∈ M : g m}. We shall call obs(g) the ’intension of g’. In fact, any element g appears through the series of its observable properties, so that obs(g) is actually the intensional description of g. The intension of a point g is, therefore, its description through the observable properties listed in M . We shall also say that if g m (i. e. if m ∈ obs(g)), then m is an observable property connected with g and that g belongs to the ﬁeld of m. 8

By the way, note that in [3], classiﬁcation is achieved at ”type” level.

Formal Topology and Information Systems

263

Symmetrically we can introduce a ”substance function” sub : M → ℘(G) deﬁned by setting g ∈ sub(m) ⇔ g m. (2) This symmetry reﬂects the intuition that a point can be intensionally conceived as the set of properties it is connected with, just as a property may be extensionally conceived as the set of points belonging to its ﬁeld. Dually to obs, given a property m ∈ M , sub(m) = {g ∈ G : g m}, so that sub(m) is the ’extension’, or the ﬁeld, of m. The link between these two functions is the relation : g ∈ sub(m) ⇔ m ∈ obs(g) ⇔ g m, ∀g ∈ G, ∀m ∈ M

(3)

Fig. 8. A ﬁrst level perception process

We now notice that since the set M is given and ﬁxed, any P − system will provide only partial observations of the members of G so that a single point x possibly fails to be uniquely described by its intension obs(x). We shall also say that obs(x) is an intensional approximation of a ’partially describable’ member x of G and claim that if obs(x) = obs(y), then x and y cannot be discerned by means of the observable approximating properties (or ”partial descriptions”) at hand, so that x and y will be said to be indiscernible in the given P-system. If x and y are indiscernible they will collapse into the same intentional description. Indeed, if obs fails to be injective then we know that it cannot have a retraction and this means that the identity 1℘(G) cannot be determined by means of the properties at our disposal (that is, the subsets of M mapped by obs), so that a ”loss of identity” literally happens. However, we can deﬁne, by means of obs and sub some approximation operators. 4.1

The Basic Perception Constructors

The second step after observing, is an initial interpretation of what we have observed. Thus we shall introduce the ”perception constructors” that are induced by a P-system.

264

P. Pagliani and M.K. Chakraborty

These constructors will make it possible to deﬁne diﬀerent kinds of structures over ℘(G) and ℘(M ). Since such structures are deﬁned as extensions of the two functions obs and sub and since, in turn, these two functions are linked by the relation (3), it is clear that any structurization on points will have a dual structurization on observables, and vice-versa. Deﬁnition 6 (Basic contructors). Let P = G, M, be a P-system. Then: – – – – – –

e : ℘(M ) −→ ℘(G); e(Y ) = {g ∈ G : ∃m(m ∈ Y & g ∈ sub(m))}; [e] : ℘(M ) −→ ℘(G); [e](Y ) = {g ∈ G : ∀m(g ∈ sub(m) =⇒ m ∈ Y )}; [[e]] : ℘(M ) −→ ℘(G); [[e]](Y ) = {g ∈ G : ∀m(m ∈ Y =⇒ g ∈ sub(m))}; i : ℘(G) −→ ℘(M ); i(X) = {m ∈ M : ∃g(g ∈ X & m ∈ obs(g))} [i] : ℘(G) −→ ℘(M ); [i](X) = {m ∈ M : ∀g(m ∈ obs(g) =⇒ g ∈ X)}; [[i]] : ℘(G) −→ ℘(M ); [[i]](X) = {m ∈ M : ∀g(g ∈ X =⇒ m ∈ obs(g))}.

Fig. 9. Basic constructors derived from a basic pair

An intuitive interpretation of the above functions is in order. As for the constructors decorated with ’e’ (because the result of the function is an extent), we notice that if we want to extend function sub from elements to subsets of M , we have essentially two choices: a ”disjunctive” or ”existential” extension and a ”conjunctive” or ”universal” extension. The former is e while the latter is [[e]]. Obviously, e = sub→ = obs← . It is not diﬃcult to see that [e] is the dual of e, hence it is the ”co-existential extension” of sub (the dual of [[e]] is not discussed in this paper). Given Y ⊆ B, the set [[e]](Y ) collects the points that fulﬁll at least all the properties from Y (and, possibly, others), while e(Y ) gives the set of points which fulﬁll at least one property from Y . Finally, [e](Y ) collects the points which fulﬁll at most all the properties from Y (but possibly not all the properties in Y ). The same considerations apply symmetrically to the operators decorated with ’i’ (because the result of these functions is an intent). Indeed i and [[i]] are the disjunctive and, respectively, conjunctive extensions to subsets of G of the function obs and i = obs→ = sub← . More precisely, i(X) collects the set of properties that are fulﬁlled at least by one point of X, while [[i]](X) collects the set of properties that are fulﬁlled at least by all the points of X, that is, the properties com-

Formal Topology and Information Systems

265

mon at least to all the points of X. Finally, [i](X), the ”co-existential extension” of obs, gives the set of properties that are fulﬁlled at most by all of the points in X. In particular, [i]({x}) is the set of properties which uniquely characterize x. We summarize these remarks in the following table: ... property/ies in Y ... point/s in X at least one ... e(Y ) i(X) at least all ... [[e]](Y ) [[i]](X) at most all ... [e](Y ) [i](X) If one of the above or following operators on sets, say Op, is applied to a singleton {x} we shall also write Op(x) instead of the correct Op({x}), if there is no risk of confusion. 4.2

A Modal Reading of the Basic Constructors

At this point a modal reading of the basic constructors is in order. This will be formalized in Section 6. Indeed, we can read these constructors by means of operators taken from extended forms of modal logic, namely, possibility, necessity and suﬃciency. x∈X b∈B if x ∈ X then b is enjoyed by i(X) = B it is possible that some element x enjoys elements in B collected in X Operator

[i](X) = B

to enjoy elements in B b is enjoyed by it is necessary at most all the to be in X element of X

[[i]](X) = B

to enjoy elements in B b is enjoyed by it is suﬃcient at least all the to be in X elements of X

Example reading there are examples of elements in X that enjoy b there are not examples of elements enjoying b that are not in X there are not examples of elements of X that do not enjoy b

Remarks: Suﬃciency was introduced in modal logic by [10]. Recently it was discussed in [8] and in [16] from an informational point of view. Suﬃciency happens to be the fundamental operator to deﬁne Formal Concepts, which are pairs of the form [[e]][[i]](X), [[i]](X), for X ⊆ G (see [34]). From the point of view of pointless topology, the operators [i], i, [e] and e have been studied by [26], (where the notation ””, ”♦”, ”rest” and, respectively, ”ext” is used) and in [19]. From an informational point of view they have been investigated in [8], [7], [36], [21] and [22]. It is worth noticing that variations of concept lattices have been introduced by means of these operators, namely ”object oriented concepts” of the form (e[i](X), [i](X) (by Y. Y. Yao) and ”property oriented concepts” of the form [e]i(X), i(X) (by D¨ untsch and Gegida). From Proposition 2 and Corollary 3 below one can easily deduce some of the properties discussed in the quoted papers. A pictorial description of the above modal reading follows:

266

P. Pagliani and M.K. Chakraborty

Fig. 10. Possibility, Necessity, Suﬃciency

The above functions are linked by some structural relationships. First, recall that our operators are deﬁned on Boolean algebras of type ℘(A), ∩, ∪, A, −, ∅, where A is either G or M , so that negation coincides with settheoretical complementation. We say that if an operator Opo is obtained by negating all of the deﬁning subformulas in an operator Op, and by further applying the contraposition law according to negations (or, equivalently, by ﬁrst putting the deﬁnition in disjunctive normal form), then Opo and Op are called opposite or orthogonal (to each other), or ”o” in symbols9 . If Opd (X) =∼ Op(∼ X) then Opd is called the dual of Op and we denote the relation of duality with ”d”. Furthermore we can easily observe that functions decorated with e and functions decorated with i are symmetric with respect to the relation , and we denote this fact with ”s”. The following table summarizes these relationships between basic operators (some of these connections are well known in literature: cf. [26], [7] and [8] - but see also the literature about Galois connections): e i [e] [i] [[e]] [[i]] e = s d sd od ods i s = sd d ods od [e] d sd = s o os [i] sd d s = os o [[e]] od ods o os = s [[i]] ods od os o s = Obviously, symmetric functions fulﬁll the same formal properties, opposite functions fulﬁll opposite properties, while dual and symmetric-dual operators fulﬁll dual properties.

5

Fundamental Properties of the Basic Constructors

Let us investigate the fundamental properties of basic constructors. We carry on this job in a more general dimension concerning binary relations at large. 9

So, for instance, if α =⇒ β appears in a deﬁning formula, of Op, then in Opo we have ∼ β =⇒∼ α.

Formal Topology and Information Systems

267

Deﬁnition 7. Let R ⊆ A × B and Q ⊆ C × D be binary relations, X ⊆ A, Y ⊆ B, x ∈ A, y ∈ B. Then we deﬁne10 : 1. R = {y, x : x, y ∈ R} - the inverse relation of R. 2. R (X) = {y ∈ B : ∃x ∈ X(x, y ∈ R)} - the left Peirce product of R and X. We shall also call R (X) the R- neighborhood of X. In particular, if X is a singleton {x}, then we shall usually write R (x) instead of R ({x}). 3. R(Y ) = {x ∈ A : ∃y ∈ Y (x, y ∈ R)} - the left Pierce product of R and Y . Clearly, R(Y ) is the R -neighborhood of Y . 4. [R ](X) = {y ∈ B : ∀x(x, y ∈ R =⇒ x ∈ X)} - the right residual of R and X. 5. [R](Y ) = {x ∈ A : ∀y(x, y ∈ R =⇒ y ∈ Y )} - the right residual of R and X. 6. [[R ]](X) = {y ∈ B : ∀x(x ∈ X =⇒ x, y ∈ R)} - the left residual of X and R . 7. [[R]](Y ) = {x ∈ A : ∀y(y ∈ Y =⇒ x, y ∈ R)} - the left residual of X and R. 8. R⊗Q = {a, d : ∃z ∈ B∩C(a, z ∈ R & z, d ∈ Q)} - the right composition of R with Q or the left composition of Q with R. If deﬁned, R ⊗ Q ⊆ A × D. Lemma 1. Let R ⊆ A × B. Then for any X ⊆ A, Y ⊆ B, a ∈ A, b ∈ B: 1. (a) b ∈ [R ] (X) iﬀ R(b) ⊆ X; (b) a ∈ [R] (Y ) iﬀ R (a) ⊆ Y ; 2. (a) a ∈ [[R]] (Y ) iﬀ Y ⊆ R (a); (b) b ∈ [[R ]] (X) iﬀ X ⊆ R(b); 3. (a) [[R]](∅) = A, (b) [[R ]](∅) = B, (c) R (∅) = R(∅) = ∅, (d) [R ](A) = B. (e) If R is onto then R (A) = B and [R ](∅) = ∅; (f) [R](B) = A. (g) If R is onto then R(B) = A and [R](∅) = ∅. 4. If X and Y are singletons, then (a) R (X) = [[R ]] (X); (b) R(Y ) = [[R]] (Y ); 5. (a) If R is onto, [R ](X) ⊆ R (X); (b) If R is onto [R](Y ) ⊆ R(Y ); 6. If R is a functional relation then [R](Y ) = R(Y ); 7. If R is a functional relation then [R ](X) = R (X). Proof. (1) (a) By deﬁnition b ∈ [R ] (X) iﬀ ∀a(a, b ∈ R =⇒ a ∈ X) iﬀ R(b) ⊆ X.(b) By symmetry. (2) from (1) by swapping the position of the 10

As the reader will probably note, the operations we denote with R (X) and R(Y ) are often denoted with R(X) and, respectively, R (Y ). Moreover, the left composition of R with Q is usually denoted with R; Q in mathematical literature. However, there are several reasons which suggest to adopt the following symbols, mostly depending on both logic and relational algebra. In particular logical reasons are related to Kripke models for Modal Logics and, as to the left composition, to Linear Logic. Apart from symbols, R(Y ) coincides with what in [26] is called the ”extension” of Y along R and [R](Y ) the ”restriction” of Y along R . In Formal Concept Analysis [[R]](Y ) is the ”(derived) extent” of Y , while [[R ]](X) is called the ”(derived) intent” of X. The terminology used here is that of Relation Algebra and connected topics, (strictly speaking, residuals are deﬁned provided X and Y are right ideal elements - see for instance [16]).

268

P. Pagliani and M.K. Chakraborty

relations ∈ and R. (3) (a), (b) and (c) are obvious. (d) For any b ∈ B, either a, b ∈ R for some a ∈ A or the premise of the implication deﬁning the operator [R ] is false. (e) If R is surjective then for all a ∈ A there is a b ∈ B such that a, b ∈ R. Moreover, in [R ](∅) the consequence is always false. Similar proofs for (f) and (g). (4) Applied on singletons the deﬁnitions of [[α]] and α operators trivially coincide, for α = R or α = R . (5) For all b ∈ B, b ∈ [R ](X) iﬀ R(b) ⊆ X iﬀ (for isotonicity of R ) R (R(b)) ⊆ R (X). But b ∈ R (R(b)). Hence b ∈ R(X). Symmetrically for [R] and R. (6) If R is a functional relation, by deﬁnition R is onto, thus from point (5) [R](Y ) ⊆ R(Y ) for any Y ⊆ B. Suppose x ∈ R(Y ) and x ∈ / [R](Y ). Then there is y ∈ Y such that x, y ∈ R and there is a y ∈ / Y such that x, y ∈ R and x, y ∈ R. Hence R is not functional. (7) It is an instance of (6). 5.1

Solving the Divisor Inequalities

Now we come back for a while to the divisor diagram of Deﬁnition 5. Our instance of this diagram reads as in Figure 3. Thus we have to understand under what conditions we can have ”best approximating” maps. Therefore, suppose in general φ is a function which maps subsets of a set A into subsets of a set B (possibly in dependence on how the elements of A are related via a binary relation R ⊆ A × B with the members of B). If φ(X) ⊇ Y we can say that X approximates Y from above via φ. The smallest of these X can therefore be thought of as a ”best approximation from above” via φ, for its image is the closest to Y . In order to get such a best approximation, if any, we should take φ← (↑ Y ), where ↑ Y = {Y ⊆ B: Y ⊇ Y }. In fact φ← (↑ Y ) = {X : φ(X) ⊇ Y }. Dually, if we take φ← (↓ Y ), where ↓ Y = {Y ⊆ B : Y ⊆ Y }, we should obtaina ”best approximation from below” of Y , if any, via φ, because φ←1 (↓ Y ) = {X : φ(X) ⊆ Y }. To be sure, this approach is successful if φ( φ← (↑ Y )) ⊇ Y and, dually, φ( φ← (↓ Y )) ⊆ Y . So we now shall examine, in an abstract setting, the conditions under which the above operations are admissible and behave as expected. Indeed, we have a mathematical result which states rigorously these informal intuitions11 . 11

We remind that in a preordered set O: (a) ↑ X = {y : ∃x(x ∈ X & x y)} = (X) is called the order ﬁlter generated by X. In particular ∀p ∈ A, ↑ p =↑ {p} is called the principal order ﬁlter generated by p. If O is partially ordered p = min(↑ p), where, given a set X, min(X) is the minimum element of X. (b) ↓ X = {y : ∃x(x ∈ X & y x)} = (X) is called the order ideal generated by X. In particular, ∀p ∈ A, ↓ p =↓ {p} is called the principal order ideal generated by p. If O is partially ordered p = max(↓ p), where, given a set X, max(X) is the maximum element of X. From now on O = A, ≤ and O = A , ≤ will denote preordered or partial ordered sets. Furthermore, with L and L we shall denote two arbitrary complete and bounded lattices L = L, ∨, ∧, 0, 1 and, respectively, L = L , ∨ , ∧ , 0 , 1 .

Formal Topology and Information Systems

269

Proposition 1. Let A and B be partially ordered sets, φ a functor (i. e. an isotone map) between A and B. Then, the following conditions are equivalent: 1. (a) there exists a functor ψ : B −→ A such that φ◦ψ ≥A 1A and ψ◦φ ≤B 1B ; (a’) for all b ∈ B, φ← (↓ b) is a principal order ideal of A. 2. (b) there exists a functor ϑ : A −→ B such that ϑ◦φ ≥B 1B and φ◦ϑ ≤A 1A ; (b’) for all a ∈ A, ϑ← (↑ a) is a principal ﬁlter of B. The proof can be found in [5]. 5.2

Galois Adjunctions and Galois Connections

The conditions stated in Proposition 1 deﬁne a basic mathematical notion which is at the core of our construction (NOTE: the following materials are known and we have included them to render completeness to this paper). Deﬁnition 8. Let σ : O −→ O and ι : O −→ O be two maps between partial ordered sets. Then we say that ι and σ fulﬁll an adjointness relation if the following holds: ∀p ∈ O, ∀p ∈ O , ι(p ) ≤ p if and only if p ≤ σ(p)

(4)

If the above conditions hold, then σ is called the upper adjoint of ι and ι is called the lower adjoint of σ. This fact is denoted by O ι,σ O

(5)

and we shall say that the two maps form an adjunction between O and O . If the two preorders are understood we shall denote it with ι σ, too12 . When an adjointness relationship holds between two preordered structures we say that the pair σ, ι forms a Galois adjunction or an axiality. This name is after the notion of a Galois connection which is deﬁned by means of a similar but covariant condition where, indeed, ι and σ are antitone: ∀p ∈ O, ∀p ∈ O , ι(p) ≥ p if and only if p ≤ σ(p )

(6)

We read this fact by saying that the pair σ, ι forms a Galois connection or a polarity. Clearly, a Galois connection is a Galois adjoint with the right category O turned into its opposite Oop . In other words, σ, ι is a polarity if and only if O ι,σ Oop . 12

Sometimes in mathematical literature, the lower adjoint is called ”left adjoint” and the upper adjoint is called ”right adjoint”. However, the reader must take care of the fact that we have two levels of duality. The ﬁrst swaps the partial order (≤ into ≥ and vice-versa). The second swaps the order of application of the functors (ι ◦ σ into σ ◦ ι, and the other way around) and the position of the two structures (by the way, we notice that in usual literature given a map φ, the upper residual is denoted with φ∗ and the lower residual is denoted with φ∗ ).

270

P. Pagliani and M.K. Chakraborty

We now state without proof a number of properties fulﬁlled by adjoint maps. Proposition 2. Let σ : O −→ O and ι : O −→ O be mappings, p ∈ O and p ∈ O . Then, (a) the following statements are equivalent: (a.1) O ι,σ O; (a.2) σι(p ) ≥ p and ισ(p) ≤ p, and both ι and σ are isotone; (a.3) σ is isotone and ι(p ) = min(σ ← (↑ p )); (a.4) ι is isotone and σ(p) = max(ι← (↓ p)); If O ι,σ O, then: (b) σ preserves all the existing infs and ι preserves all the existing sups; (c) ι = ισι, σ = σισ; (d) σι and ισ are idempotent. (e) σ is surjective iﬀ ι(p ) = min(σ ← ({p })) iﬀ σι(p ) = p iﬀ ι is injective; (f) σ is injective iﬀ σ(p) = max(ι← ({p})) iﬀ ισ(p) = p iﬀ ι is surjective. Notice that (e) and (f) are the reader’s digest of the story about retraction and coretraction we told in Section 2.1. Now we have a good stock of results in order to ”implement” a suﬃciently large body of useful operators, actually those operators which will constitute the backbone of all the present story. Deﬁnition 9. Let φ : O → O be an operator on a partially ordered set and ϑ : L → L be an operator between two lattices. Then, (1) φ is a projection operator on O iﬀ it is isotone and idempotent; (2) a projection operator is a closure operator iﬀ it is increasing; (3) a projection operator is an interior operator iﬀ it is decreasing; (4) ϑ is a modal operator iﬀ it is normal (i. e. ϑ(0) = 0 ) and additive; (5) a closure operator ϑ on a lattice is topological iﬀ it is modal; (6) ϑ is a co-modal operator iﬀ it is co-normal (i. e. ϑ(1) = 1 ) and multiplicative; (7) an interior operator ϑ on a lattice is topological iﬀ it is co-modal; (8) ϑ is an anti-modal operator iﬀ it is anti-normal (i. e. ϑ(0) = 1 ) and antiadditive (i. e. ϑ(x ∨ y) = ϑ(x) ∧ ϑ(y)). Notice that in our deﬁnition of modal operators we do not require L = L . Then from Proposition 2 we immediately obtain: Corollary 1. Let O ι,σ O and O ε,ς Oop hold (hence the latter is a Galois connection between O and O ). Then, 1. (a) σι is a closure operator on O ; (b) ισ is an interior operator on O. 2. (a) ςε is a closure operator on O ; (b) ες is a closure operator on O. It is worth underlining that none of these operators needs to be topological. Moreover given the above adjointness situations we can underline what follows:

Formal Topology and Information Systems

271

(m1) σ is half of a co-modal operator: it lacks co-normality; (m2) ι is half of a modal operator: it lacks normality; (m3) and ς are half of an anti-modal operator: they lack anti-normality. The lack of properties concerning preservation of operations may be partially amended when we restrict domains to the families of ﬁxed points of the operators σι, ισ, ες and ςε. To this end the following two results are fundamental: Lemma 2. Let φ : O −→ O be a map. Then, o o (a) if φ is closure then O φ ,φo Imφ ; (b) if φ is interior then Imφ φo ,φ O. Corollary 2. Let φ : L −→ L be a map. Then, 1. if φ is closure then φo is additive, φo (Imφ ) is closed under infs, and Satφ (L) = Imφ , ∧, , 1, where for all x, y ∈ Imφ , x y = φ(x ∨ y), is a lattice; 2. if φ is interior then φo is multiplicative, φo (Imφ ) is closed under sups, and Satφ (L) = Imφ , , ∨, 0, where for all x, y ∈ Imφ , x y = φ(x ∧ y), is a lattice. We want to point out that if φ is closure then sups in L and sups in Imφ may diﬀer. Hence, although for all x ∈ L, φ(x) = φo (x) and φo is additive, nonetheless φ in general is not sup-preserving so that φo (Imφ ) is not closed under sups (dually if φ is interior). These results give the following proposition (where we have just to notice that turning Lop upside-down interiors turns into closures, sups into infs and viceversa): Proposition 3. Let L ι,σ L and L ε,ς Lop hold. Then: 1. Satισ (L) = Imισ , , ∨, 0, where for all x, y ∈ Imισ , x y = ισ(x ∧ y), is a lattice; 2. Satσι (L ) = Imσι , ∧ , , 1 , where for all x, y ∈ Imσι , x y = σι(x ∨ y), is a lattice; 3. Satςε (L ) = Imςε , ∧ , , 1 , where for all x, y ∈ Imςε , x y = ςε(x ∨ y), is a lattice; 4. Satες (L) = Imες , ∧, , 1, where for all x, y ∈ Imες , x y = ες(x ∨ y), is a lattice.

6

Formal Operators on Points and on Observables

Now let us come back to our basic constructors. Proposition 4 (Fundamental relationships). Let A and B be two sets, X ⊆ A, Y ⊆ B, R ⊆ A × B a relation and fˆ ⊆ A × B a functional relation. Then the following holds: 1. (a) R(Y ) ⊆ X iﬀ Y ⊆ [R ](X); (b) R (X) ⊆ Y iﬀ X ⊆ [R](Y ) . 2. Y ⊆ [[R ]](X) iﬀ X ⊆ [[R]](Y );

272

P. Pagliani and M.K. Chakraborty

3. fˆ (X) ⊆ Y iﬀ X ⊆ fˆ(Y ); 4. The operators R, R , [R] and [R ] are isotone; [[R]] and [[R ]] are antitone. Proof. (1) (a) R(Y ) ⊆ X iﬀ R (y) ⊆ X, ∀y ∈ Y iﬀ (from Lemma 1.(1)) Y ⊆ [R ](X). (b) By symmetry. (2) X ⊆ [[R]](Y ) iﬀ ∀x ∈ X(Y ⊆ R(x)) (from Lemma 1.(2)), iﬀ ∀x ∈ X, ∀y ∈ Y (y ∈ R(x)) iﬀ ∀y ∈ Y (X ⊆ R (y)) iﬀ ∀y ∈ Y (y ∈ [[R ]](X)) iﬀ Y ⊆ [[R ]](X) (in view of Lemma 1.(1).(b)). (3) Directly from Proposition 1. (4) Easily from the position of the subformula ”y ∈ Y ” and ”x ∈ X” in the deﬁnitions. From the above discussion we trivially have: = i = e [ ] = [i] [] = [e] [[ ]] = [[i]] [[]] = [[e]] Therefore it is clear that if given a P-system, P, we set M = ℘(M ), ⊆ and G = ℘(G), ⊆, in view of Proposition 4 the following adjointness properties hold: (a) M e,[i] G; (b) G i,[e] M; (c) M [[e]],[[i]] Gop ; (d) G [[i]],[[e]] Mop . The lack of properties involving top and bottom elements, such as ”normality” and ”co-normality”, for generic adjoint functions, is quite obvious since they depend on the adjoint structures. But in the case of the basic constructors Lemma 1.(3) and Proposition 2 immediately prove that e and i are modal operators, [e] and [i] are co-modal operators and, ﬁnally, [[e]] and [[i]] are anti-modal operators. Moreover, in view of these adjunction properties, some sequences of constructors with alternate decorations provide a number of useful operators on ℘(G) and ℘(M ). Indeed axiality says that if one operator lowers an element then its conjugate operator lifts it, and vice-versa, so that by combining them either we obtain the maximum of the lowering elements or the minimum of the lifting elements of a given argument. Deﬁnition 10. Let G, M, be a P -system. Then: – – – – – –

int : ℘(G) −→ ℘(G); int(X) = e([i] (X)). cl : ℘(G) −→ ℘(G); cl(X) = [e] (i(X)). est : ℘(G) −→ ℘(G); est(X) = [[e]] ([[i]] (X)). A : ℘(M ) −→ ℘(M ); A(Y ) = [i] (e(Y )). C : ℘(M ) −→ ℘(M ); C(Y ) = i([e] (Y )). IT S : ℘(M ) −→ ℘(M ); IT S(Y ) = [[i]] ([[e]] (Y )).

The above operators inherit ’d’, ’o’, ’s’, ’sd’, ’os’, ’od’ and ’ods’ reciprocal relationships from the outermost constructors which deﬁne them. Proposition 5. In any P-system G, M, , for any X ⊆ G, Y ⊆ M , g ∈ G, m ∈ M: 1. (a) m ∈ A(Y ) iﬀ e(m) ⊆ e(Y ), (b) g ∈ cl(X) iﬀ i(g) ⊆ i(X); 2. (a) g ∈ int(X) iﬀ i(g) ∩ [i] (X) = ∅, (b) m ∈ C(Y ) iﬀ e(m) ∩ [e] (Y ) = ∅; 3. (a) g ∈ est(X) iﬀ [[i]](X) ⊆ i(g), (b) m ∈ IT S(Y ) iﬀ [[e]](Y ) ⊆ e(m).

Formal Topology and Information Systems

273

Proof. (1) (a) By deﬁnition m ∈ A(Y ) iﬀ m ∈ [i] (e(Y )). Hence from Lemma 1.(1), m ∈ A(Y ) iﬀ e(m) ⊆ e(Y ). (b) By symmetry. (2) (a) g ∈ int(X) iﬀ g ∈ e([i] (X)) iﬀ g ∈ e({m : e(m) ⊆ X}), iﬀ i(g) ∩ {m : e(m) ⊆ X} = ∅, iﬀ i(g) ∩ [i] (X) = ∅. (b) By symmetry. (3) (a) Directly from Lemma 1.(2) and the deﬁnition of ”est”. (b) By symmetry13 . Therefore, g ∈ est(X) if and only if g fulﬁlls at least all the properties that are shared by all the elements of X. In this sense est(X) is the extent of the set of properties that characterises X as a whole. Symmetrically, m ∈ IT S(Y ) if and only if m is fulﬁlled by at least all the objects that enjoy all the properties from Y . In this sense IT S(Y ) is the intent of the set of objects that are characterised by Y as a whole. In order to understand the meaning of the other operators, let us notice that the elements of M can be interpreted as ”formal neighborhoods”14 . In fact, in topological terms a neighborhood of a point x is a collection of points that are linked with x by means of some nearness relation. For a member m of M is associated, via e with a subset X of G, m may be intended as a ’proxy’ of X itself. Thus if X is a concrete neighborhood of a point x, then m may be intended as a formal neighborhood of x, on the basis of the observation that the nearness relation represented by X states that two points are close to each other if they both fulﬁll property m15 . It follows that obs(g) is the family of formal neighborhoods of g (symmetrically for sub(m) we have the concrete neighborhoods of m). This is the intuitive content of the following gift of the adjointness relationships between basic constructors: Interior operators int, C Closure operators cl, A, est, IT S In view of the observation after Corollary 1 one easily notices that none of the above operators needs to be topological. 6.1

Fundamental Properties of the Formal Perception Operators

Deﬁnition 11. Let P be any P-system. Then we deﬁne the following families of ﬁxpoints of the operators induced by P: 1. Ωint (P) = {X ⊆ G : int(X) = X}; Γcl (P) = {X ⊆ G : cl(X) = X}; 2. Γest (P) = {X ⊆ G : est(X) = X}; ΩA (P) = {Y ⊆ M : A(Y ) = Y }; 3. ΓC (P) = {Y ⊆ M : C(Y ) = Y }; ΓIT S (P) = {Y ⊆ M : IT S(Y ) = Y }. 13

14

15

Moreover, [[i]](X) = {m : X ⊆ e(m)} = {m : ∀x ∈ X(x m)}. Henceforth [[i]](X) ⊆ i(g) iﬀ g m for all m such that x m, for any member x of X, that is, iﬀ ∀m ∈ M ((∀x ∈ X(x m)) =⇒ g m). Indeed, this is the framework in which the operators cl, int, A and C are introduced, although not by means of adjointness properties, by the Padua School of Formal Topology (see [26]). This interpretation is close to the approach of [11].

274

P. Pagliani and M.K. Chakraborty

It is understood that the partial order between saturated subsets is inherited from the category they are derived from. Thus, for instance, we shall have ΓC (M ), ⊆. Proposition 6. Let P be a P-system. Then the following are complete lattices: 1. Satint (P) = Ωint (P), ∪, ∧, ∅, G, where i∈I Xi = int( i∈I Xi ); 2. SatA (P) = ΩA (P), ∨, ∩, ∅, M , where i∈I Yi = A( i∈I Yi ); 3. Satcl (P) = Γcl (P), ∨, ∩, ∅, G, where i∈I Xi = cl( i∈I Xi ); 4. SatC (P) = ΓC (P), ∪, ∧, ∅, M , where i∈I Yi= C( i∈I Yi ); 5. Satest (P) = Γest (P), ∩, ∨, est(∅), G, where i∈I X i ); i = est( i∈I X 6. SatIT S (P) = ΓIT S (P), ∩, ∨, IT S(∅), M , where i∈I Yi = IT S( i∈I Yi ). Proof. Much work has already been done in Proposition 3. We just need to justify the choice of top and bottom elements. To this end, remember that in any P-system both and are onto. Hence in view of Lemma 1.(3). int(G) = e[i](G) = [ ](G) = (M ) = G, and analogously for the other operators. The only diﬀerence is for IT S and est because [[]](∅) = G but [[ ]](G) = {m : (m) = G} ⊇ ∅, dually for [[]][[ ]](∅). Lemma 3. Let P be a P-system. Then for all X ⊆ G, Y ⊆ M , X ∈ Ωint (P) iﬀ X = e(Y ) X ∈ Γcl (P) iﬀ X = [e](Y ) X ∈ Γest (P) iﬀ X = [[e]](Y ) Y ∈ ΩA (P) iﬀ Y = [i](X ) Y ∈ ΓC (P) iﬀ Y = i(X ) Y ∈ ΓIT S (P) iﬀ Y = [[i]](X )

for some Y ⊆ M, X ⊆ G. Proof. If X = e(Y ) then X = e[i]e(Y ), from Proposition 2.(c). Therefore, by deﬁnition of int, X = int(e(Y )) = int(X). Vice-versa, if X = int(X), then X = e[i](X). Hence, X = e(Y ) for Y = [i](X). The other cases are proved in the same way, by exploiting the appropriate equations of Proposition 2.(c). Corollary 3. Let P be a P-system. Then the following are isomorphisms: 1. (a) e : SatA (P) −→ Satint (P); (b) [i] : Satint (P) −→ SatA (P); 2. (a) [e] : SatC (P) −→ Satcl (P); (b) i : Satcl (P) −→ SatC (P); 3. The following are anti-isomorphisms (where − is the set-theoretical complementation): (a) [[i]] : Satest (P) −→ SatIT S (P); [[e]] : SatIT S (P) −→ Satest (P); (b) − : Satcl (P) −→ Satint (P); − : SatC (P) −→ SatA (P). Proof. Let us notice, at once, that the proof for an operator requires the proof for its adjoint operator. Then, let us prove (1).(a) and (b) together: First, let us prove bijection for e and [i]. From Lemma 3 the codomain of e is Ωint (P) and the codomain of [i] is ΩA (P). Moreover, for all X ∈ Ωint (P), X = e[i](X) and for all Y ∈ ΩA (P), Y = [i]e(Y ). From the adjointness properties we have:

Formal Topology and Information Systems

275

(i) e is surjective on Ωint (P) and (ii) [i] is injective from Ωint (P). (iii) e is injective from ΩA (P) and (iv) [i] is surjective onto ΩA (P). Moreover, if [i] is restricted to Ωint (P), then its codomain is the set H = {Y : Y = [i](X) & X ∈ Ωint (P)}. Clearly, H ⊆ ΩA (P). In turn, if e is restricted to ΩA (P), then its codomain is the set K = {X : X = e(Y ) & Y ∈ ΩA (P)}. Clearly K ⊆ Ωint (G). Therefore, (i) and (iii) give that e is bijective if restricted to ΩA (P), while (ii) and (iv) give that [i] is a bijection whenever restricted to Ωint (P)16 . Now we have to showthat e and [i] preserve joins and meets. For e we proceed as follows: (v) e( i∈I (A(Yi ))) =def e(A( i∈I (A(Yi ))). But eA = e, from Proposition2.(c). Moreover, e distributes over unions. Hence the right side of (v) equals to i∈I e(A(Yi )). But in view of Proposition 6, the union of extensional open subsets is concrete open and from Lemma 3 e(A(Y i )) belongs indeed to ∈ Ω (P), so that the right side of (v) turns into int( int i∈I e(A(Yi ))) =def e(A(Y )). i i∈I (vi) e( i∈I A(Yi )) = e( i∈I [i]e(Yi )). Since [i] distributes over intersections, the right side of (vi) turns into e[i]( i∈I e(Y )) = int( i i∈I e(Yi )). But e = eA, so that the last term is exactly i∈I e(A(Yi )). Since [i] is the inverse of e, qua isomorphism, we have that [i] preserves meets and joins, too. As to (2) the results come by symmetry. (3) (a) As in the above proof by noticing that in polarities the right structure is reversed upside-down (we can optimize a passage by noticing that [[e]] and [[i]] are both upper and lower adjoints). (b) By duality between the operators. 6.2

Pre-topological Approximation Spaces

Now we are in position to show how the above mathematical machinery may be used to generalise the upper and lower approximation operators provided by Rough Set Theory. Given X ⊆ G we know that [e]i(X) ⊇ X and e[i](X) ⊆ X. We can interpret these relationships by saying that – cl is an upper approximation of the identity map on ℘(G); – int is a lower approximation of the identity map on ℘(G). More precisely, i(X) = min([e]← (↑ X)) = min{X ⊆ G : [e](X ) ⊇ X}, it follows that [e]i(X) (i. e. cl) is the best approximation from above to X via function [e]. Dually, [i](X) = max(e← (↓ X)) = max{X ⊆ G : e(X ) ⊆ X}. Hence, e[i](X) (i. e. int) is the best approximation from below to X, via function e. Of course, if i is injective (or, equivalently, [e] is surjective), then we can exactly 16

As side results, we have: (i) ΩA (P) = H and (ii) Ωint (P) = K. This is not surprising, because if Y ∈ ΩA (P) then Y = [i]e(Z) for some Z ⊆ M and e(Z) ∈ Ωint (P), any Z ⊆ M . Vice-versa, if X ∈ Ωint (P), then X = e(Z). Hence [i](X) = [i]e(Z) belongs to ΩA (P). Symmetrically for (ii).

276

P. Pagliani and M.K. Chakraborty

reach X from above by means of [e]. The element that must be mapped is, indeed, i(X). Dually, if [i] is injective (or e is surjective), then we can exactly reach X from below by means of e applied to [i](X).

7

Information, Concepts and Formal Operators

So far we have discussed a number of instruments that act on either abstraction sides we are dealing with, that is, points and properties. Indeed, the introduced operators are based on well-deﬁned mathematical properties, such as adjointness, and feature proper informational and conceptual interpretations. Also, the use of the terms ”open” and ”closed” is not an abuse because, on the contrary, these operators translate the usual topological deﬁnitions of an interior and, respectively, a closure of a set X ⊆ G, into the language of observation systems, provided the elements of M are interpreted as formal neighborhoods. For instance, the usual deﬁnition tells us that for any subset X of G, a point a belongs to the interior of X if and only if there is a neighbourhood of a included in X. If the elements of the set M are intended as formal neighbourhoods, then the relation a b (hence, a ∈ e(b)) says that b is a formal neighborhood of a and e(b) ⊆ X says that the extension of this neighbourhood b is included in X. But this is precisely a reading of a ∈ e[i](X), because in view of the adjunction properties, e(b) ⊆ X if and only if b ∈ [i] (X). Thus we have made a further step beyond M. Smyth’s seminal observation that semi-decidable properties are analogous to open sets in a topological space, with the aid of the interpretation of basic constructors elaborated by the Padua School on Formal Topology17 . Moreover, we have seen that int and cl provide us with lower and upper approximations of any set X ⊆ G. But are we really happy with this machinery? The answer is ”yes and no”. Yes, for we have found a mathematically sound way to deal with approximations which enjoy a reliable intuitive interpretation. No, for both int and cl are discontinuous (non topological) operators because int is not multiplicative and cl is not additive, so that we have to face ”jumps” which can be too wide and make us miss information. EXAMPLE 1 Here we give an example of a P-system and its induced operators: a a1 a2 a3

17

See for instance [25] and [27].

b b1 1 1 0 1 0 1 0 0

b2 0 0 1 0

b3 0 1 1 1

Formal Topology and Information Systems

277

Let us try and compute some instances of basic formal operators on system P: 1) Extensional operators: int({a, a1 }) = e[i]({a, a1 }) = e({b}) = {a}; int({a2 }) = e({b2 }) = {a2 }. cl({a, a1 }) = [e]i({a, a1 }) = [e]({b, b1 , b3 }) = {a, a1 , a3 }; cl({a, a2 }) = [e](M ) = G. est({b1 , b2 }) = [[i]][[e]]({b1 , b2 }) = [[i]]({a2 }) = {b1 , b2 , b3 }. int(int({a, a1 })) = int({a}) = e[i]({a}) = e({b}) = {a}. cl(cl({a, a1 })) = cl({a, a1 , a3 }) = [e]i({a, a1 , a3 }) = [e]({b, b1 , b3 }) = {a, a1 , a3 }. est(est({b1 , b2 }))= est({b1 , b2 , b3 })= [[i]][[e]]({b1 , b2 , b3 })= [[i]]({a2 })= {b1 , b2 , b3 }. Thus, this is also an example of the fact that int is decreasing while cl and est are increasing and all of them are idempotent. Moreover, one can see that int({a, a1 }) ∪ int({a2 }) = {a, a2 } ⊆ {a, a1 , a2 } = int({a, a1 } ∪ {a2 }) and cl({a, a1 }) ∩ cl({a, a2 }) = {a, a1 , a2 } ⊇ {a} = cl({a, a1 } ∩ {a, a2 }). 2) Intensional operators: A({b, b1 }) = [i]e({b, b2 }) = [i]({a, a1 , a2 }) = {b, b1 , b2 }. C({b2 , b3 }) = i[e]({b2 , b3 }) = i({a3 }) = {b3 }. IT S({b1 , b2 }) = [[i]][[e]]({b1 , b2 }) = [[i]]({a2 }) = {b1 , b2 , b3 }. A(A({b, b1 })) = A({b, b1 , b2 }) = [i]e({b, b1 , b2 }) = [i]({a, a1 , a2 }) = {b, b1 , b2 }. C(C({b2 , b3 })) = C({b3 }) = i[e]({b3 }) = i({a3 }) = {b3 }. IT S(IT S({b1 , b2 }))= IT S({b3 })= [[i]][[e]]({b1 , b2 , b3 }) = [[i]]({a2 }) = {b1 , b2 , b3 }. Thus, this is also an example of the fact that C is decreasing while A and IT S are increasing and all of them are idempotent. Let us now visualise the lattices of saturated sets: Satint (P)

Satcl (P)

G G bb bb "" "" {a1 , a2 , a3 } {a, a1 , a2 } {a1 , a2 , a3 } {a, a1 , a3 } bb bb bb bb "" bb bb {a1 , a3 } {a, a2 } bb bb bb bb "" {a2 } {a} {a3 } {a} bb bb "" "" ∅ ∅ SatA (P) SatC (P) M M bb bb "" "" {b2 , b3 } {b, b1 , b2 } {b1 , b2 , b3 } {b, b1 , b3 } bb bb bb bb "" bb bb {b1 , b3 } {b, b2 } bb bb bb bb "" {b2 } {b} {b3 } {b, b1 } bb bb "" "" ∅ ∅

278

P. Pagliani and M.K. Chakraborty

SatIT S (P)

Satest (P)

G G bb bb "" "" {b, b1 } {b1 , b2 , b3 } {a1 , a2 , a3 } {a, a1 , a2 } bb bb bb bb "" bb bb {a1 , a2 } {b1 , b3 } bb bb bb bb "" {b1 } {b3 } {a2 } {a} bb bb "" "" ∅ ∅

Pay attention that SatIT S (P) and SatA (P) have the same shape just by chance (idem for Satest (P) and Satcl (P)). The non topological nature of these operators is openly visible in the above pictures. For instance in Satint (P) we do not have the intersection of {a1 , a2 , a3 } and {a, a1 , a2 }. Hence int does not distribute over intersections. in Satcl (P) we lack the union of {a3 } and {a}, so that cl does not distribute over unions. However we have a few results which will be useful. Proposition 7. Let θ and φ be two dual basic operators. Then, (a) θ is a closure operator if and only if φ is an interior operator; (b) θ is topological if and only if φ is topological. Proof. (a) Trivially, since complementation reverses the order. (b) Suppose θ is additive, then φ(X ∩Y ) = −θ−(X ∩Y ) = −θ(−X ∪−Y ) = −(θ(−X)∪θ(−Y )) = −θ(−X) ∩ −θ(−Y ) = φ(X) ∩ φ(Y ). Dually for the opposite implication. In order to try and solve the above issue, we must notice that any answer and solution depends on the nature of the P-system at hand. Generally, the nature of points is not really important. More important is the nature of properties. And, even more important is the nature of the operator supposed to better represent the basic perception act. 7.1

Choosing the Initial Perception Act

We have assumed that our ﬁrst act of knowledge is a grouping act, a sort of ”data abstraction”. However, we can basically perform this act in two opposite ways: either collect around an object g the elements which fulﬁlls at least all the properties (or attribute-values) of g, or the elements fulﬁlling at most all the properties (or attribute-values) of g. Otherwise stated, in the ﬁrst case we collect the objects which are characterised at least as g by the properties (attributes) at hand, while in the second case we collect the objects which are characterised at most as g. However if we consider attribute-values the two conditions collapse (see later on). Moreover notice that the grouping rule just asserted does not imply any form of symmetry. Indeed, g could manifest all the properties of g but also additional properties that are not manifested by g. To put it another way, we are claiming

Formal Topology and Information Systems

279

that our basic grouping act is not based on the notion ”to manifest exactly the same properties”, but on the notion ”to manifest at least (or at most) the same properties”. Indeed, the set of properties which are manifested by g, is the attracting phenomenon around which we form our perception. Thus from an analytical point of view we have just to focus on these properties and not to take into account additional properties. In the present paragraph we shall see that the former notion is subsumed by the latter. That is, if we deﬁne this way the basic ”cells” of our categorisation process, we shall be able to cover a wider range of cases. Let G, At, {Va }a∈At be an Attribute System. Then for all g ∈ G deﬁne: Qg = {g : ∀a ∈ At, ∀x ∈ Va ((a(g) = x) =⇒ (a(g ) = x))} Let G, M, be a Property System. Then for all g ∈ G deﬁne: Qg = {g : ∀p ∈ M (p ∈ i(g) =⇒ p ∈ i(g ))} Qg will be called the quantum of information at g. In view of the previous discussion we adopt quanta of information because they reﬂect the idea ”g is perceived together g whenever it manifests at least the same properties as g”. Therefore, given (the properties manifested by) an object g the perception cell organised around g should be Qg which should be referred to as the ”minimum perceptibilium at location g”, for it is not possible to perceive g without perceiving (the manifestations of) the other members of that ”perception parcel”. Therefore, we call such perception parcel a quantum of perception or a quantum of information at g. This terminology drew its inspiration from [4] and this term expresses a sort of programme that may be epitomized by the slogan ”any information is a quantum of information”. As to quanta of information from a Property System, we can elaborate a little further (we shall resume Attribute Systems later on). Proposition 8. Let G, M, be a P-system and g, g ∈ G. Then, 1. 2. 3. 4.

Qg = est(g); g ∈ Qg iﬀ i(g) ⊆ i(g ) g ∈ Qg iﬀ for all X ∈ Γest , g ∈ X =⇒ g ∈ X; g ∈ Qg iﬀ for all p ∈ M, g ∈ e(p) =⇒ g ∈ e(p); g ∈ Qg iﬀ g ∈ cl(g ).

Proof. (1) Indeed, g ∈ est(g) iﬀ i(g ) ⊇ [[i]](g). But [[i]](g) = i(g), whence g ∈ est(g) if and only if p ∈ i(g) =⇒ p ∈ i(g ) if and only if g ∈ Qg if and only if i(g) ⊆ i(g ). (2) () Suppose X ∈ Γest and g ∈ X =⇒ g ∈ X. Then i(g) ⊇ [[i]](X) implies i(g ) ⊇ [[i]](X). Since this happens for all est-saturated X, it happens for [[i]](g) too and we trivially obtain i(g ) ⊇ i(g), so that g ∈ Qg . () If g ∈ Qg then i(g ) ⊇ i(g). If, moreover, g ∈ X, for X = est(X), then i(g) ⊇ [[i]](X). By transitivity, i(g ) ⊇ [[i]](X), whence g ∈ X too. (3)

280

P. Pagliani and M.K. Chakraborty

Indeed, g ∈ e(x) if and only if g x if and only if x ∈ i(g). Hence for all p ∈ M, g ∈ e(p) =⇒ g ∈ e(p) if and only if for all p ∈ M, p ∈ i(g) =⇒ p ∈ i(g ). (4) So, g ∈ Qg if and only if i(g) ⊆ i(g ) if and only if g ∈ cl(g ). Indeed, these results, trivial consequences of Lemma 1 and Proposition 4, formally state that g is perceived together with g if and only if it fulﬁlls at least all the properties fulﬁlled by g. Moreover, a quantum of perception at location g is the universal extension of function sub to the set of properties (g) fulﬁlled by g. In turn, since we start from a singleton {g}, (g) is both a universal and an existential extension of function obs. When we have to move from grouping maneuvers around a single object to grouping maneuvers around two or more objects we have essentially two kinds of choice: universal extensions from singletons to a generic set X and existential extensions. The existential extension is deﬁned as Qx (7) Q∪ X = x∈X

This is not the sole choice, but for the very reasons discussed so far, we shall adopt it. Moreover, it makes a uniform treatment of both P-systems and Asystems possible. As for universal extensions we brieﬂy discuss only the following alternative: Q⊗ X = [[e]][[i]](X) = est(X).

(8)

The superscript ⊗ underlines the fact that in est(X) we consider the properties which glue the elements of X together. Otherwise stated, we extract from i(X) those properties P which are shared by all the elements of X. Then we make exactly the same thing with respect to P . Thus, according to this universal extension, an object g belongs to Q⊗ X if g fulﬁlls all the properties fulﬁlled by all the elements of X. Whenever we need to distinguish the system inducing a quantum of information, we shall use the name of the system as an exponent. The same for P any operator (for instance we shall write QP g and int if needed). Moreover, if O = X, R, we set −O = X, −R and O = X, R . 7.2

Information Quantum Relational Systems

Now that we have chosen the basic mechanisms (basis and step) leading from atomic perception (or ”elementary perception cells”) to complex perception, let us analyse what kinds of a relation arise between elements of G from these grouping maneuvers. Let us then set the following deﬁnition: Deﬁnition 12 (Information Quantum Relational System). Let S be an A-system or a P-system over a set of points G. Let R be a binary relation on G. We say that R is induced by S whenever the following holds, for all g, g ∈ G: g, g ∈ R iﬀ g ∈ Qg .

Formal Topology and Information Systems

281

We call R the information quantum relation - or i-quantum relation or quantum relation in short - induced by S and it will be denoted as RS . Moreover, Q(S) will denote the relational system G, RS , called the Informational Quantum Relational System - IQRS, or Quantum Relational System in short, induced by S. Finally we set ΩQ (S) = {RS (X) : X ⊆ G}18 . Since g ∈ Qg says that g fulﬁlls at least all the properties fulﬁlled by g , then g , g ∈ RS has the same meaning. Clearly the properties of i-quantum relations depend on the patterns of objects induced by the given systems. However, they uniformly fulﬁll some basic properties. In view of the additivity property of generalised quanta, we can conﬁne our attention to i-quanta at a location. Lemma 4. In any A-system S over a set G, for all g, g , g ∈ G: 1. (a) g ∈ Qg (q-reﬂexivity); (b) g ∈ Qg & g ∈ Qg =⇒ g ∈ Qg (qtransitivity). 2. If S is an A-system, a functional or a dichotomic P-system then g ∈ Qg implies g ∈ Qg (AFD-q-symmetry). Proof. The ﬁrst two statements are obvious consequences of transitivity and reﬂexivity of the relation ⊆. Notice that antisymmetry does not hold because of the obvious fact that g ∈ Qg and g ∈ Qg does not imply g = g. Now, let S be an A-system. Suppose g ∈ Qg and a(g) = x, then a(g) = x for some x = x so that a(g ) = x , because g ∈ Qg , whence a(g ) = x too. Therefore, g ∈ Qg implies g ∈ Qg , so that the induced relation is also symmetric. If S is a functional P-system then we trivially obtain the proof from deﬁnitions and the fact that i(g) is a singleton. Finally, if S is dichotomic and g ∈ Qg , then g fulﬁlls at least the same properties as g. Now, if g p while g p, then g p, where p is a complementary copy of p. But g p, since it fulﬁlls p. Hence we cannot have i(g) ⊆ i(g ), whence g ∈ Qg . Contradiction. So notice that in A-systems the universal quantiﬁcation over attribute-values hides a bi-implication because the set of attribute-values of g and that of g must coincide in order to have g ∈ Qg . As an immediate consequence of the above result we have: Proposition 9. Let S be an A-system or a P-system. Then: 1. The i-quantum relation RS induced by S is a preorder; 2. If S is an A-system or an FP or DP system, then RS is an equivalence relation; 3. If S is an FP-system then RS = ⊗ and g ∈ Qg iﬀ g ∈ [g]k . 18

I-quanta and i-quantum relations from A-systems were introduced in [13], with different names. If the entire set M is considered as a multi-valued property, then i-quantum relations coincide with the so-called ”forward inclusion relations” introduced in [12].

282

P. Pagliani and M.K. Chakraborty

Proof. We just have to prove statement (3). In view of Proposition 8.(4) we just need to show that if is a map then for all x ∈ G, cl({x}) = [x]κ . From Proposition 7.1, a ∈ cl({a }) if and only if i({a}) ⊆ i({a }). Therefore, if happens to be a map, we have that a ∈ cl({a }) if and only if i({a}) = i({a }), since exactly one value is admitted. We can conclude that for all x ∈ G, cl({x}) = {x : i(x ) = i(x)} = ei({x}) = [x]k (that is, the kernel of ). EXAMPLE 2 Consider the P-system P of EXAMPLE 1. Let us compute some quanta of information of P: Qa = {a}, Qa = {a , a }, Qa = {a }, Qa = {a , a , a }, Q{a,a } = {a, a , a } and so on addictively. Thus a , a ∈ RS because a ∈ Qa but the opposite does not hold. Here below the i-quantum relations RP and RQ(P) are displayed: RP a a a a

a a 1 0 0 1 0 0 0 1

a 0 1 1 1

a RQ(P) 0 a 0 a 0 a 1 a

a a 1 0 0 1 0 1 0 0

a 0 0 1 0

a 0 1 1 1

It is easy to verify that both of the above relations are reﬂexive and transitive, Q(P) = RQ(P) . Moreover one can see that, for instance, Qa = {a , a } and RP or Qa = {a , a , a }. Indeed we have that a ∈ Qa and a ∈ Qa , or Q(P) a ∈ Qa whereas a ∈ Qa , and so on. Now consider the A-system A = G = {a, a1 , a2 , a3 }, At = {A, B, C}, {VA = {0, 1, 3}, VB = {b, c, f }, VC = {α, δ}} such that A(a) = A(a2 ) = 1, A(a1 ) = 0, A(a3 ) = 3, B(a) = B(a2 ) = b, B(a1 ) = c, B(a3 ) = f, C(a) = C(a1 ) = C(a2 ) = A A A α, C(a3 ) = δ. We have: QA a = Qa2 = {a, a2 }, Qa1 = {a1 }, Qa3 = {a3 }. The resulting i-quantum relation RA = {a, a2 , a2 , a, a, a, a2 , a2 , a1 , a1 , a3 , a3 } is an equivalence relation. Q(P)

Q(P)

From the above results we obtain immediately some interesting consequences about functional P-systems: Corollary 4. Let P be an FP-system. Then, (a) cl is a topological closure operator; (b) int is a topological interior operator. Proof. From Proposition 8.(4) and Proposition 9.(3) we have that cl(x) = [x]k . But k is the kernel of and the kernel of a function is a congruence. It follows by induction that cl(X) ∪ cl(Y ) = [X]k ∪ [Y ]k = [X ∪ Y ]k = cl(X ∪ Y ). Hence cl is additive. Since int is dual of cl we immediately obtain that int is multiplicative. We now list some results in terms of IQRSs. Since i-quantum relations are preorders, it is useful to prove some general facts about this kind of relations:

Formal Topology and Information Systems

283

Proposition 10. Let O = X, R be any preordered set. Then for any x, y ∈ X the following are equivalent: (1) y ∈ R (x); (2) R (y) ⊆ R (x); (3) x ∈ QO y ; (4) x ∈ R(y); (5) O y ∈ Qx . Proof. (1 2) y ∈ R(x) iﬀ x, y ∈ R. Suppose y, y ∈ R. Since R is transitive, for all y ∈ X, y, y ∈ R =⇒ x, y ∈ R so that R (x) ⊇ R (y). Conversely, since R is reﬂexive, y ∈ R(y) holds. Thus if R (x) ⊇ R (y) then y ∈ R (x). All the other equivalences are obvious consequences or even just deﬁnitions. Corollary 5. Let S be an A-system or a P-system over a set G. Then, g ∈ QS iﬀ g ∈ Qg g iﬀ g ∈ RS (g) iﬀ g ∈ Qg g ∈ Q−S g . Q(S)

Q(Q(S))

iﬀ g ∈ RS (g ) iﬀ

Proof. The ﬁrst equivalence is just a deﬁnition. Now, g ∈ RS (g) iﬀ RS (g ) ⊆ Q(S) RS (g) iﬀ g ∈ Qg iﬀ g ∈ RQ(S) (g ) iﬀ RQ(S) (g) ⊆ RQ(S) (g ) iﬀ g ∈ Qg . From this we have that Qg = RS (g) so that in view of trivial set-theoretic considerations, (X ⊆ Y iﬀ −Y ⊆ −X) we obtain the last two equivalences. Q(Q(S))

Q(S)

These equivalences show that IQRSs of level higher than 1 do not provide any further information. Corollary 6. If S is an A-system, an FP-system or a DP-system over a set Q(S) G, then for all g, g ∈ G, X ⊆ G, (a) g ∈ QS ; (b) RS (X) = g iﬀ g ∈ Qg RQ(S) (X). Moreover, since a P-system is a generic relational system we have that all facts valid for P-systems are valid for any relational system. The notion of a quantum of information is asymmetric for P-systems, because if g fulﬁlls strictly more properties than g, we have g ∈ Qg but g ∈ Qg . On the contrary it is symmetric in the case of A-systems and dichotomic or functional P-systems.

8

Higher Level Operators

Let S be an A-system and let Q(S) = G, G, RS be its induced IQRS. What kinds of patterns of data can we collect by applying our operators to these derivative systems? First of all, since in IQRSs there is no longer the distinction between objects and properties and intension or extension, it is better we change once more our symbols and notation:

284

P. Pagliani and M.K. Chakraborty

The operator deﬁned as turns into i i(X) = {g : ∃g (g ∈ X & g , g ∈ RS )} RS e e(X) = {g : RS (g) ∩ X = ∅} RS [i] [i](X) = {g : RS (g) ⊆ X} [RS ] [e] [e](X) = {g : RS (g) ⊆ X} [RS ] Let us call the above operators decorated with RS ”quantum operators” (notice that in this context [[RS ]] and [[RS ]] are not quantum operators). Quantum operators behave in a very particular way, because, we remind, they fulﬁll adjoint properties. Namely RS [RS ] and RS [RS ]. Actually, the following results apply to any preorder. Proposition 11. Let Q(S) = G, G, RS be a IQRS. Let Oi and Oj be any two adjoint quantum operators from the set {RS , RS , [RS ], [RS ]}. Then (a) Oi Oj = Oj ; (b) Oj Oj = Oj ; (c) the ﬁxpoints of Oi and Oj coincide. Proof. (a) (i) In view of Proposition 5, for all g ∈ G and X ⊆ G, g ∈ [RS ]RS (X) iﬀ RS (g) ⊆ RS (X), iﬀ (from Proposition 10) g ∈ RS (X). One can prove g ∈ [RS ]RS (X) iﬀ g ∈ RS (X) similarly. (ii) In view again of Proposition 5, a ∈ RS [RS ](X) iﬀ RS (a) ∩ [RS ](X) = ∅, iﬀ there is a such that a ∈ RS (a ) and a ∈ [RS ](X). But a ∈ [RS ](X) iﬀ RS (a ) ⊆ X iﬀ RS (a) ⊆ X. Hence, a ∈ RS [RS ](X) iﬀ a ∈ [RS ](X). (b) From point (a) and Proposition 6.3, Oj Oj = Oi Oj Oi Oj = Oi Oj = Oj . (c) Let X = Oj (X). Then from point (a) Oi (X) = Oi Oj (X) = Oj (X) = X. Deﬁnition 13. Let S be an A-system or a P-system. With ΩQ (S) we shall denote the family {QS X : X ⊆ G}. With Qn (S) we denote the n-nested application of the functor Q to S. In view of these results we can prove a number of properties. Lemma 5. For any P-system P, (a) Ωint (P) ⊆ ΩQ (P); (b) Γcl (P) ⊆ Ωint (Q(P)). Proof. (a) Assume X QX . Thus we must have some x such that x ∈ / X but x ∈ QX . Thus there is g ∈ X such that (x) ⊇ (g), so that for all / intP (X) m ∈ M such that g m surely (m) X. It follows that g ∈ P P and, hence, int (X) = X. (b) We remind that x ∈ cl (X) iﬀ i(x) ⊆ i(X) iﬀ (x) ⊆ (X). Moreover, if x ∈ X, i(x) ⊆ i(X). Now suppose X = intQ(P) (X). Then there is x ∈ X such that RP (x) X. Hence {y : x ∈ QP y} X. Thus {y : i(y) ⊆ i(x)} X. This means that there is g ∈ / X such that i(g) ⊆ i(x) ⊆ i(X) so that i(g) ⊆ i(x) ⊆ i(X). It follows that X clP (X). Corollary 7. Let Q(S) be an IQRS. Then, Q(S)

(a) QS (...) = RS = cl; (b) Q(...) = RS = A; (c) [RS ] = int; (d) [RS ] = C.

where the operators cl, int, C, A are intended over Q(S).

Formal Topology and Information Systems

285

Corollary 8. Let S be an A-system or a P-system and A, B ⊆ G. Then, Q(S)

1. QS (...) , Q(...) , RS and RS are topological closure operators and their images are closed under intersections. 2. [RS ] and [RS ] are topological interior operators and their images are closed under unions. Q(S)

Proof. (1) From Corollary 7, we have that QS (...) , Q(...) , RS and RS are closure operators. Moreover, since they are lower adjoints in the category ℘(G), ⊆ they preserve unions. Finally, from Proposition 7.2 they are normal and their images are closed under intersections. (2) From Corollary 7, [RS ] and [RS ] are interior operators. Moreover as they are lower adjoints in the category ℘(G), ⊆ they preserve intersections. Finally from Proposition 7.2 they are conormal and their images are closed under unions.

Corollary 9 (I-quantum systems). Let S be an A-system or a P-system. Then, 1. SatQ (S) = ΩQ (S), ∪, ∩, ∅, G is a distributive lattice, called the I-quantum system - IQS induced by S. 2. SatQ (Q(S)) = ΩQ (Q(S)), ∪, ∩, ∅, G is a distributive lattice, called the coI-quantum system - co-IQS induced by S. 3. The set theoretical complement is an antisomorphism between SatQ (S) and SatQ (Q(S)). 4. Satint (Q(S)), Satcl (Q(S)), SatA (Q(S)) and SatC (Q(S)), equipped with the set-theoretical operations, are distributive lattices. 5. G, ΩQ (S) and G, ΩQ (Q(S)) are topological spaces, where the interior operators are intQ2 (S) and, respectively, intQ1 (S) . Proof. (1) We know that the operator Q(...) is additive. Thus ΩQ (S) is closed under unions. From Corollary 8 it is closed under intersections too. Moreover, for ΩQ (S) is a (ﬁnite) lattice of sets Sat(S) inherits distributivity from the corresponding property of unions and intersections. (2) Since Q(S) is a P-system the above considerations apply to this structure. (3) From Corollary 5 we know that −RS = RS , so that we obtain immediately the thesis. (4) From Proposition 7.2 and Corollary 8. (5) Any family of open sets of a topological space enjoys distributivity of arbitrary unions over ﬁnite intersections and of intersection over arbitrary unions. Moreover, from Proposition 11 and Corollary 7 we obtain that, ΩQ (S) = Γcl (Q(S)) = Ωint (Q2 (S)) and ΩQ (Q(S)) = ΩA (Q(S)) = Ωint (Q(S)). Now by means of the above mathematical machinery we prove a key statement in the theory of Approximation Space and Rough Sets, namely the well-known fact that the family of deﬁnable sets can be made into a Boolean algebra. Proposition 12 (Quantum relations and Boolean algebras). Let S be an Information system. If RS is an equivalence relation, then SatQ (S) is a Boolean algebra.

286

P. Pagliani and M.K. Chakraborty

Proof. We show that if RS is an equivalence relation then any element QX of ΩQ (S) is complemented by QX = z∈QX Qz . First, let us prove that QX ∪QX = G. In fact for all g ∈ G if g ∈ QX then g ∈ QX because g ∈ Qg (q-reﬂexivity). Now we prove that QX ∩ QX = ∅. Assume z ∈ QX . We have just to prove that if z ∈ Qz then z ∈ QX . So let z ∈ Qz . For q-symmetry z ∈ Qz . So, if there is an x ∈ X such that z ∈ Qx we have z ∈ Qx too (for q-transitivity), hence z ∈ QX . Contradiction.19 Corollary 10. Let S be an Information system. Then, if S is an A-system, a dichotomic or a functional P-system, then SatQ (S) is a Boolean algebra. About the family of co-prime elements of SatQ (S) we have: Lemma 6. Let S be an A-system or a P-system. Then for any X ∈ J (SatQ (S)), X = Qg for some g ∈ X. Proof. Trivial from the very additive deﬁnition of the operator Q and its increasing property. Lemma 7. Let P be a P-system and g ∈ G. Then Qg = {e(m) : m ∈ i(g)} Proof. Indeed, x ∈ Qg iﬀ i(x) ⊇ i(g) iﬀ x ∈ e(m), ∀m ∈ i(g). Proposition 13. Let P be a P-system such that cl (int) is topological. Then SatQ (P) = Satint (P). Proof. We have seen in Lemma 5 that Ωint (P) ⊆ ΩQ (P). Now we need just to show that if X ∈ J (SatQ (P)) then X = int(X). The proof is immediate. Indeed, the family {e(m) : e(m) ⊆ X} is a base of Ωint (S). Moreover, if int is topological then it is multiplicative and since for all m ∈ M , int(e(m)) = e(m) (from Lemma 3), in view of the above Lemma 7 we have the result. Corollary 11. Let F be an FP-system. Then ΩQ (Qn (F)) = Ωint (Qn (F)), n ≥ 0. Proof. From Corollary 4 and an inductive extension of Proposition 13. Hence we can note that P-systems such that int and cl are topological behave like functional systems. Corollary 12. If S is a preordered set (that is, G = M and R ⊆ G × G is a preorder), then ΩQ (S) = Ωint (S). 19

There is another way to obtain this result. In fact, J. L. Bell proved that if T = A, T is a relational structure with T a tolerance relation (that is, reﬂexive and symmetric) then the family ΩQL (T) of all unions of principal order ﬁlters ↑T x (i. e. T (x) i. e., for symmetry of T , T (x)) can be made into an ortholattice. But if RS is an equivalence relation then it is a tolerance relation too and for any x ∈ A, ↑RS x =↓RS x = cl(x) = Qx so that SatQ (S) can be made into a distributive ortholattice, that is, a Boolean algebra.

Formal Topology and Information Systems

287

Proof. From Corollary 11 and Corollary 9 (3). At this point we can end this subsection with an analogue of the duality between distributive lattices and preorders in the context of i-quantum relations and Psystems. Proposition 14 (Duality between preorders and P-systems) 1. Let O = G, R be a preorder, then there is a P-system I(O) over G such that RI(O) = R (hence, Q(I(O)) ∼ =I O). 2. Let S be an A-system or a P-system. Then I(Q(S)) ∼ =I S. Proof. (1) Let F (O) be the set of order ﬁlters of O. Thus F (O) = ΩQ (Q(O)) (i. e. ΩQ (O ))), so that we know that F(O) can be made into the distributive lattice SatQ (Q(O)). Then let J (SatQ (Q(O))) be the set of co-prime elements of SatQ (Q(O)). Notice that co-prime elements have the form ↑R x, i. e. R (x), for some element x ∈ G and that they may be understood as properties fulﬁlled by the elements of G such that g x only if x, g ∈ R. Let us then deﬁne I(O) I(O) as G, J (SatQ (Q(O))), . Thus, g, g ∈ RI(O) iﬀ g ∈ Qg , iﬀ i(g) ⊆ i(g ), iﬀ g ∈ R (x) =⇒ g ∈ R (x) for all R (x) ∈ J (SatQ (Q(O))). In particular, since R is reﬂexive, g ∈ R (g) so that g ∈ R (g) holds, i. e. g, g ∈ R. Conversely, if g, g ∈ R and x, g ∈ R, for transitivity x, g ∈ R too. It follows that g ∈ R (x) g ∈ R (x), all x ∈ G. (2) For Q(S) is a preorder, from the previous point we have Q(I(Q(S))) ∼ =I Q(S) so that trivially I(Q(S)) ∼ =I S. EXAMPLE 3 Consider the P-system P and the A-system A of Example 2. Here below we display the lattices SatQ (P) and SatQ (A): SatQ (P)

SatQ (A)

G , ll G ,, l " ll " {a1 , a2 , a3 } {a, a1 , a2 } " l " ll , ll {a, a1 , a2 } {a, a2 , a3 } {a1 , a2 } , l , l " l ll "" " {a1 , a2 } {a, a2 } " " ll l " " ll , ll {a, a2 } {a1 } {a3 } l ,, l " l " {a2 } {a} ll " " ll , ∅ l ,, ∅

It is easy to verify that both of them are distributive lattices and that, moreover, SatQ (A) is a Boolean algebra. Indeed, for instance, the element QA {a,a1 } = A A Q = Q = {a {a, a1 , a2 } is complemented by the element z∈Q 3 }. z / A {a3 } {a,a1 }

288

P. Pagliani and M.K. Chakraborty

We can straightforwardly verify that Satcl (Q(P)) is Satint (P) plus some missed elements which are the diﬀerence between SatQ (P) and Satint (P). Indeed, the missed element is {a1 , a2 } which equals clQ(P) ({a1 , a2 }). On the contrary, intP ({a1 , a2 }) = ∅.

9

Generalised Topological Approximation Operators

In view of the duality between Information Systems and preorders, we can develop the rest of the theory from a more abstract point of view. Thus, from now on we shall deal with preordered structures and assume, intuitively, that they represent some information quantum relation system. Corollary 13. Let O = G, G, R be a preordered set. Let X ⊆ G. Then: The application is the least ﬁxpoint of including X R [R ] A int R(X) R [R] cl C R (X) The application is the largest ﬁxpoint of included in X [R] R C cl [R](X) [R ] R int A [R ](X) Proof. First notice that from Proposition 11 the listed ﬁxpoints collapse. Nonetheless it is worthwhile proving the ﬁrst two cases by means of two diﬀerent approaches. (a) Obviously R(X) ⊇ X and for idempotence R(X) is a ﬁxed point of R. Suppose Z is a ﬁxed point of R such that X ⊆ Z. From monotonicity R(X) ⊆ R(Z) = Z. Hence R(X) is the least ﬁxpoint of R including X. (b) From Proposition 6.3 [R]R (X) is the smallest image of [R] greater than or equal to X. Since [R] is idempotent it is the least ﬁxpoint of [R] which includes X and from Proposition 11, it is also the least ﬁxpoint of R (X) including X. The remaining cases are proved analogously. Corollary 14. Let O = G, G, R be a preordered set. Then for all X ⊆ G, (i) R(X) = {Z : Z ∈ ΩA (O) & Z ⊇ X}; (ii) [R](X) = {Z : Z ∈ ΓC (O) & Z ⊆ X}; (iii) R (X) = {Z : Z ∈ Γcl (O) & Z ⊇ X}; (iv) [R ](X) = {Z : Z ∈ Ωint (O) & Z ⊆ X}. Henceforth, for obvious reasons we shall adopt the following terminology: R(X) direct upper R-approximation of X, also denoted with (uR)(X) R (X) inverse upper R-approximation of X, also denoted with (uR )(X) [R](X) direct lower R-approximation of X, also denoted with (lR)(X) [R ](X) inverse lower R-approximation of X, also denoted with (lR )(X)

Formal Topology and Information Systems

289

The information-oriented reading of the above operators is: R(X)

Set of the elements specialised by some member of X (or, which approximate some member of X) R (X) Set of the elements approximated by some member of X (or, which specialise some member of X) [R](X) Set of the elements specialised only by members of X (or, which approximate only members of X) [R ](X) Set of elements approximated only by members of X (or, which specialise only elements of X) Particularly we can give an information-oriented interpretation to some combinations of operators: Set of the elements which are specialised just by elements specialised by some member of X (x ∈ [R]R(X) only if each element which specialises x is specialised by some member of X) Set of the elements which are approximated just by elements [R ]R (X) approximated by some member of X (x ∈ [R ]R (X) only if each element which approximates x is approximated by some member of X )

[R]R(X)

Besides these operators we add also the interpretation of [[R]] and [[R ]]: [[R]](X)

Set of the elements specialised by all the members of X (or, which approximate all the members of X) [[R ]](X) Set of the elements approximated by all the members of X (or, which specialise all the members of X) 9.1

Topological Approximation Spaces

Eventually we deﬁne some interesting examples of topological Approximation Spaces. Deﬁnition 14. Let Q(S) = G, G, RS be an IQRS. Then, 1. 2. 3. 4.

G, [RS ], RS - will be called a Direct Intuitionistic Approximation Space. G, [RS ], RS - will be called an Inverse Intuitionistic Approximation Space. G, [RS ], RS - will be called a Galois Intuitionistic Approximation Space. G, [RS ], RS - will be called a co-Galois Intuitionistic Approximation Space.

Deﬁnition 15. Let G, G, E be a relational structure such that E is an equivalence relation. Let I and C be the interior and, respectively, topological operators of the topological space induced by taking {[x]E : x ∈ G} as a subbasis. Then G, I, C is called a Pawlak Approximation Space From the above discussion the following statement is obvious: Proposition 15. Let E = G, G, E be a relational structure such that E is an equivalence relation. Then G, intE , clE is a Pawlak Approximation Space.

290

P. Pagliani and M.K. Chakraborty

But we can prove a further fact. To this end we introduce the notion of an Approximation Equivalence, or a-equivalence between (pre) topological Approximation Spaces: Deﬁnition 16. Let A = G, α, β and A = G, γ, δ two topological or pretopological Approximation Spaces. Then we say that A and A are a-equivalent, in symbols, A ∼ =a A if and only if Ωα (G) = Ωγ (G) and Γβ (G) = Γδ (G). Clearly, by duality one equality implies the other. We use this deﬁnition in the following statement: Proposition 16. Let S be an A-system or FP-system or DP-system. Let us set ♦ = RS and = [RS ]. Then ♦ = RS , = [RS ] and G, , ♦ is a Pawlak Approximation Space. Moreover, if S is an FP-system then G, , ♦ ∼ =a G, intS , clS . Proof. Immediate, from the fact, proved in Proposition 9.(3), that RS in this case is an equivalence relation. For the last part it is suﬃcient to use in addition Proposition 4.(3) together with Proposition 7, or the latter Proposition and Proposition 9.(3) which together with Proposition 8.(4) states that clS (g) = [g]kf , any g ∈ G. Note that the former system of the previous Proposition is, actually, G, G, , ♦ while the latter is G, M, intS , clS . Therefore, we cannot put G, , ♦ = G, intS , clS . However, if S is an FP-system, then Single-agent (pre)topological Approximation Spaces, and Pawlak Approximation Spaces induce the same family of ﬁxed points.

10

Comparing Information Systems

The notion of an i-quantum makes it possible to compare Information Systems. First of all we should ask whether it is possible to compare two quanta of information Qg and Qg . At ﬁrst sight we would say that Qg is ﬁner than Qg if Qg ⊆ Qg . However, this intuition works for P-systems, but not for A-systems because from Proposition 9.(2) if Qg ⊆ Qg then Qg ⊆ Qg . Thus non trivial comparisons of quanta of information in A-systems require a specialised notion of an i-quantum, which, in any case, is useful for P-systems too. Deﬁnition 17 (Relativised quanta of information) – Let A be an A-system. The quantum of information of g relative to a subset A ⊆ At is deﬁned as: Qg A = {g ∈ G : ∀a ∈ A, ∀x ∈ Va ((a(g) = x) =⇒ (a(g ) = x))}. – Let P be a P-system. The quantum of information of g relative to a subset A ⊆ M is deﬁned as: Qg A = {g ∈ G : ∀a ∈ A(g a =⇒ g a)}. Deﬁnition 18 (I-quantum dependence) Let S be an A-system or a P-system. Let A, A ⊆ At (or A, A ⊆ M ), g ∈ G.

Formal Topology and Information Systems

291

1. We say that A functionally depends on A at g, in symbols A →g A , if for all g ∈ G, g ∈ Qg A =⇒ g ∈ Qg A (that is, if Qg A ⊆ Qg A ). 2. We say that A functionally depends on A, in symbols A → A , if for all g ∈ G, A →g A . 3. If A → A and A → A, we say that A and A are informationally equivalent, A∼ =I A (thus, A ∼ =I A if for all g ∈ G, Qg A = Qg A ). So, a set of attributes (properties) A functionally depends on a set of attributes (properties) A if A has a higher discriminatory capability than A . Clearly, if S is an A-system then the notion of an i-quantum dependence relation turns into the usual notion of a functional dependence. From now on, if X denotes the relation with co-domain restricted to X then with S X we shall denote the subsystem G, X, X. If S is an A-system and X ⊆ At, with S X we shall denote the subsystem G, X, {Va }a∈X . The following statement formalises the above intuitions with respect to iquantum relations: Proposition 17. Let S be an A-system or a P-system. Let A, A ⊆ At (A, A ⊆ M ) such that A → A . Then R(AA) ⊆ R(AA ) . Proof. The proof is immediate. Suppose A → A . Then for all g ∈ G, Qg A ⊆ Qg A , so that g, g ∈ R(AA) implies g, g ∈ R(AA ) . It follows that we can naturally extend the notion of a functional dependence in order to compare two sets X and X of properties or attributes from two distinct (property or attribute) systems S and S over the same set of points G. Thus, we can extend the notion of ”informational equivalence” to entire systems: Deﬁnition 19. Let S and S be A-systems or P-systems over the same set of points G. Let S and S be the sets of attributes (properties) of S and, respectively, S . We say that S and S are informationally equivalent, in symbols S ∼ =I S , if S and only if for any g ∈ G, Qg S = Qg S , that is, if and only if Qg = QS g . Informational equivalence tells something about the behaviour of cl and int: Proposition 18. Let P and P be P-systems and P ∼ G, =I P . Then for all x ∈ P P P P cl (x) = cl (x). If both cl and cl are topological, then clP (X) = clP (X) and intP (X) = intP (X), for any X ⊆ G.

Proof. Suppose clP (x) = clP (x). Then there is g ∈ G such that, say, g ∈ clP (x) and g ∈ / clP (x). It follows that (g) ⊆ (x) but (g) (x). Thus x ∈ QP (g) and x ∈ / QP (g), so that P ∼ operators =I P . If both closure are additive, then by easy induction we obtain that clP (X) = clP (X) for any X ⊆ G. Moreover, suppose intP (X) = intP (X). Then −intP (X) = −intP (X), so that clP (−X) = clP (−X) - contradiction.

Notice that if either clP or clP is not topological then the equality between clP and clP is guaranteed just for singletons so that clP (−X) = clP (−X) is not

292

P. Pagliani and M.K. Chakraborty

a contradiction. Moreover, we can have P and P such that intP (x) = intP (x) but still P ∼ =I P . Therefore, the relation ∼ =I is far to be considered the ”best” way to compare P-systems, though very useful for our purposes. Now we want to stress the fact that we can compare not only the informational behaviour of the same point g with respect two diﬀerent sets of properties (attributes) X and X , but we can also compare the behaviours of two diﬀerent points g and g with respect to the same set of properties (attributes) P . Deﬁnition 20. Let S be an A-system or a P-system, X ⊆ M (or X ⊆ At) and g, g ∈ G. 1. We say that g is an X-specialisation of g (or that g is an X-approximation of g), in symbols g X g, if and only if the following condition holds: ∀x ∈ G(g ∈ Qx X =⇒ g ∈ Qx X). 2. We say that g is a specialisation of g , g g, if and only if g M g. Since for q-reﬂexivity x ∈ Qx , any x ∈ G, if g X g then g ∈ Qg X, so that g X g says that g fulﬁlls at least all the properties from X that are fulﬁlled by g . Therefore, g g implies g , g ∈ RS . Conversely, if g , g ∈ RS then g ∈ Qg . Hence g ∈ Qx implies g ∈ Qx , any x ∈ G, from transitivity of RS . It follows that the two relations and RS coincide. In fact they are the same instance of the usual topological notion of a specialisation preorder. In view of Proposition 9.(1) we can construct a topological space G, ImQ on G whose specialisation preorder is indeed (that is, RS ).

11

Transforming Perception Systems

Now we are equipped with a suﬃcient machinery in order to compare transformed systems. Let A be an A-system. To get a P-system out of A, the basic step derives from the observation that any attribute a is actually a set of properties, namely the possible attribute values for a. Thus we start associating each attribute a with the family N (a) = {av }v∈Va . We set N (At) = a∈At N (a). For each value v, av is the property ”taking value v for attribute a”. This transform is usually called a ”scale nominalisation”. Now let us set a relation N as: g N av if and only if a(g) = v, all g ∈ G, a ∈ At, v ∈ Va . We call the resulting system, N (A) = G, N (At), N , the ”nominalisation of A”. N (A) will be called a nominal A-system or NA-system. Proposition 19. Let A be an A-system. Then: (a) N (A) is a P-system; (b) N (A) ∼ =I A. Proof. (a) is obvious. (b) Let us prove that for any g ∈ G, Qg At = Qg N (At). Indeed, if g ∈ Qg At, then a(g) = x if and only if a(g ) = x, all a ∈ At.

Formal Topology and Information Systems

293

Therefore for any x ∈ N (a), g ax if and only if g ax , whence g ∈ Qg N (a). Finally, g ax for any other x = x, so we have the reverse implication. Moreover, if we formally consider P-systems as binary A-systems, we can also nominalise P-systems. But in this case we have a further property: Proposition 20. Let P be a P-system. Then N (P) is a dichotomic system. Proof. This is obvious, because for any property p, the nominalisation N (p) = {p1 , p0 } forms a pair of complementary properties, since for all g ∈ G, g N p1 if and only if g p and g N p0 if and only if g p. Nominalisation of dichotomic or functional systems does not give rise to any further result. Proposition 21. If P is a DP system or an FP system, then N (P) ∼ =I P. Proof. If P is dichotomic let p, p be a pair of complementary properties. After nominalisation we shall obtain two pairs N (p) = {p1 , p0 } and N (p) = {p1 , p0 }. Clearly, for any g ∈ G, g p in P if and only if g N p1 in N (P). But g N p1 if and only if g N p0 if and only if g N p0 . Conversely, g p if and only if g N p0 if and only if g N p1 if and only if g N p1 . If P is functional and g ∈ Qg M then g m if and only if g m, since (g) = (g ) = m. Thus the proof runs as in Proposition 19.(b). For N (A) is not only a P-system but it is still an A-system with At = {0, 1}, we obtain the following corollary: Corollary 15. Let S be an A-system or a P-system. Then N (S) ∼ =I N (N (S)). Proof. If S is a P-system then N (S) is a dichotomic systems so that from Proposition 21 N (N (S)) ∼ =I N (S). If S is an A-system then N (S) is a binary A-system and from Proposition 19.(b) N (S) ∼ =I N (N (S)). Corollary 16. If A is an A-system then there is a dichotomic system D such that D ∼ =I A. Proof. Since N (A) is a P-system, from Proposition 20 N (N (A)) is dichotomic. But from Proposition 19.(b) and Corollary 15 A ∼ =I N (A) ∼ =I N (N (A)). As a side result we again obtain Proposition 9.(2). Notice that this Proposition, as well as Corollary 15, relies on the fact that we are dealing with deterministic A-systems so that either two objects converge on the same attribute-value, or they diverge, but not both. EXAMPLE 4 Here are some examples: a P-system P = G, M, , an FP-system F = G, M , fˆ and an A-system A = G, At, V over the same set G:

294

P. Pagliani and M.K. Chakraborty

a a a a

b b 11 01 01 00

b 0 0 1 0

b 0 1 1 1

fˆ a a a a

m m 1 0 0 1 1 0 0 0

m 0 0 0 1

a a a a

A A 1 b 0 c 1 b 3 f

A α α α δ

Considering the system P let A = {b, b } and B = {b , b }. Then Qa A = {a, a , a } while Qa B = {a }. It follows that B →a A. On the contrary, Qa A = {a, a , a } and Qa B = {a , a , a } are not comparable. Hence B → A does not hold. If we compare the above systems we notice what follows: P a) A ∼ I P because QA = a = {a, a } while Qa = {a}. Neither P → A because P A F ∼ Qa = {a , a } while Qa = {a }. b) F =I A, because for all g ∈ G, QA g = Qg . Let us now nominalise the systems A and P: N A a a a a

A0 0 1 0 0

A1 1 0 1 0

A3 0 0 0 1

Ab 1 0 1 0

Ad 0 1 0 0

Af 0 0 0 1

Aα 1 1 1 0

Aδ 0 0 0 1

N P a a a a

b1 1 0 0 0

b0 0 1 1 1

b1 1 1 1 0

b0 0 0 0 1

b1 0 0 1 0

b0 1 1 0 1

b 1 0 1 1 1

b 0 1 0 0 0

Thus N (A) = {A0 , A1 , A3 }, N (b) = {b1 , b0 } and so on. It is evident that, for N (A) N (A) instance, a ∈ Qa and a ∈ Qa . But the same happens already in A. N (P) A A = {a }. Indeed, Qa = {a, a } = Qa . On the contrary, QP a = {a , a } but Qa P In fact a ∈ Qa because it fulﬁlls all the properties fulﬁlled by a (i. e. b and b ) plus the additional property b . But in N (P) this latter fact prevents N (P) a from belonging to Qa , because property b splits into the pair b0 , b1 N and a N P b0 while a P b1 , what are mutually exclusive possibilities. If we further nominalise N (P) and split, for instance, b0 , b1 into b01 , b00 , b11 , b10 , it is obvious that the pairs b01 , b10 and b00 , b11 give the same information as b0 and, respectively, b1 . It is not diﬃcult to verify that RN (A) = RA so that N (A) ∼ =I Q(A). 11.1

Dichotomic, Functional and Nominal Systems

First notice that the reverse of Proposition 9.(2) does not hold. For instance, if P is such that G = {1, 2, 3, 4}, M = {A, B, C} and (1) = {A, B}, (2) = {A, B}, (3) = {B, C} and (4) = {B, C}, Qg is an equivalence class, any g ∈ G though P is neither dichotomic nor functional. Also, if A is an A-system, then N (A) is not necessarily dichotomic. However N (A) ∼ =I N (N (A)) which is dichotomic (see Corollary 15). Indeed, notice that N (N (A)) is informationally equivalent to the system deﬁned as follows: 1) For each av in N (A), if Va is not a singleton set ¬av = {av }v =v,v ∈Va , while if Va = {v} then set ¬av = {av }. We set P = {av }v∈Va ∪ {¬av }v∈Va . 2) For each g ∈ G set g ∗ ¬av if and only if g av and g ∗ av if and only if g av . Clearly ¬av is the complementary copy of av . Thus, 3) set S = G, P, ∗ . We can easily verify that S is a dichotomic system and that S ∼ =I N (A).

Formal Topology and Information Systems

295

In reversal, since for any P-system P, N (P) induces an equivalence relation, we can ask whether N (P) itself ”is”, in some form, an A-system. Indeed it is trivially an A-system with set of attributes values V = {0, 1} and such that m1 (g) = 1 iﬀ g m1 iﬀ g m0 iﬀ m0 (g) = 0 and m1 (g) = 0 iﬀ g m0 iﬀ g m1 iﬀ m0 (g) = 1, all m ∈ M and by trivial inspection one can verify that G, N (M ), {0, 1} ∼ =I N (P). Finally we discuss another natural equivalence. We know that if S is an Asystem, or a DP or a FP system then RS is an equivalence relation (see Proposition 9). Thus a question arises as how to deﬁne a functional system F (S) informationally equivalent to a given A or DP system S. The answer is simple. If S is a P-system consider it as an A-system. Any tuple t ∈ a∈At Va is a combination of attribute-values and has the form a1m , . . . , ajn . We setg ∗ t only if a1 (g) = aim for any ai ∈ At and aim ∈ t. The resulting system G, a∈At Va , ∗ is the required F (S). Indeed ∗ is a map because no g ∈ G can satisfy diﬀerent tuples. Thus RF (S) is an equivalence relation such that g, g ∈ RF (S) only if a(g) = a(g ) for all a ∈ At (or in M ). It follows that N (S) ∼ =I F (S) so that if S is dichotomic or it is an A-system then RS = RF (S) and S ∼ =I F (S).

12

Conclusions

We have seen how modal operators naturally arise from the satisfaction relation which links points and properties in a Property System. Combinations of two modal operators which fulﬁll a adjunction relations deﬁne pre-topological interior and closure operators, as studied in Formal Topology. Thus we have shown that approaching approximation problems by means of the mathematical machinery provided by Formal Topology and Galois Adjunction theory makes it possible to deﬁne well-founded generalization of the classical upper and lower approximation operators. Moreover Galois Adjunction theory provides a set of results that can be immediately applied to these operators, so that we have a good understanding of the structure of the system of their ﬁxed points (i. e. exact sets). We have also seen how to deﬁne higher order information systems, namely Information Quantum Relation Systems, from property systems in order to deﬁne topological (that is, continuous) approximation operators, through the notion of a ”quantum of information”. And we have shown when these operators coincide with the lower and upper approximations deﬁned in classical Rough Set Theory. Eventually, we have seen how we can make diﬀerent kinds of information systems, property systems and attribute systems, into a uniform theoretical framework, and control these manipulations by means of a particular notion of an ”informational equivalence” induced by the concept of quanta of information. This has practical consequences too. Indeed, the relational modal or/and topological operators that we have deﬁned over P-systems may be directly translated into extremely simple constructs of functional languages such as LISP or APL (see [16]), thus providing a sound implementation. Therefore, this approach directly links the logical interpretation of approximation operators to the manipulation of concrete data structures for it coherently embeds the con-

296

P. Pagliani and M.K. Chakraborty

crete operations on Boolean matrices into a very general logical framework (the same relational interpretation of a modal operator applies to any sort of binary Kripke frame).

References 1. M. Banerjee & M. Chakraborty, ”Rough Sets Through Algebraic Logic”. Fundamenta Informaticae, XXVIII, 1996, pp. 211-221. 2. Banerjee, M., Chakraborty, M. K.: Foundations of Vagueness: a Category-theoretic Approach. In Electronic Notes in Theoretical Comp. Sc., 82 (4), 2003. 3. Barwise; J., Seligman, J.: Information Flow: the Logic of Distributed Systems. Cambridge University Press, Cambridge, 1997. 4. Bell, J. L.: Orthologic, Forcing, and the Manifestation of Attributes” In C. TR. Chong & M. J. Wiks (Eds.): Proc. Southeast Asian Conf. on Logic. North-Holland, 1983, pp. 13-36. 5. Blyth, T. S., Janowitz, M. F.: Residuation Theory. Pergamon Press, 1972. 6. Chakraborty, M. K., Banerjee, M.: Dialogue in Rough Context. Fourth International Conference on Rough Sets and Current Trends in Computing 2004 (RSCTC’2004), June 1-June 5, 2004, Uppsala, Sweden. 7. D¨ untsch, I., Gegida, G.: Modal-style operators in qualitative data analysis. Proc. of the 2002 IEEE Int. Conf. on Data Mining, 2002, pp. 155-162. 8. D¨ untsch, I., Orlowska, E.: Mixing modal and suﬃciency operators. In Bulletin of the Section of Logic, Polish Academy of Sciences, 28, 1999, pp. 99-106. 9. Gierz, G., Hofmann, K. H., Keimel, K., Lawson, J. D., Mislove, M. and Scott, D. S.: A compendium of Continuous Lattices. Springer-Verlag, 1980. 10. Humberstone, I. L.: Inaccessible worlds. In Notre Dame Journal of Formal Logic, 24 (3), 1983, pp. 346-352. 11. Lin, T. Y.: Granular Computing on Binary Relations. I: Data Mining and Neighborhood Systems. II: Rough Set Representation and Belief Functions. In Polkowski L. & Skowron A. (Eds.): Rough Sets in Knowledge Discovery. 1: Methodology and Applications, Physica-Verlag, 1998, pp.107-121 and 122-140. 12. Orlowska, E.: Logic for nondeterministic information. In Studia Logica, XLIV, 1985, pp. 93-102. 13. Pagliani, P.: From Concept Lattices to Approximation spaces: Algebraic Structures of some Spaces of Partial Objects. In Fund. Informaticae, 18 (1), 1993, pp. 1-25. 14. P. Pagliani: A pure logic-algebraic analysis on rough top and rough bottom equalities. In W. P. Ziarko (Ed.): Rough Sets, Fuzzy Sets and Knowledge Discovery, Proc. of the Int. Workshop on Rough Sets and Knowledge Discovery, Banﬀ, October 1993. Springer-Verlag, 1994, pp. 227-236. 15. Pagliani, P.: Rough Set Systems and Logic-algebraic Structures. In E. Orlowska (Ed.): Incomplete Information: Rough Set Analysis, Physica Verlag, 1997, pp. 109-190. 16. Pagliani, P.: Modalizing Relations by means of Relations: a general framework for two basic approaches to Knowledge Discovery in Database. In Proc. of the International Conference on Information Processing and Management of Uncertainty in Knowledge-Based Systems. IPMU 98, July, 6-10, 1998. ”La Sorbonne”, Paris, France, pp. 1175-1182. 17. Pagliani, P.: A practical introduction to the modal relational approach to Approximation Spaces. In A. Skowron (Ed.): Rough Sets in Knowledge Discovery. Physica-Verlag, 1998, pp. 209-232.

Formal Topology and Information Systems

297

18. Pagliani, P.: Concrete neighbourhood systems and formal pretopological spaces (draft). Conference held at the Calcutta Logical Circle Conference on Logic and Artiﬁcial Intelligence. Calcutta October 13-16, 2003. 19. Pagliani, P.: Pretopology and Dynamic Spaces. In Proc. of RSFSGRC’03, Chongqing, R. P. China 2003. Extended version in Fundamenta Informaticae, 59(2-3), 2004, pp. 221-239. 20. Pagliani, P.: Transforming Information Systems. In Proc. of RSFDGrC 2005, Vol. I, pp. 660-670. 21. Pagliani, P., Chakraborty, M. K.: Information Quanta and Approximation Spaces. I: Non-classical approximation operators. In Proc. of the IEEE Int. Conf. on Granular Computing. Beijing, R. P. China 2005, pp. 605-610. 22. Pagliani, P., Chakraborty, M. K.: Information Quanta and Approximation Spaces. II: Generalised approximation operators. In Proc. of the IEEE Int. Conf. on Granular Computing. Beijing, R. P. China 2005, pp. 611-616. 23. Pawlak, Z.: Rough Sets: A Theoretical Approach to Reasoning about Data. Kluwer, 1991. 24. Polkowski, L.: Rough Sets: Mathematical Foundations. Advances in Soft Computing, Physica-Verlag, 2002. 25. Sambin, G.: Intuitionistic formal spaces and their neighbourhood. In Ferro, Bonotto, Valentini and Zanardo (Eds.) Logic Colloquium ’88, Elsevier (NorthHolland), 1989, pp. 261-285. 26. Sambin, G., Gebellato, S.: A Preview of the Basic Picture: A New Perspective on Formal Topology. In TYPES 1998, pp. 194-207. 27. Sambin, G.: Formal topology and domains. In Proc. of the Workshop on Domains, IV. Informatik-Bericht, Nr. 99-01, Universit¨ at GH Siegen, 1999. 28. Smyth, M.: Powerdomains and predicate transformers: a topological view. In J. Diaz (Ed.) Automata, languages and Programming. Springer LNCS, 154, 1983, pp. 662-675. 29. Skowron, A. & Stepaniuk, J.: Tolerance Approximation Spaces. Fundamenta Informaticae, 27 (2-3), IOS Press, 1996, pp. 245-253. 30. Skowron, A., Swiniarski, R. & Synak, P.: Approximation Spaces and Information Granulation. Transactions on Rough Sets III, LNCS 3400, Springer, 2005, pp. 175-189. 31. Skowron, A., Stepaniuk, J., Peters, J. F. & Swiniarski, R.: Calculi of approximation spaces. Fundamenta Informaticae, 72 (1-3), 2006, pp. 363–378. 32. Vakarelov, D.: Information systems, similarity relations and modal logics. In E. Orlowska (Ed.) Incomplete Information - Rough Set Analysis Physica-Verlag, 1997, pp. 492-550. 33. Vickers, S.: Topology via Logic. Cambridge University Press, 1989. 34. Wille, R.: Restructuring Lattice Theory. In I. Rival (Ed.): Ordered Sets, NATO ASI Series 83, Reidel, 1982, pp. 445-470. 35. Yao, Y. Y.: Granular computing using neighborhood systems. In R. Roy, T. Fumhashi, and P.K. Chawdhry (Eds.): Advances in Soft Computing: Engineering Design and Manufacturing, Springer-Verlag, London, U.K., 1999. 36. Yao, Y. Y.: A comparative study of formal concept analysis and rough set theory in data analysis. Manuscript 2004. 37. Zhang, G. -Q.: Chu spaces, concept lattices and domains. In Proc. of the 19th Conf. on the Mathematical Found. of Programming Semantics. March 2003, Montreal, Canada. Electronic Notes in Theor. Comp. Sc., Vol. 83, 2004.

On Conjugate Information Systems: A Proposition on How to Learn Concepts in Humane Sciences by Means of Rough Set Theory Maria Semeniuk–Polkowska Chair of Formal Linguistics, Warsaw University Browarna 8/12, 00991 Warsaw, Poland m polkowski@hotmail.com

To the memory of Professor Zdzislaw Pawlak Abstract. Rough sets, the notion introduced by Zdzislaw Pawlak in early 80’s and developed subsequently by many researchers, have proved their usefulness in many problems of Approximate Reasoning, Data Mining, Decision Making. Inducing knowledge from data tables with data in either symbolic or numeric form, rests on computations of dependencies among groups of attributes, and it is a well–developed part of the rough set theory. Recently, some works have been devoted to problems of concept learning in humane sciences via rough sets. This problem is distinct as to its nature from learning from data, as it does involve a dialogue between the teacher and the pupil in order to explain the meaning of a concept whose meaning is subjective, vague and often initially obscure, through a series of interchanges, corrections of inappropriate choices, explanations of reasons for corrections, ﬁnally reaching a point, where the pupil has mastered enough knowledge of the subject to be able in future to solve related problems fairly satisfactorily. We propose here an approach to the problem of learning concepts in humane sciences based on the notion of a conjugate system; it is a family of information systems, organized by means of certain requirements in order to allow a group of students and a teacher to analyze a common universe ofobjects and to correct faulty choices of attribute value in order to reach a more correct understanding of the concept. Keywords: learning of cognitive concepts, rough sets, information systems, conjugate information systems.

1

Introduction

In addition to a constant ﬂux of research papers on inducing knowledge from data expressed in either symbolic or numerical form, there are recently papers on learning cognitive concepts by means of the rough set theory, see, e.g., [2], [12], [13], [14], [15]. We propose in this work an approach to the problem of learning/teaching of concepts in humane sciences that stems from an analysis of the process of J.F. Peters et al. (Eds.): Transactions on Rough Sets VI, LNCS 4374, pp. 298–307, 2007. c Springer-Verlag Berlin Heidelberg 2007

On Conjugate Information Systems

299

learning in humane sciences, and of learning approach in library sciences, in particular, that has been worked out during our seminars at Warsaw University, [12], [13], [14], [15]. In the process of learning of humane concepts, in general, a teacher, a tutor, is directing the pupil, the student, toward understanding of the problem, and toward its correct solutions, by means of a dialogue that involves expositions, responses, corrections, explanations etc., etc., aimed at developing a satisfactory understanding of the concept meaning by the student. In order to formally render this mechanism and to study this problem, we recall here a notion of a conjugate information system introduced in [14] (under the name of SP–systems), and discussed shortly in [12]. The main idea underlying this approach can be introduced as follows: 1. Both the tutor and the student are equipped with information/decision systems (see Sect. 2, for all relevant notions of the rough set theory) that possess identical sets of attributes and the same universe of objects, and they differ from each other only in value assignment to attributes; the assumption is, that the values of attributes are correctly assigned in the tutor system whereas the student can initially assign those values incorrectly, which results in a faulty classiﬁcation of objects to decision classes, and those values are gradually corrected during the interchange of messages with the tutor; 2. In order to learn a correct assignment of values to attributes, the student has also in his disposal a family of auxiliary decision systems, one for each attribute. Attributes in those decision systems are for simplicity (and, actually, in conformity with the practice of learning in many parts of humane sciences) assumed to be Boolean; this means that the value of an attribute on an object is selected on the basis of whether the object has/has not given Boolean features (for instance, when deciding whether romance books should be acquired for a library, one may look at the feature: majority/minority in a poll opted for romance books in the library). In what follows, we present a formal framework for conjugate information systems. We refrain from discussing here the interface between the tutor and the pupil, being satisﬁed with presenting the formal apparatus of conjugate information systems. Our methodology presented in what follows was tested in our courses given to the students in the Department of Library and Information Sciences at the University of Warsaw. The author does express gratitude to her students to whom she is indebted for many works on applying the ideas presented in this paper.

2

Auxiliary Notions of Rough Set Theory

All basic notions relevant to rough sets may be found in [5], [9], or in [3]. We recall here, for the reader’s convenience, some basic notions that are used in the sequel.

300

M. Semeniuk–Polkowska

2.1

Information Systems

An information system A is deﬁned as a triple (U, A, h) where: 1. U is a ﬁnite set of objects; 2. A is a ﬁnite set of conditional attributes. In the sequel, we will use the term attribute instead of the term conditional attribute. Attributes act on objects, to each pair of the form (a, u), where a is an attribute and u is an object, a value a(u) is assigned. In our setting, we wish to work with systems where U and A are ﬁxed, but value assignments are distinct, hence the need for the component h, representing the value assignment in a given information system; 3. A mapping h : U × A → V , with h(u, a) = a(u), is an A–value assignment, where V = {Va : a ∈ A} is the attribute value set. 2.2

Decision Systems

Decision systems are a variant of information systems in which a new attribute d∈ / A, called the decision, is introduced; formally, a decision system is a quadruple (U, A, d, h) where U, A are as in sect. 2.1, d : U → Vd is the decision with values in the decision value set Vd , and the value assignment h does encompass d, i.e., h : U × (A ∪ {d}) → V ∪ Vd with the obvious proviso that values of h on pairs of the form (d, u) belong in Vd . 2.3

Indiscernibility and Its Extensions

The crucial notion on which the idea of rough sets does hinge is that of the indiscernibility relation [5], [6]. For an information system I = (U, A, h), the indiscernibility relation IN DI (B), induced by a set B ⊆ A of attributes, is deﬁned as follows, IN DI (B) = {(u, v) : h(a, u) = h(a, v) for a ∈ B},

(1)

and equivalence classes of IN DI (B) generate by means of the set–theoretical operation of the union of sets the family of B–exact sets (or, concepts); concepts that are not B–exact are called B–rough. Rough set theory deals with constructs that are invariant with respect to indiscernibility relations hence they can be expressed in terms of indiscernibility classes. Indiscernibility classes may be generalized to μ–granules, where μ is a rough inclusion [11]. Rough inclusions are relations of the form μ(u, v, r) read as: ”u is a part of v to degree at least r”. Examples of rough inclusions and deeper applications can be found, e.g., in [10]; let us quote from there an example of the rough ukasiewicz t–norm tL (x, y) = max{0, x+y−1} by inclusion μL induced from the L ) ≥ r in which DIS(u, v) = {a ∈ means of the formula μL (u, v, r) ⇔ g( |DIS(u,v)| |A| A : h(a, u) = h(a, v)}, |A| stands for the cardinality of A, and g is a function from

On Conjugate Information Systems

301

the reals into [0, 1] that ﬁgures in the representation : tL (x, y) = g(f (x) + f (y) (see,e.g., [7]). As g(x) = 1 − x, the formula for μL is ﬁnally: μL (u, v, r) iﬀ

|IN D(u, v)| ≥ r. |A|

(2)

In case r = 1, one obtains from (2) indiscernibility classes as sets (granules) of the form g(u)1 = {v : μ(u, v, 1)}; for r < 1, one obtains a collection of granules being unions of indiscernibility classes with respect to various sets of attributes. 2.4

Approximations to Rough Concepts

One more crucial notion due to Zdzislaw Pawlak is the notion of an approximation. In the classical case, any set (concept) X ⊆ U is approximated with indiscernibility classes [u]B of the relation IN DI (B), where B ⊆ A, from below (the lower approximation) and from above (the upper approximation): (3) BX = {[u]B : [u]B ⊆ X}, BX =

{[u]B : [u]B ∩ X = ∅}.

(4)

More generally, one can replace in above deﬁnitions classes [u]B with the μ–granules g(u)r of a ﬁxed or subject to some conditions radius r.

3

Conjugate Information Systems

The notion of a conjugate information system reﬂects the mechanism of learning in the interaction between the tutor and the student (or, students). In this process, the correct evaluation scheme is transferred from the tutor to students, who initially may have incorrect evaluation schemes and gradually learn better evaluations in order to ﬁnally come up with schemes satisfactorily close to the correct one. 3.1

On Conjugate Information Systems

By a conjugate information system, we understand a triple, d : a ∈ A, i ∈ I}, i0 ), C = ({Ai = (Ui , Ai , hi ) : i ∈ I}, {Fa,i

(5)

where I is a set of participants in the learning process, with i0 denoting the tutor and i ∈ I \ {i0 } denoting students, consisting of: 1. a family of information systems {Ai = (Ui , Ai , hi ) : i ∈ I} such that for some ﬁnite sets U, A we have Ui = U, Ai = A for i ∈ I; d 2. a family of decision systems {Fa,i : a ∈ A and i ∈ I}. Thus the diﬀerence between information systems Ai , Aj with i = j, i, j ∈ I is in functional assignments hi , hj . The information system corresponding to i0 is said to be the tutor system;

302

M. Semeniuk–Polkowska

d 3. for each pair (a, i), a ∈ A, i ∈ I, the decision system Fa,i is a decision system (U, F eata , a, ha,i ), where U is the C–universe, F eata is the set of a–features, each f ∈ F eata being a binary attribute, a is the decision attribute of the d , and ha,i is the value assignment. system Fa,i d will be regarded as the system belonging to the tutor, The system Fa,i 0 d while its realization by an agent i ∈ I, i = i0 will be the system Fa,i of the student i for the evaluation of the attribute a ∈ A. d An Assumption. We assume that each system Fa,i is reduced in the sense that for each value va of the attribute a, there exists at most one object u with the property that ha,i (u, a) = va . In the case when such an object u exists (and then, by our assumption, it is unique) we will denote by h← a,i (va ) the information vector (f (u) : f ∈ F eata ), i.e.,

h← a,i (va ) = (f (u) : f ∈ F eata ).

(6)

← The symbol (h← a,i (va ))f will denote the f −th coordinate of the vector ha,i (va ). We deliberately omit the communication aspect of the process; formally, its presence could be marked with some mappings between the corresponding systems. However, we deem it unnecessarily complicating the picture for the purpose of this paper.

3.2

A Metric on Conjugate Systems

It is now important to introduce some means for organizing the unordered as of now set of participants in the learning process; to this end, we exploit the idea in [12] of some distance function on a conjugate system. We introduce a distance function dist on the conjugate system C. To this end, we ﬁrst introduce for an object u ∈ U the set, DISi,j (u) = {a ∈ A : hi (a, u) = hj (a, u)},

(7)

of attributes discerning on u between systems Ai and Aj . Thus, DISi,j (u) collects attributes which are assigned distinct values on the object u by students i = j. Now, we let, (8) dist(Ai , Aj ) = maxu |DISi,j (u)|. The function dist(Ai , Aj ) is a pseudo–metric, i.e., it has all properties of a metric (see, e.g., [1]) except for the fact that its value may be 0 whereas the arguments may be formally distinct as discerned by distinct indices i and j; we oﬀer a simple argument showing that dist(., .) is a pseudo–metric. Only the triangle inequality may need a proof. Thus, assume that information systems Ai , Aj , and Ap are given. If hi (a, u) = hj (a, u) and hj (a, u) = hp (a, u) then hi (a, u) = hp (a, u); thus, hi (a, u) = hp (a, u) implies that either hi (a, u) = hj (a, u) or hj (a, u) = hp (a, u).

On Conjugate Information Systems

303

In consequence, DISi,p (u) ⊆ DISi,j (u) ∪ DISj,p (u),

(9)

and from (9) one gets that, maxu |DISi,p (u)| ≤ maxu |DISi,j (u)| + maxu |DISj,p (u)|.

(10)

The formula (10) is the required triangle equality. Learning starts with the pupil(s) closest to the tutor, and continues in the decreasing order of the distance. 3.3

Basic Parameters for Learning

At the learning stage, each agent Aj (represented by the corresponding information system) learns to assign values of attributes in its set A from features in d decision systems {Fa,j : a ∈ A}. Parameters for Learning Feature Values and Attribute Values. First, at the training stage, each agent student learns to assign correct values to features in sets F eata = {fk,a : k = 1, 2, ..., na } for each attribute a ∈ A. We assume that values at the tutor system are already established as correct. The measure of learning quality is the assurance–level–function mj (k, a); for each triple (j, k, a), where j ∈ I \ {i0 }, k ≤ na , and a ∈ A, it is deﬁned as follows: mj (k, a) =

pos(j, k, a) , ex(j, k, a)

(11)

where ex(j, k, a) is the number of examples for learning fk,a and pos(j, k, a) is the number of positively classiﬁed examples in the set U by the agent Aj. The process of learning, as mentioned above, proceeds in a dialogue between the tutor and the student, aimed at explaining the meaning of the attribute a, and its dependence on features in the set F eata ; after that discussion, the j − th student proceeds with assigning values to features for objects from the universe U , in the training sample. The assignment is evaluated by the tutor and on the basis of that evaluation, assurance levels are calculated, to judge the understanding of the pupil. d according to decreasing value of mj (k, a); the reWe order features in Fa,j d sulting linear order is denoted ρaj , and the system Fa,j with values assigned by d the agent Aj is denoted with the symbol Fa,j (ρ). Metrics on Value Sets of Attributes. We set a distance function φja (v, w) on values of the attribute a for each a ∈ A, v, w ∈ Va , and j ∈ I, estimated in d the system Fa,j by letting,

304

M. Semeniuk–Polkowska

where,

φja (v, w) = |DISa,j (v, w)|,

(12)

← DISa,j (v, w) = {f ∈ F eata : (h← a,j (v))f = (ha,j (w))f }.

(13)

d in This deﬁnitions are possible, due to our assumption about systems Fa,j Sect.3.1. Thus, φja (v, w) does express the distance at the pair v, w of values of the attribute a measured as the number of diﬀerently classiﬁed features in the row deﬁned by v, w, respectively.

4

Learning of Attribute Values

We now address the problem of learning from the tutor of the correct evaluation of attribute values. Objects u ∈ U are sent to each agent Ai for i ∈ I one–by–one. Step 1. Assignment of attribute values based on training examples. At that stage the values dist(i, i0 ) of distances from agents Ai to the tutor Ai0 are calculated. Step 2. The feedback information passed from the tutor to the agent Ai is the following: Infi = (r, Error− set− i = {ai1 , ...aipr }, Error− vector− i = [si1 , ..., sipr ]), (14) where: 1. r is the value of the distance dist (Ai0 , Ai ) from the student i to the tutor i0 ; 2. pr is the number of misclassiﬁed attributes in A between agents i, i0 . Clearly, p(r) ≤ |U | · r, depends on r; 3. aij is the j th misclassiﬁed attribute; 4. for j ∈ {1, .., pr }, the value sij is the distance φaij (vj , wj ) where vj is the correct (tutor) value of the attribute aij and wj is the value assigned to aij by the agent Ai . Step 3. The given agent Ai begins with the attribute a = aiError−seti for which the value of the assurance-level-function is maximal (eventually selected at random from attributes with this property). For the attribute a, the value s = sa is given, meaning that s × 100 percent of features has been assigned incorrect values by Ai in the process of determining the value of a. d (ρ) according to deStep 4. Features in F eata are now ordered into a set Fa,i creasing values of the assurance–level–function mi (k, a) i.e. by ρai : starting with the feature f = fFi,aeata giving the minimal value of the function mi (k, a), the agent i goes along the ordered set changing the value at subsequent nodes. If the value of φ remains unchanged after the change at the node, the error counter remains unchanged, otherwise its value is decremented/incremented by one.

On Conjugate Information Systems

305

Step 5. When the error counter reaches the value 0, stop and go to the next feature. Step 6. Go to the next attribute in the set A.

5

An Example

Our example is a simple case that concerns grading essays written by students in French, taken from [13], [12]. Grading is done on the basis of three attributes: a1 : grammar, a2 : structure, and a3 : creativity. We present below tables showing the tutor decision systems Fa for a = a1 , a2 , a3 . Example 1. Decision systems Fa1 , Fa2 , Fa3 Table 1. Decision systems Fa1 , Fa2 , Fa3 fa11 + -

fa21 + -

fa31 + + -

a1 3 2 1

fa12 +

fa22 + -

fa32 + -

a2 3 2 1

fa13 + +

fa23 + +

fa33 +

a3 3 2 1

where fa11 takes value +/- when the percent of declination errors is ≥ /< 20 ; fa21 is +/- when the percent of conjugation errors is ≥ / < 20, and fa31 is +/- when the percent of syntax errors is ≥ / < 20; fa12 takes value +/- when the structure is judged rich/not rich, fa22 is +/- when the structure is judged medium/not medium, and fa32 is +/- when the structure is judged to be weak/ not weak. fa13 takes value +/- when the lexicon is judged rich/not rich, fa23 is +/- when the source usage is judged extensive/not extensive, and fa33 is +/- when the analysis is judged to be deep/ not deep. Consider a pupil A1 and a testing information system with U = {t1 , t2 , t3 }, A = {a1 , a2 , a3 } which is completed with the following value assignments. Example 2. Information systems A0 of the tutor and A1 of the pupil. Table 2. Decision systems Fa1 , Fa2 , Fa3 t t1 t2 t3

a1 1 1 3

a2 2 1 2

a3 1 1 3

t t1 t2 t3

a1 1 1 3

a2 2 1 2

a3 2 2 3

The distance dist(A0 , A1 ) is equal to 1 as DIS0,1 (t1 ) = {a3 } = DIS0,1 (t2 ); DIS0,1 (t3 ) = ∅.

306

M. Semeniuk–Polkowska

Thus, the pupil misclassiﬁed the attribute a3 due to a faulty selection of feature values: in case of t1 , the selection by the tutor is +,+,+ and by the pupil: +,+,-. The distance φa3 ,1 is equal to 1 and the information sent to the pupil in case of t1 is Inf1 = (1, {a3 }, (1)). Assuming the values of assurance–level–function m(1, k, a3 ) are such that f 3,a3 = fa33 , the pupil starts with fa33 and error–counter =1 and changing the value at that node reduces the error to 0. This procedure is repeated with t2 etc.

6

Conclusion

We have presented a skeleton on which the mechanism of learning cognitive concepts can be developed. It has been the principal aim in this paper to show that the notion of a conjugate information system may be helpful in fulﬁlling this task as a model of dependence between the tutor and the student.

Acknowledgement The topic of the paper has been discussed at seminars conducted by the author at the Institute of Library and Information Sciences at Warsaw University. The author wishes to thank the participants in those seminars. Thanks go also to Professor Lech Polkowski for useful discussions and valuable help with the preparation of this note.

References ´ ements de Math´ematique. Topologie G´en´erale. Hermann, Paris, 1. Bourbaki, N.: El´ France (1960). 2. Dubois, V., Quafafou, M.: Concept learning with approximations: rough version spaces. In: Lecture Notes in Artiﬁcial Intelligence vol. 2475, Springer–Verlag, Berlin, Germany. (2002) 239–246. 3. Komorowski, J., Pawlak, Z., Polkowski, L., Skowron, A.: Rough sets: A tutorial. In: Pal, S. K., Skowron, A. (Eds.): Rough–Fuzzy Hybridization: A New Trend in Decision Making. Springer–Verlag, Singapore Pte. Ltd. (1999) 3–98. 4. Pal, S. K., Polkowski, L., Skowron, A. (Eds.): Rough-Neural Computing. Techniques for Computing with Words. Springer–Verlag, Berlin, Germany (2004). 5. Pawlak, Z.: Rough Sets: Theoretical Aspects of Reasoning about Data. Kluwer, Dordrecht, the Netherlands (1991). 6. Pawlak, Z.: Rough sets. International Journal of Computer and Information Science 11 (1982) 341–356. 7. Polkowski, L.: Rough Sets. Mathematical Foundations. Physica–Verlag, Heidelberg, Germany (2002). 8. Polkowski, L., Tsumoto, S., Lin, T. Y. (Eds.): Rough Set Methods and Applications. Physica–Verlag, Heidelberg, Germany (2000). 9. Polkowski, L., Skowron, A. (Eds.): Rough Sets in Knowledge Discovery 1,2. Physica–Verlag, Heidelberg, Germany (1998).

On Conjugate Information Systems

307

10. Polkowski, L., Semeniuk–Polkowska, M.: On rough set logics based on similarity relations. Fundamenta Informaticae 64 (2005) 379–390. 11. Polkowski, L., Skowron, A.: Rough mereology: A new paradigm for approximate reasoning. International Journal of Approximate Reasoning 15 (1997) 333-365. 12. Semeniuk-Polkowska, M., Polkowski, L.: Conjugate information systems: Learning cognitive concepts in rough set theory. In: Lecture Notes in Artiﬁcial Intelligence 2639, Springer–Verlag, Berlin, Germany. (2003) 255–259. 13. Semeniuk–Polkowska, M.: Applications of Rough Set Theory. Seminar Notes (in Polish), Fasc. II, III, IV. Warsaw University Press, Warsaw, Poland (2000–2002). 14. Semeniuk–Polkowska, M.: On Some Applications of Rough Sets in Library Sciences (in Polish). Warsaw University Press, Warsaw, Poland (1997). 15. St¸epie´ n, E.: A study of functional aspects of a public library by rough set techniques.PhD Thesis, Warsaw University, Department of Library and Information Sciences, M. Semeniuk-Polkowska, supervisor (2002).

Discovering Association Rules in Incomplete Transactional Databases Grzegorz Protaziuk and Henryk Rybinski Institute of Computer Science, Warsaw University of Technology gprotazi@ii.pw.edu.pl, hrb@ii.pw.edu.pl

Abstract. The problem of incomplete data in the data mining is well known. In the literature many solutions to deal with missing values in various knowledge discovery tasks were presented and discussed. In the area of association rules the problem was presented mainly in the context of relational data. However, the methods proposed for incomplete relational database can not be easily adapted to incomplete transactional data. In this paper we introduce postulates of a statistically justiﬁed approach to discovering rules from incomplete transactional data and present the new approach to this problem, satisfying the postulates. Keywords: association rules, frequent itemsets, incompleteness, transactional data.

1

Introduction

Very often one of the main restrictions in using data mining methodology is imperfection of data, which is a common fact in real-life databases, especially those exploited for a long period. Imperfection can be divided into several different categories: inconsistency, vagueness, uncertainty, imprecision and incompleteness [19]. In the paper we consider the problem of discovering knowledge from incomplete database. Within the knowledge discovery process the incompleteness of data can be taken into consideration at two stages, namely (1) at the preprocessing step, and (2) at the data mining step. The objective of (1) is to ﬁll missing values in order to pass to the next steps of the process and process data as they were complete. Here, one can use simple approaches, such as replacing unknown values by special ones (e.g. average or dominant value), as well as more advanced methods, such as e.g. completing data methods based on classiﬁers or sets of rules [8]. In the case of (2), missing or unknown values are subject of processing by the data mining algorithms. In the literature many such approaches for diﬀerent data mining tasks were introduced. In particular, the problem of classifying incomplete objects has been addressed in the context the rough set theory [20,21]. The main idea of the approach is based on the indiscernibility relation and lower and upper approximation of a given set X. Originally proposed for complete information system,

Research has been supported by the grant No 3 T11C 002 29 received from Polish Ministry of Education and Science.

J.F. Peters et al. (Eds.): Transactions on Rough Sets VI, LNCS 4374, pp. 308–328, 2007. c Springer-Verlag Berlin Heidelberg 2007

Discovering Association Rules in Incomplete Transactional Databases

309

the approach was successfully extended to deal with incomplete data. Various modiﬁcations have been proposed and discussed in the papers [10,16,28,29,30]. Yet another group of data mining algorithms dealing with incomplete data can be distinguished, the algorithms from this group are based on the methods for building decision tree. The modiﬁcation of the C4.5 algorithm was introduced in [25]. In the CART algorithm [6] the surrogate tests are used for dealing with missing values. Diﬀerent aspects of using decision trees for working with incomplete data have been presented in [9,17,32]. In the paper we concentrate on the algorithms discovering association rules from incomplete data sets. In [12] a notion of legitimate approach has been deﬁned. It consists in satisfying a set of postulates resulting from statistical properties of the support and conﬁdence parameters and being necessary conditions. We claim that any method dealing with incompleteness should satisfy the postulates of the approach in order to properly asses expected support and conﬁdence. The original deﬁnition of the postulates referred to the relational database. Here we generalize it, so that it also covers discovering association rules from transactional data. In addition we deﬁne a novel data mining algorithm (DARIT), very well suited for discovering rules from transactional databases and satisfying the postulates of the statistically justiﬁed approach. The rest of the paper is organized in the following manner. Section 2 formally deﬁnes association rules and their properties. Section 3 reviews the methods of discovering association rules from incomplete data. Section 4 presents the concept of support and conﬁdence under incompleteness. In Section 5 we discuss details of the presented DARIT algorithm. The results of experiments are presented in Section 6, whereas Section 7 concludes this paper.

2

Association Rules and Their Properties

Below we introduce basic notions necessary for analyzing the process of discovering rules from incomplete data. We consider two types of databases, namely transactional and relational ones. A transactional database, denoted as DT , consists of ﬁnite set of transactions, DT = {t1 , t2 , t3 , . . . , tn }. Each transaction has a unique identiﬁer and a non empty set of elements (items). Each element belongs to the ﬁnite set of items I = {elem1, elem2 , . . . , elemm}. A relational database, denoted by DR, is a ﬁnite set of records DR = {r1 , r2 , . . . , rk }. Each record consists of n scalar values, belonging to the domains of n attributes respectively. The set of attributes is further denoted by A. By k-itemset we denote a set of k items from the database. In the sequel, if we do not distinguish between relational and transactional database, we denote it by D. Association rules are one of the simplest and the most comprehensive forms for representing discovered knowledge. The problem of discovering association rules was ﬁrst deﬁned in [1], in the context of market basket data with the goal to identify customers’ buying habits. An exemplary association rule would state that 70% customers who buy bread also buy milk.

310

G. Protaziuk and H. Rybinski

The basic property of an itemset is its support. It is deﬁned as percentage of those transactions in the database D, which contain given itemset. It is referred to as a relative support, and is formally deﬁned as: support(X) =| {t ∈ D | X ⊆ t}|/ | D | .

(1)

where X – itemset, t – transaction or record. Sometimes the notion of absolute support of an itemset is used, which is deﬁned as a number of the transactions supporting given itemset. If a given transaction (record) includes an itemset X we say that the transaction (record) supports the itemset. Frequent itemset is an itemset with support not less than a given minimal level called minimal support and denoted by minSup. Association rule is an expression in the form: X → Y , where X, Y are itemsets over I and X = ∅, Y = ∅ and X ∩ Y = ∅. X is called an antecedent of the rule, Y is called a consequent of the rule. The support of the rule X → Y is equal to support(X ∪ Y ). Conﬁdence of the rule X → Y , denoted by conﬁdence(X → Y ), is deﬁned as: conf idence(X → Y ) = support(X → Y )/support(X).

(2)

The parameter minConf is deﬁned by the user and indicates minimal conﬁdence that the discovered rules should satisfy. The basic task concerning association rules is to ﬁnd all such rules which satisfy the minimal support and minimal conﬁdence requirements.

3

Related Works

The most known algorithm for discovering all association rules is the Apriori algorithm, proposed in [2]. In the algorithm candidate sets (potentially frequent itemsets) of the length k are generated from frequent itemsets of the length k −1. Another approach proposed in [11] is based on the special data structure called FP-tree (frequent pattern tree). Various aspects of discovering association rules are widely presented and discussed in the literature. In [3,15,24,27,31] the problem of generating only interesting rules with respect to additional measure is considered. In [4,5,23] the methods for discovering rules from dense data were introduced. The lossless concise representations of frequent patterns were proposed in [7,14,22]. However, the problem of discovering association rules from incomplete data is discussed relatively rarely, especially with respect to transactional data sets. In the case of missing values the fundamental problem is evaluating the real support of a given itemset, so that one can determine if the given itemset is frequent or not. In [26] the new deﬁnitions of support and conﬁdence of the rule were introduced for the relational databases with missing values. In the deﬁnitions, the notions of disabled data and valid database are used. A record r is disabled for a given set X if, it includes at least one unknown value for the attributes, for which there are elements in X. A set of disabled records for X is denoted by

Discovering Association Rules in Incomplete Transactional Databases

311

Dis(X). A valid database vdb for the set X consists only of those records from DR which are not disabled for X: vdb = DR\Dis(X). Given these notions the authors deﬁne support as: support(X) = |DX|/|vdb(X)|, where DX stands for a set of records containing the set X. The conﬁdence was deﬁned as follows: conﬁdence(X → Y ) = |DXY |/|DX | − |Dis(Y ) ∩ DX |). It is worth mentioning that the proposed deﬁnition of support may give rise to the situations in which a multielement set has greater support than its subsets. This drawback has been eliminated in [13], where the probabilistic approach was presented. The approach is dedicated to relational databases. It is based on the assumption that the missing values of an attribute do not depend on the values of other attributes. Given an attribute a, unknown values in all records are assigned a probability distribution over all known values from the domain of the attribute. Each value v has assigned a probability, denoted by prob(v, a), which is equal to its frequency of occurring in all records having known values for the considered attribute. The main idea of the approach is based of the notion of probable support. Probable support, denoted by probSupr , is calculated for the element elem(v, a) of the value v from the domain of the attribute a for single record in the following manner: 1 if r.a = v probSupr (elem(v, a)) = prob(v, a) if r.a is unknown 0 otherwise

(3)

where r.a stands for the value for the attribute a in the record r. The support probSupr of a set X = {elem(v1 , a1 ), elem(v2 , a2 ), . . . , elem(vk , ak )} is computed by the following formula: probSupr(X) = probSupr (elem(v1 , a1 )) ∗ probSupr (elem(v2 , a2 )) ∗ . . . ∗ (4) probSupr (elem(vk , ak )) A similar approach has been applied for transactional databases in [18] where the algorithm ∼AR has been proposed. The algorithm is a modiﬁcation of the well known Apriori algorithm. Also here the main idea is based on partial support of itemsets, in this case by transactions. The following way of calculating support of itemsets for single transaction was introduced: given k-itemset Z k , each element included in a transaction t and the set Z k contributes in 1/k to total value of support of the Z k set calculated for the transaction t. The total value of support of the Z k set is computed by summing up all values contributed by elements included in the transaction. If the set Z k is contained in the transaction t then the value of support of the set calculated for the transaction is equal to k ∗ (1/k) = 1. The ∼AR algorithm starts from replacing each unknown element u elem in the transactions by all known items k elem corresponding to the unknown ones in other transactions. Each element k elem replaces the u elem with the probability evaluated based on its frequency of occurring in the transaction, of which

312

G. Protaziuk and H. Rybinski

the k elem elements come form. The authors assumed that such replacing is possible because of presence of names of classes of elements or uniﬁed ordering of items in the transactions. A value contributed by a single k elem to the total value of support of the Zk set calculated for the transaction t is additionally multiplied by the probability of its occurring.

4

Discovering of Association Rules Under Incompleteness

Incompleteness of data creates problem in interpreting minimal support and minimal conﬁdence thresholds given by the user. In the case of missing or unknown values it is not possible to calculate exact values of the support and conﬁdence measures. Instead, we can provide the estimation of these values and the range, limited by optimistic and pessimistic support (conﬁdence), in which the true value is placed. 4.1

Support and Conﬁdence Under Incompleteness

In the order to express the properties of data incompleteness we will apply the following notions: • by minSet(X) we denote the maximal set of records (transactions) which certainly support the itemset X. • by maxSet(X) we denote the maximal set of records (transactions) which possibly support the itemset X. • by nkD we denote the maximal set of records (transactions) which are incomplete (i.e. include at least one item with unknown or missing value). • by kmD we denote the maximal set of records (transactions) which include only items with known values. Deﬁnition 1. Minimal (pessimistic) support of an itemset X, denoted further as pSup(X), is deﬁned as the number of records (transactions) which certainly support the itemset X, i.e. pSup(X) = |minSet(X)|. Deﬁnition 2. Maximal (optimistic) support of an itemset X, denoted further as oSup(X), is deﬁned as the number of records (transactions) which possibly support the itemset X : oSup(X) = |maxSet(X)|. The estimated (probable) support of an itemset X is denoted further as probSup(X). Deﬁnition 3. Minimal (pessimistic) conﬁdence of a rule X → Y , denoted further as pConf (X → Y ), is deﬁned as: pConf (X → Y ) = |minSet(X ∪ Y )|/|maxSet(X)|. Deﬁnition 4. Maximal (optimistic) conﬁdence of a rule X → Y , denoted further as oConf (X → Y ), is deﬁned as: oConf (X→Y )=|maxSet(X∪Y )|/(|maxSet(X∪Y )|+|minSet(X)\minSet(Y )|).

Discovering Association Rules in Incomplete Transactional Databases

313

Deﬁnition 5. Estimated (probable) conﬁdence of a ruleX → Y , denoted further as probConf (X → Y ), is deﬁned as: probConf (X → Y ) = probSup(X ∪ Y )/probSup(X). 4.2

Support Calculation for Single Record or Transaction

The standard method of calculating support for a complete record can be expressed by the formula: Supr (X) =

1 if X is present in a record r 0 otherwise

(5)

where Supr (X) denotes support of the itemset X calculated for the record r, r ∈ DR. Instead, in the case of incomplete records we estimate probable support, for which we can use a generalization of (5) in the form: probSupr (X) = prob(sat(X, r)) if X may be present in a record r

(6)

where prob(sat(X, r)) denotes probability that the itemset X is present in the record r. Obviously if X is certainly present in a record r (all needed values in r are present) then probSupr (X) = 1. If X cannot be present in r then probSupr (X) = 0. For transactional data the method of calculating support in a single transaction is analogous. 4.3

Postulates for Relational Data

The simplest approach to the calculation of support and conﬁdence based on optimistic or pessimistic estimation does not promise good results, especially if the incompleteness of data is signiﬁcant and cannot be neglected. To obtain result of a higher quality more advanced techniques for the support and conﬁdence estimation should be used. In the literature some proposals in this direction have been published. As mentioned above, we claim that any such method should satisfy the postulates deﬁned in [12], in order to properly asses statistically justiﬁed expected support and conﬁdence. The original postulates refer to the relational database. Let us recall them: (P1) (P2) (P3) (P4) (P5)

probSup(X) ∈ [pSup(X), oSup(X)] probSup(X) ≥ probSup(Y ) for X ⊂ Y probConf (X → Y ) = probSup(X ∪ Y )/probSup(X) probConf (X) ∈ [pConf (X → Y ), oConf (X → Y )] XInstances(A) probSup(X) = 1 for any A ⊆ AT

where X, Y are itemsets, and Instances(A) is the set of all possible tuples over the set of attributes A. Postulate P1 assures the natural limitation of estimated support – it can not be greater than optimistic support and less than pessimistic support. The second

314

G. Protaziuk and H. Rybinski

postulate says that an itemset can not be present in database more often than its proper subset. The postulate 3 introduces the standard way for conﬁdence calculation. The fourth one is analogous to P1 but refers to conﬁdence. The last one states that for freely chosen set of attributes the sum of support of all itemsets consisting of items which belong to a domain of selected attributes is equal to the number of records included in a database. In the [16] it is shown that P4 is redundant. With the satisﬁed condition of P2 the estimated support is consistent with the fact that any superset of an itemset X does not occur more often than X. However, satisfying these criteria does not necessary means that the sum of support of supersets of an itemset X, speciﬁed on the same set of attributes, is not greater than support of X. Example: For the record r8 from the base DR n1 (Table 1) the probabilities P r o of occurrences are deﬁned for the following sets: P r w(z1 = {atr3 = a}) = 2/3, P r w(z2 = {atr3 = a1}) = 1/3, P rw (z3 = {atr4 = b}) = 1/2, P rw (z4 = {atr4 = c}) = 1/2, P rw (z5 = {atr3 = a, atr4 = b}) = 1/2, P rw (z6 = {atr3 = a, atr4 = c}) = 1/2. The support of an itemset Y calculated for the record r8 is as follows: P r w(zN ) if Y = zN ∪ r8.knowni if Y ⊂ r8.known probSup(Y ) = 1 0 Otherwise where: zN – is one of the sets zi deﬁned above, r8.known – is a set of all sets which can be generated from known values included in the record r8, r8.knowni – one of sets included in the r8.known set. The support deﬁned in this way satisﬁes all the postulates, however after counting supports of the following set [Y 1 = {atr2 = x, atr3 = a}, probSup(Y 1) = 42/3], [Y 2 = {atr2 = x, atr3 = a, atr4 = b}, probSup(Y 2) = 2/1], [Y 3 = {atr2 = x, atr3 = a, atr4 = c}, probSup(Y 3) = 2/1] one can conclude that the sets Y 2 and Y 3 occur in the database more often than the set Y 1, which of course is not possible. Table 1. Relational database DR n1 id atr2 atr3 atr4 atr5 atr6 r1 x A b Y v1 r2 x A b Y v2 r3 x A c Y v3 r4 x A c Y v4 r5 x a1 b1 Z v1 r6 x a1 c1 P v3 r7 k a2 b2 P v3 r8 x ∗ ∗ Y v2

* – the missing value

Discovering Association Rules in Incomplete Transactional Databases

315

To exclude such situations the postulate P2 has to be modiﬁed. Below the new version of the postulate is presented. (P2n) : probSup(X) = probSup(X ∪ Z) for any A ⊆ AT. (7) Z∈Instances(A)

where X = ∅, Z = ∅ and X ∩ Z = ∅ for any set Z. Rationale: Support of a set X can be counted by testing values in records only for those attributes over which the set X is deﬁned. If we consider larger set of attributes then of course the support of the set X does not change. In this case we can say that we examine whether in a given record the set X is present along with an additional set Z deﬁned over attributes out of those in X. If in place of Z consecutively all the sets deﬁned over the additional attributes will be considered then ﬁnally we will obtain valid value of the support of X. 4.4

Postulates for Transactional Data

For transactional data the postulates from [12] can be used directly, except for P2n and P5. In the deﬁnitions below we use the following notations: • superset(X, k) is a set of all supersets of X which include k more elements than X. • tmax is a maximal transaction i.e. transaction belonging to the given DT which has the most known elements. • |t| is the length of transaction t – number of elements included in this transaction • DTk – set of transaction which include at least k elements • sets(Ik ) denotes sets of k-itemsets • ti denotes the ith transaction. The postulate 2 for transactional data is deﬁned as follows: tmax |−|X|

probSup(X) ≥

k=1

(−1)(k+1)

probSup(Y ).

(8)

Y ∈superset(X,k)

The inequality in the formula results from taking into consideration the diﬀerent lengths of transactions. The postulate 5 for transactional data is deﬁned as follows: X∈sets(Ik ) probSup(X) =1 (9) |Dk | k (C|ti | ti ∈ DTk k where C|t| is a number of all combinations of k items from t. In the deﬁnition of the postulate the absolute support is used.

Rationale for Postulate 2: Calculation of a support of an itemset X can be done by summing number of transactions which contain only all elements

316

G. Protaziuk and H. Rybinski

included in X with the number of transactions that include the set X and at least one more element i.e. which contain a set z = {X ∪{y}|{y}∩X = ∅, y −an item}. There are up to n such sets z, where n is a number of items y, which occur in the database but not in X. A number of transaction in which at least one of the sets z occurs as the percentage of the number of all transaction can be calculated by adopting the following formula (the probability of the sum of events): P(A1 ∪ A2 ∪ . . . ∪ An ) = P(A1 ) + P(A2 ) + . . . + P(An ) − P(A1 ∩ A2 ) − P(A1 ∩ A3 ) − . . . − P(An−1 ∩ An ) + P(A1 ∩ A2 ∩ A3 ) + P(A1 ∩ A2 ∩ A4 ) + . . . + P(An−2 ∩ An−1 ∩ An ) + . . . + (−1)n−1 P(A1 ∩ . . . ∩ An ).

(10)

Assuming that an event Ai represents occurrence of sets zi1 in a transaction (where zi1 is an ith set composed from item yi and elements included in X) we can write the following equation: P (Z 1 ) =

n

(−1)k+1

P (zik )

(11)

zik ∈Z k

k=1

where Z k = superset(X, k), and P(Z 1 ) is the probability that the transaction t contains X and at least one additional item. If we replace the probability by frequencies, and express the frequencies by relative support of the sets we obtain the following formula: ptobSup(Z 1) =

n

(−1)(k+1)

probSub(zik )

(12)

zik ∈Z k

k=1

As we are interested in calculating support of itemsets, we should consider only such z k sets for which k + |X| ≤ |tmax |. All the more numerous sets have support equal to 0. Hence, we can rewrite our formula in the following manner: |tmax |−|X| 1

probSup(Z ) =

(−1)(k+1)

probSup(zik ).

(13)

z1k ∈Z k

To calculate support of the set X we have to sum the values: (i) resulting from the formula (13) and (ii) support of X for the transaction with exactly the same items as in X. This leads us ﬁnally to the formula (8) above. Rationale for postulate 5: Each transaction supports n k-itemsets, where k ≤ |t| and n is the number of all diﬀerent k-elements sets which can be created the transaction. Hence for each transaction from the items belonging to k t: X∈sets(Ik ) probSupt (X) = C|t| , where probSupt (X) denotes the support of the itemset X counted for the transaction t. The fulﬁllment of the postulate 5 requires estimation of support for each possible itemset, even if the given itemset is certainly infrequent. In the tasks of discovering association rules only frequent sets are interested, so there is no need to

Discovering Association Rules in Incomplete Transactional Databases

317

take into consideration infrequent itemsets. In our opinion the postulate 5 can be weaken by replacing the equality relation (=) by the relation ≤ in the formula (9). This modiﬁcation preserves natural deﬁnition of the postulate and makes it more practical. The modiﬁed postulates will be denoted as SJATD (Statistically Justiﬁed Approach for Transactional Data). 4.5

Postulate for Single Record or a Transaction

The postulates SJATD presented above provide conditions for the support calculated for the entire database. On the other hand they say nothing about required properties of support counted for a single record or transaction. For the methods that estimate support of an itemset X by summing its support calculated for single record it seems that more practical is to deﬁne conditions that should be fulﬁlled by support calculated for single record or transaction rather than calculated for entire database. Below we present such requirements. Postulates for single record (P1r) probSupR (X) ∈ [0, 1] (P2r) probSupR (X) ≥ probSupR (Y ) for X ⊂ Y (P3r) probSupR (X) ≥ Z∈sets(At) probSupR (X∪Y )X = ∅, Z = ∅ and X∩Z = ∅ for any set Z (P4r) X∈set(At) , probSupR (X) = 1 for each set of attributes At ⊆ A. For transactional data the postulates: (P3r) and (P4r) have to be redeﬁned. The appropriate formulas are given below. Postulate P3r for transactional data: |T |−|X|

probSupT (X) ≥

(−1)(k+1)

probSupT (Y )

(14)

Y ∈superset(X,k)

k=1

Postulate P4r for transactional data X∈sets(Ik )

probSupT (X)

k C|T |

=1

(15)

where probSupT (X) is a support of an itemsets X counted for transaction T .

5

Algorithm DARIT

In this section we present the new algorithm for discovering association rules from incomplete transactional data, called in the sequel DARIT (Discovering Association Rules in Incomplete Transactions). In our approach we allow that incomplete transaction may have any number of missing element. We start from the description of the data structure, called mT-tree, which is used in the algorithm.

318

5.1

G. Protaziuk and H. Rybinski

The mT-Tree

The tree-like mT-tree structure is used for storing information concerning sets tested by DARIT. Each node of this tree has assigned level, (the root is at level 1). Data associated with itemsets (supports, items) are stored in a node in a table of elements, denoted further as tblElem. Each ﬁeld of the tblElem table in a node at level l may have pointer to another node at level l + 1. With each element stored in the tblElem table there are three values associated: val pSup, val probSup, val oSup, which are used for calculating pessimistic, estimated and optimistic supports respectively. On Figure 1 a simpliﬁed structure of mT-tree is presented. In the tables are only shown elements and name of a val pSup parameter with itemsets to which this parameter concerns. Items belonging to a path in the mT-tree, beginning from the root and composed of the ﬁelds of the tblElem tables, form a frequent or potentially frequent itemset. The items stored in the root of the mT-tree are frequent 1 itemsets. For instance, the set {a, b, c, d} is represented by the path (n1[a], n2[b], n5[c], n8[d]) on Figure 1, where nN[x] stands for the ﬁeld in tblElem where the item x is stored in a node of number N .

n1 - level 1. ( root of tree ) val_pSup{a} val_pSup{b} val_pSup{c} val_pSup{d} a b c d

n4 - level 2. val_pSup{c,d} d

n2 - level 2. val_pSup{a,b} val_pSup{a,c} val_pSup{a,d} b c d

n3 level 2. val_pSup{b,c} val_pSup{b,d} c d n5 - level 3. val_pSup{a,b,c} val_pSup{a,b,d} c d

n8 - level 4. val_pSup{a,b,c,d} d

n6 - level 3. val_pSup{a,c,d} d

n7 - level 3. val_pSup{b,c,d} d

Fig. 1. mT-tree - simpliﬁed schema

5.2

Algorithm DARIT

The pseudo code of the algorithm DARIT is presented below.

Discovering Association Rules in Incomplete Transactional Databases

319

Algorithm DARIT 1) mTd: mT-tree; 2) Apriori Adapt(kmD, nkD, mTd) 3) for each transaction t ∈ nkD 4) begin 5) Generate Set NZ(t, mTd); 6) mTd.Modify probSup(t,t.NZ); 7) end 8) Generate Rules(mTd); At the beginning the set of potentially frequent itemsets is generated (line 2) by calling the Apriori Adapt procedure. In the procedure for determining potentially frequent itemsets the pessimistic support and minimal pessimistic supports thresholds are used. The minimal pessimistic support, denoted further as min minSup, is an additional parameter of the algorithm. This parameter deﬁnes threshold which should be exceeded by the pessimistic support of each itemset in order to consider the itemset as potentially frequent. It allows for appropriate limitation of number of sets taken into consideration during execution of the procedure Apriori Adapt, especially in the case of signiﬁcant data incompleteness. In the next step of the DARIT algorithm for each incomplete transaction t, based on the sets stored in the mT-tree a set t.N Z is generated; it consists of the sets which may occur in place of the special element null, which indicates missing elements in the transactions (line 5). For each element of the t.N Z set the probability of occurring in the considered transaction is assigned. Based on the results obtained in this step, the values of the probSup support of itemsets represented in the mT-tree are modiﬁed (line 6). At the end of the algorithm the procedure Generate Rules is called – it produces the association rules from the mT-tree using values of estimated support of itemsets. Procedure Apriori Adapt Procedure Apriori Adapt(Set of transaction kmD,nkD; mT-tree mTd) 1) Add Frequent Items(kmD, nkD, mTd) 2) p=2; 3) while(Generate Candidates(p, mTd)> 0) 4) begin 5) for each transaction t ∈ kmD 6) Calculate Support(t, mTd); 7) for each transaction t ∈ nkD 8) Calculate Support Incomplete(t, mTd); 9) mTd.Remove NotFrequent( minSup, min minSup); 10) p=p+1; 11) end The procedure Apriori Adapt starts from adding 1-itemset potentially frequent to the root of the mT-tree (line 1). The function Generate Candidates creates

320

G. Protaziuk and H. Rybinski

candidate sets and returns their number. Candidate sets in a node n at level p are generated by creating child nodes at level p + 1, for each ﬁeld in the table tblElem, except for the last one. In the child node cn created for the j th ﬁeld, the table of elements consists of all those elements from the table tblElem of the parent node, which are stored in the ﬁelds of the index greater than j. The procedure Calculate Support increases value of the optimistic support and of the pessimistic support for those of the candidate sets which are supported by a complete transaction t. The procedure Calculate Support Incomplete (line 8) diﬀers from the procedure Calculate Support in that it increases values of the optimistic support for each candidate set. The method Remove NotFrequent removes the candidate sets which certainly will not be frequent, and for the remaining candidate itemsets it sets value of probSup support to the value of the pessimistic support. Procedure Generate SetNZ Procedure Generate Set NZ(transaction t, mT-tree mTd) 1) k = min(mT d.max length set −1, t.nb known items) 2) while(Stop condition = false and k > 0) 3) begin 4) t.NZ= t.NZ ∪ mTd.Find set nZ(t.knownk ); 5) k = k − 1; 6) Stop condidtion = Check stop condition(); 7) end 8) Calculate Probability(mTd); At the beginning of procedure Generate Set NZ the initial value of k is deﬁned. It is calculated as a minimum of the 2 values (i) number of known items in the transaction t, and (ii) the number of elements in the most numerous potentially frequent set stored in mT-tree. Next, in the method Find set nZ for each incomplete transaction t the (k + j)-itemsets, (denoted as zpc) are looked for in the mT-tree. Formally, zpc = {nZ ∪ t.knownki } and t.knownki is a set consisting of k known elements from the transaction t. First, 1-item nZ sets are found, and then the mT-tree is traversed deeper in order to ﬁnd more numerous nZ sets. The nZ set is added to the t.N Z set if it does not include known elements from the transaction t. With each set nZ the parmProb parameter is associated. Further on, the parameter is used to estimate probability of occurrence of the given set nZ instead of the special element null in the transaction, thus to estimate the value of the probSup support. The value of the parmP rob is equal to the value of minSup(zpc) of the currently examined set zpc, or if the given nZ set is already in the t.N Z set, the value of the parameter is increased by minSup(zpc). This procedure is repeated for k = k − 1 down to k = 0, or until the stop condition is met. The stop condition is fulﬁlled when the sum of the parmProb parameters of 1-item nZ sets included in the t.N Z set exceeds the following value: max nb unknown = max len trans − t.nb known elem

(16)

Discovering Association Rules in Incomplete Transactional Databases

321

where max len trans is the maximal number of the potentially frequent items included in the single transaction, and t.nb known elem is the number of known items contained in the transaction t. The Calculate Probability procedure The way in which the value of the parameter parmP rob of the nZ sets is computed causes that it can not be directly used as probability of occurrence of the given nZ set in a transaction. Generally, we have to deal with the following basic problems: 1) A value of the parameter for a single set or a sum of values for group of sets exceeds thresholds. In the former case it is 1 – the maximal value of probability, in the latter the threshold is associated with number of items included in a transaction, for instance, the sum of probabilities of occurrences of single items in a given transaction cannot be greater than the maximal length of the transaction. This condition may be expressed as: i=1..n prob(itemi , t) ≤ max(|t|), where prob(itemi , t) is the probability of occurring of ith item in the transaction t, max(|t|) is the maximal length of the transaction t, n is the number of items potentially frequent. The solution of this problem is a simple normalization of values to the required level. 2) Values of the parameter are very small. In this case the results obtained by applying the DARIT algorithm is comparable with the results obtained by using the methods in which the incompleteness of data is neglected (pessimistic support is used), but with much greater computational cost. To solve the problem these values are multiplied by certain ratio, which is calculated based of most probable number of items which should be present in the transaction in the place of the element null. The ratio is computed in such a way that the sum of probabilities of occurring of single items in a given transaction is not greater than the possible maximal length of the transaction. The Modify probSup method The M odif y probSup method increases the value of the probSup support for each such zpc set stored in the mT-tree, that zpc = {nZ ∪zZ}, where nZ ∈ t.N Z and zZ ∈ t.known. Note that the set zZ may be empty. The value of the probSup support is increased by the value of the parmP rob for the given nZ set. The Generate Rules procedure The Generate Rules procedure generates association rules from the sets stored in the mT-tree in the following way: for each set cZ for which the minimal support requirements are fulﬁlled, all its supersets nZc are found, such that the value probSup(nZc)/probSup(cZ) is not less than the minimal conﬁdence threshold. Next, the rule cZ → {nZc\cZ} is generated with support equal to probSup(nZc) and conﬁdence equal to probSup(nZc)/probSup(cZ). 5.3

The SJATD Postulates

Theorem: DARIT satisﬁes the SJATD postulates.

322

G. Protaziuk and H. Rybinski

Proof: Postulate 1. Value of the probSup support for each set is of course not less than 0. The way of calculating probSup applied in the Calculate Probabilty procedure ensures that the value of probSup support never exceeds 1. Postulate 2. The method of computing the values of the parmSup parameters - summing the values of pessimistic support, ensures that the value of the parameter parmSup for any set will be not less than parmSup for its supersets. Multiplication of these values by the same factor does not inﬂuence this relation. Additionally, the way of adding the sets nZ to the set t.N Z by traversing mT-tree in depth, and adding more and more numerous sets guarantees that for any given set all its subsets have been taken into consideration. Postulate 3. Fulﬁlling this postulate follows from the method of computing the probSup support described in the proof for Postulate 2 and from the fact that the pessimistic support meets this postulate. Postulate 5 (weakened). The postulate says that sum of the support of nitemsets calculated for a single transaction cannot be greater than the number of k-elements sets which can be created from the items included in the transaction. According to the deﬁnition for an incomplete transaction, support of the set dowZ is equal to: • 1, if dowZ ∈ t.known, • nZ.parmSup, if dowZ = nZ or dowZ = nZ∪zZ, nZ ∈ t.N Z, zZ ∈ t.known, • 0 otherwise where t.known contains all the sets which can be formed from the known elements included in the transaction t. Assuming that the number of elements nb elem t in the transaction t is equal to max(|t.N Zk |) + t.nb known, where t.nb known denotes the number of known elements in the transaction t and t.N Zk denotes k − itemsets from the set t.N Z, then number on n-elements sets, created from the items included in the n transaction t is equal to Cnb elem t . The sum Sm of the support of n-itemsets counted for the transaction t can be split into tree parts: Sm = Sknown + Sunknown + Sjoined

(17)

Of course, not all parts occur in all cases. The partial sums in the equation above means, respectively: • Sknown — the sum of the support of the sets dowZ ∈ t.known. The sum is equal to the number of n-elements sets created from the known items included in the transaction t. • Sunknown — the sum of supports of the sets dowZ ∈ t.N Z. In this case the sum is less than the number of the sets created from max(|t.N Zk |) elements. It is ensured by the way of computing the normalization ratio in the method Calculate Probability. • Sjoined — the sum of supports of the sets dowZ = nZ ∪ zZ, nZ ∈ t.N Z, zZ ∈ t.known, nZ = ∅, zZ = ∅. The sum is smaller than the number of n-element sets created by joining the i-itemsets created from t.nb known

Discovering Association Rules in Incomplete Transactional Databases

323

elements and j-itemsets created from the max(|t.N Zk |) elements for j, i > 0 and j + i = k. The number of such sets can be computed from the following equation:

i=min(n−1,t.nb known) max(t.N Zk )

known Ct.nb ∗ Cn−1 i

(18)

i=1

for i ≤ t.nb known and n − i ≤ max(|t.N Zk |). During the calculation of the sum Sjoined the second factor is replaced by the sum of supports of the sets nZ ∈ t.N Zn−i , which as it follows from explanation presented for Sunknown is smaller than the value of this factor in the original formula.

6

Experiments

percent differenceof supportsof sets

The objective of the performed experiments was to evaluate the practical value of the proposed approach. In order to evaluate the quality of the results obtained by executing the DARIT algorithm the following procedure was applied. First, the sets of association rules and frequent itemsets, further denoted as the referential sets, were generated from a complete transactional database. Next, the incomplete database was created by random removing some elements from 40% 35% d_1

30%

d_2 25% d_5 20%

d_10

15%

ign_1

10%

ign_2

5%

ign_5

0% 0%

ign_10 10%

20% 30% 40% 50% 60% 70% percent of incomlete transactions

80%

90%

100%

Fig. 2. Percent diﬀerence of supports of sets 100% not discoveredsetsinpercent

90% 80%

d_1

70%

d_2

60%

d_5

50%

d_10

40%

ign_1

30% ign_2

20%

ign_5

10% 0% 0%

ign_10 10%

20% 30% 40% 50% 60% 70% percent of incomlete transactions

80%

90%

Fig. 3. Percent of non-discovered sets

100%

324

G. Protaziuk and H. Rybinski 100%

additional setsinpercent

sumof not discoveredand

90% 80%

d_1

70%

d_2

60%

d_5

50%

d_10

40%

ign_1

30% ign_2 20% ign_5

10% 0% 0%

ign_10 10%

20% 30% 40% 50% 60% 70% percent of incomlete transactions

80%

90%

100%

Fig. 4. Erroneous sets as percentage of the size of the referential set

percent differenceof supportsof rules

32% 28% d_1

24%

d_2

20%

d_5 16%

d_10

12%

ign_1

8%

ign_2

4%

ign_5

0% 0%

ign_10 10%

20% 30% 40% 50% 60% 70% percent of incomlete transactions

80%

90%

100%

Fig. 5. Percent diﬀerence of supports of association rules

8%

percent difference

of confidencesof rules

7% d_1

6%

d_2 5% d_5 4%

d_10

3%

ign_1

2%

ign_2 ign_5

1% 0% 0%

ign_10 10%

20% 30% 40% 50% 60% 70% percent of incomlete transactions

80%

90%

100%

Fig. 6. Percent diﬀerence of conﬁdences of association rules

some transactions. The number of incomplete transactions and the number of unknown elements in a single transaction vary in the experiments. The last step in the procedure was to use the DARIT algorithm to discover association rules in the incomplete data. The evaluation of quality of the results was based on the diﬀerence between the referential sets of association rules or frequent sets and the sets obtained from analyzing incomplete data. The diﬀerence was described by the following measures: – percent diﬀerence between the supports of rules and frequent itemsets and between the conﬁdence measures

Discovering Association Rules in Incomplete Transactional Databases

325

not discovered rulesinpercent

110% 100% 90%

d_1

80%

d_2

70%

d_5

60% 50%

d_10

40%

ign_1

30%

ign_2

20%

ign_5

10% 0% 0%

ign_10 10%

20% 30% 40% 50% 60% 70% percent of incomlete transactions

80%

90%

100%

Fig. 7. Percent of non-discovered association rules 110%

additional rulesinpercent

sumof not discoveredand

100% 90%

d_1

80% d_2

70% 60%

d_5

50%

d_10

40%

ign_1

30%

ign_2

20%

ign_5

10% 0% 0%

ign_10 10%

20% 30% 40% 50% 60% 70% percent of incomlete transactions

80%

90%

100%

Fig. 8. Erroneous rules as a percentage of the size of the referential set

– percentage of the sets (rules) presented in the referential set that have not been discovered from incomplete data – number of additional sets (rules) discovered from incomplete data but not present in the referential set, expressed as a percentage of the referential set. The results obtained from DARIT were also compared with the results obtained from the other data mining algorithms, where the pessimistic support estimation is used. In the experiments the synthetic data were used. The set consisted of 592 transactions, the average length of the transactions was 20 elements. It was 1116 diﬀerent items in the database. The test databases were incomplete in the degree between 10% up to 90% of transactions. For each case four situations with diﬀerent number of missing elements in a single transaction were analyzed, namely the situation in which 1, 2, 5, or 10 elements in a single transaction were missing. In all experiments the minimal support threshold was set to 0.06 and the minimal conﬁdence threshold was set 0.9. The tests were carried out for four values of minimal pessimistic support threshold calculated as a percentage of the minimal support value, namely 10%, 20%, 50%, 80% of this value. On the charts for Figure 2 to Figure 8 the average results for the minimal pessimistic support are presented. On the legend d denotes the results obtained from DARIT, ign stands for the ignore method, and the index at d or ign indicates the number of missing elements in a single transaction. The results concerning frequent itemsets are presented on the ﬁgures 2, 3 and 4. The proposed approach is generally much better than the ignore method.

326

G. Protaziuk and H. Rybinski

On Figure 4 the error of the algorithms DARIT and ignore is presented in the form of the sum of numbers of missing and erroneously discovered sets. Such a sum better presents diﬀerences between results obtained by the considered methods, since the ignore method does not produce erroneous sets. The results coming from those two methods are comparable only in the situations where the incompleteness of data is small. The presented results show that by applying the DARIT algorithm one can achieve good results also in the case of quite high incompleteness of data (up to 50% of incomplete transactions). On Figures 5 – 8 we present the results concerning comparisons of the sets of discovered association rules. Also with respect to association rules DARIT is much better than the ignore method, though here the diﬀerence is smaller. The obtained results are generally worse than the results obtained for the frequent sets. The errors in estimating support rather cumulate then eliminate. In the case of association rules only the best conﬁgurations concerning the minimal pessimistic support thresholds allow to obtain good results when the incompleteness of data is higher.

7

Conclusions

In the paper we extended some postulates of legitimate approach to discovering association rules, as deﬁned in [12], so that they may be applied also to transactional data. We have presented in detail a new approach for discovering association rules from incomplete transactional databases. In the presented DARIT algorithm we do not use any knowledge that is external to the dataset, but exploit only relations discovered in the investigated database. We have shown that the approach satisﬁes all the postulates of SJATD. We presented a number of experimental results using synthetic data. The performed experiments show that the proposed algorithm well foresees real values of the support and outperform the methods based on ignoring unknown values.

References 1. Agrawal R., Imielinski T., Swami A.: Mining Associations Rules between Sets of Items in Large Databases. In: Proc. of the ACM SIGMOD Conference on Management of Data, Washington, USA (1993) 207–216 2. Agrawal R., Srikant R.: Fast Algorithms for Mining Association Rules. In: Proc. of the 20th International Conference on Very Large Databases Conference (VLDB), Santiago, Chile, 1994. Morgan Kaufmann (1994) 487–499 3. Bayardo R.J., Agrawal R.: Mining the Most Interesting Rules. In: Proc. of the Fifth ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD), San Diego, CA, USA, 1999. ACM (1999) 145–154 4. Bayardo R.J., Agrawal R., Gunopulos D.: Constraint-Based Rule Mining in Large, Dense Databases. Data Mining and Knowledge Discovery, Vol. 4, No. 2/3 (2000) 217–240 5. Bayardo R. J. Jr.: Eﬃciently Mining Long Patterns from Databases, Proceedings of ACM SIGMOD International Conference on Management of Data, Seattle, (1998).

Discovering Association Rules in Incomplete Transactional Databases

327

6. Breiman L., Friedman J. H., Olshen R. A. Stone C. J.: Classiﬁcation and regression trees, Belmont Wadsworth, (1984). 7. Calders T., Goethals B.: Mining All Non-derivable Frequent Item Sets, Proc. of Principles of Data Mining and Knowledge Discovery, 6th European Conf., Helsinki, (2002) 8. Dardzi´ nska-Gl¸ebocka A., Chase Method Based on Dynamic Knowledge Discovery for Prediction Values in Incomplete Information Systems, PhD thesis, Warsaw, 2004. 9. Friedman H. F., Kohavi R., Yun Y., Lazy decision trees, Proceedings of the 13th National Conference on Artiﬁcial Intelligence, Portland, Oregon, (1996) 10. Grzymala-Busse J. W.: Characteristic Relations for Incomplete Data: A Generalization of the Indiscernibility Relation, Proceedings Rough Sets and Current Trends in Computing, 4th International Conference, Uppsala, (2004). 11. Han J., Pei J., Yin Y.: Mining Frequent Patterns without Candidate Generation. In: Proc. Of the 2000 ACM SIGMOD International Conference on Management of Data, Dallas, Texas, USA, 2000. SIGMOD Record, Vol. 29, No. 2 (2000) 1–12 12. Kryszkiewicz M., Rybinski H.: Legitimate Approach to Association Rules under Incompleteness. In: Foundations of Intelligent Systems. Proc. of 12th International Symposium (ISMIS), Charlotte, USA, 2000. Lecture Notes in Artiﬁcial Intelligence, Vol. 1932. Springer-Verlag (2000) 505–514 13. Kryszkiewicz M.: Probabilistic Approach to Association Rules in Incomplete Databases, Proceedings of Web-Age Information Management, First International Conference, WAIM 2000, Shanghai, (2000). 14. Kryszkiewicz M.: Concise Representation of Frequent Patterns based on Disjunction-Free Generators. In: Proc. of the 2001 IEEE International Conference on Data Mining (ICDM), San Jose, California, USA, 2001. IEEE Computer Society (2001) 305–312 15. Kryszkiewicz M.: Representative Association Rules. In: Research and Development in Knowledge Discovery and Data Mining. Proc. of Second Paciﬁc-Asia Conference (PAKDD). Melbourne, Australia, 1998. Lecture Notes in Computer Science, Vol. 1394. Springer (1998) 198–209 16. Kryszkiewicz M.: Concise Representations of Frequent Patterns and Association Rules Habilitation Thesis, Warsaw University of Technology, (2002) 17. Liu W. Z., White A. P.: Thompson S. G., Bramer M. A.: Techniques for Dealing with Missing Values in Classiﬁcation, Proceedings of Advances in Intelligent Data Analysis, Reasoning about Data, Second International Symposium, London, (1997) 18. Nayak J. R., Cook D. J.: Approximate Association Rule Mining, Proceedings of the Fourteenth International Artiﬁcial Intelligence Research Society Conference, Key West, Florida, (2001) 19. Parsons S.: Current Approach to Handling Imperfect Information in Data and Knowledge Bases, IEEE Transaction on knowledge and data engineering Vol. 8, (1996) 20. Pawlak Z.: Rough Sets. International Journal of Information and Computer Sciences No. 11 (1982) 341–356 21. Pawlak Z.: Rough Sets: Theoretical Aspects of Reasoning about Data. Kluwer Academic Publishers, Vol. 9 (1991) 22. Pasquier N., Bastide Y., Taouil R., Lakhal L.: Discovering Frequent Closed Itemsets for Association Rules. In: Proc. of Database Theory - ICDT ’99. Proc. of 7th International Conference (ICDT), Jerusalem, Israel, 1999. Lecture Notes in Computer Science, Vol. 1540. Springer (1999) 398–416

328

G. Protaziuk and H. Rybinski

23. Protaziuk G., Sodacki P., Gancarz ., Discovering interesting rules in dense data, The Eleventh International Symposium on Intelligent Information Systems, Sopot, (2002). 24. Bastide Y., Pasquier N., Taouil R., Stumme G., Lakhal L.: Mining Minimal Nonredundant Association Rules Using Frequent Closed Itemsets. Comp. Logic (2000) 972–986 25. Quinlan J. R., C4.5 Programs for Machine Learning, San Mateo, California, (1993) 26. Ragel A., Cremilleux B.: Treatment of Missing Values for Association Rules. In: Research and Development in Knowledge Discovery and Data Mining. Proc. of Second Paciﬁc-Asia Conference (PAKDD). Melbourne, Australia, 1998. Lecture Notes in Computer Science, Vol. 1394. Springer (1998) 258–270 27. Srikant R., Vu Q., Agrawal R.: Mining Association Rules with Item Constraints. In: Proc. Of the Third International Conference on Knowledge Discovery and Data Mining (KDD). Newport Beach, California, USA, 1997. AAAI Press (1997) 67–73 28. Stefanowski J., Tsoukias A.: Incomplete Information Tables and Rough Classiﬁcation. Int. Journal of Computational Intelligence, Vol. 17, No 3 (2001) 545–566 29. Stefanowski J.: Algorytmy indukcji regu decyzyjnych w odkrywaniu wiedzy (Algorithms of Rule Induction for Knowledge Discovery). Habilitation Thesis, Poznan University of Technology, No. 361 (2001) 30. Wang G.: Extension of Rough Set under Incomplete Information Systems, Proceedings of the 2002 IEEE International Conf. on Fuzzy Systems, Honolulu, (2002) 31. Zaki M.J.: Generating Non-Redundant Association Rules. In Proc. of 6th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Boston, MA, 2000. ACM Press (2000) 34–43 32. Zhang J., Honavar V.: Learning Decision Tree Classiﬁers from Attribute Value Taxonomies and Partially Speciﬁed Data, Proceedings of the Twentieth International Conference (ICML 2003), Washington, DC, (2003)

On Combined Classiﬁers, Rule Induction and Rough Sets Jerzy Stefanowski Institute of Computing Science Pozna´ n University of Technology, 60-965 Pozna´ n, ul.Piotrowo 2, Poland Jerzy.Stefanowski@cs.put.poznan.pl

Abstract. Problems of using elements of rough sets theory and rule induction to create eﬃcient classiﬁers are discussed. In the last decade many researches attempted to increase a classiﬁcation accuracy by combining several classiﬁers into integrated systems. The main aim of this paper is to summarize the author’s own experience with applying one of his rule induction algorithm, called MODLEM, in the framework of diﬀerent combined classiﬁers, namely, the bagging, n2 –classiﬁer and the combiner aggregation. We also discuss how rough approximations are applied in rule induction. The results of carried out experiments have shown that the MODLEM algorithm can be eﬃciently used within the framework of considered combined classiﬁers.

1

Introduction

Rough sets theory has been introduced by Professor Zdzislaw Pawlak to analyse granular information [25,26]. It is based on an observation that given information about objects described by attributes, a basic relation between objects could be established. In the original Pawlak’s proposal [25] objects described by the same attribute values are considered to be indiscernible. Due to limitations of available information, its natural granulation or vagueness of a representation language some elementary classes of this relation may be inconsistent, i.e. objects having the same descriptions are assigned to diﬀerent categories. As a consequence of the above inconsistency it is not possible, in general, to precisely specify a set of objects in terms of elementary sets of indiscernible objects. Therefore, Professor Zdzislaw Pawlak introduced the concept of the rough set which is a set characterized by a pair of precise concepts – lower and upper approximations constructed from elementary sets of objects. This quite simple, but smart, idea is the essence of the Pawlak’s theory. It is a starting point to other problems, see e.g. [27,20,26,9]. In particular many research eﬀorts have concerned classiﬁcation of objects represented in data tables. Studying relationships between elementary sets and categories of objects (in other terms, target concepts or decision classes in the data table) leads to, e.g., evaluating dependency between attributes and objects classiﬁcation, determining the level of this dependency, calculating importance of attributes for objects J.F. Peters et al. (Eds.): Transactions on Rough Sets VI, LNCS 4374, pp. 329–350, 2007. c Springer-Verlag Berlin Heidelberg 2007

330

J. Stefanowski

classiﬁcation, reducing the set of attributes or generating decision rules from data. It is also said that the aim is to synthesize reduced, approximate models of concepts from data [20]. The transparency and explainability of such models to human is an important property. Up to now rough sets based approaches were applied to many practical problems in diﬀerent domains – see, e.g., their list presented in [20]. Besides ”classical” rough sets, based on the indiscernibility relation, several generalizations have been introduced. Such data properties as, e.g., imprecise attribute values, incompleteness, preference orders, are handled by means of tolerance, similarity, fuzzy valued or dominance relations [9,20,37]. Looking into the previous research on rough sets theory and its applications, we could distinguish two main perspectives: descriptive and predictive ones. The descriptive perspective includes extraction information patterns or regularities, which characterize some properties hidden in available data. Such patterns could facilitate understanding dependencies between data elements, explaining circumstances of previous decisions and generally gain insight into the structure of the acquired knowledge. In this context presentation of results in a human readable form allowing an interpretation is a crucial issue. The other perspective concerns predicting unknown values of some attributes on the basis of an analysis of previous examples. In particular, it is a prediction of classes for new object. In this context rough sets and rules are used to construct a classiﬁer that has to classify new objects. So, the main evaluation criterion is a predictive classiﬁcation accuracy. Let us remind that the predictive classiﬁcation has been intensively studied since many decades in such ﬁelds as machine learning, statistical learning, pattern recognition. Several eﬃcient methods for creating classiﬁers have been introduced; for their review see, e.g., [16,19,23]. These classiﬁers are often constructed with using a search strategy optimizing criteria strongly related to predictive performance (which is not directly present in the original rough sets theory formulation). Requirements concerning interpretability are often neglected in favor of producing complex transformations of input data – an example is an idea of support vector machines. Although in both perspectives we could use the same knowledge representation – rules, since motivation and objectives are distinct, algorithmic strategies as well as criteria for evaluating a set of rules are quite diﬀerent. For instance, the prediction perspective directs an interest to classiﬁcation ability of the complete rules, while in the descriptive perspective each rule is treated individually as a possible representative of an ‘interesting’ pattern evaluated by measures as conﬁdence, support or coverage - for a more exhaustive discussion see, e.g., [42]. In my opinion, basic concepts of the rough sets theory have been rather considered in the way similar to a descriptive analysis of data tables. Nevertheless, several authors have developed their original approaches to construct decision rules from rough approximations of decision classes which joined together with classiﬁcation strategies led to good classiﬁers, see e.g. [1,11,20,34,37]. It seems to me that many authors moved their interest to this direction in the 90’s because of at least two reasons: (1) a research interest to verify whether knowledge

On Combined Classiﬁers, Rule Induction and Rough Sets

331

derived from ”closed world” of the data table could be eﬃciently applied to new objects coming from the ”open world” – not seen in the analysed data table; (2) as a result of working with real life applications. Let us also notice that the majority of research has been focused on developing single classiﬁers – i.e. based on the single set of rules. However, both empirical observations and theoretical works conﬁrm that one cannot expect to ﬁnd one single approach leading to the best results on overall problems [6]. Each learning algorithm has its own area of superiority and it may outperform others for a speciﬁc subset of classiﬁcation problems while being worse for others. In the last decade many researches attempted to increase classiﬁcation accuracy by combining several single classiﬁers into an integrated system. These are sets of learned classiﬁers, whose individual predictions are combined to produce the ﬁnal decision. Such systems are known under names: multiple classiﬁers, ensembles or committees [6,45]. Experimental evaluations shown that these classiﬁers are quite eﬀective techniques for improving classiﬁcation accuracy. Such classiﬁers can be constructed in many ways, e.g., by changing the distributions of examples in the learning set, manipulating the input features, using diﬀerent learning algorithms to the same data, see e.g. reviews [6,45,36]. Construction of integrated classiﬁers has also attracted the interest of some rough sets researchers, see e.g. [2,8,24]. The author and his co-operators have also carried out research, ﬁrst on developing various rule induction algorithms and classiﬁcation strategies (a review is given in [37]), and then on multiple classiﬁers [18,36,38,40,41]. The main aim of this paper is to summarize the author’s experience with applying one of his rule induction algorithm, called MODLEM [35], in the framework of diﬀerent multiple classiﬁers: the popular bagging approach [4], the n2 -classiﬁer [18] – a specialized approach to solve multiple class learning problems, and the combiner approach to merge predictions of heterogeneous classiﬁers including also MODLEM [5]. The second aim is to brieﬂy discuss the MODLEM rule induction algorithm and its experimental evaluation. This paper is organized as follows. In the next section we shortly discuss rule induction using the rough sets theory. Section 3 is devoted to the MODLEM algorithm. In section 4 we brieﬂy present diﬀerent approaches to construct multiple classiﬁers. Then, in the successive three sections we summarize the experience of using rule classiﬁers induced by MODLEM in the framework of three diﬀerent multiple classiﬁers. Conclusions are grouped in section 8.

2 2.1

Rules Generation and Rough Sets Notation

Let us assume that objects – learning examples for rule generation – are represented in decision table DT = (U, A ∪ {d}), where U is a set of objects, A is a set of condition attributes describing objects. The set Va is a domain of a. Let fa (x) denotes the value of attribute a ∈ A taken by x ∈ U ; d ∈ / A is a decision attribute that partitions examples into a set of decision classes {Kj : j = 1, . . . , k}.

332

J. Stefanowski

The indiscernibility relation is the basis of Pawlak’s concept of the rough set theory. It is associated with every non-empty subset of attributes C ⊆ A and ∀x, y ∈ U is deﬁned as xIC y ⇔ {(x, y) ∈ U × U fa (x) = fa (y) ∀a ∈ C }. The family of all equivalence classes of relation I(C) is denoted by U/I(C). These classes are called elementary sets. An elementary equivalence class containing element x is denoted by IC (x). If C ⊆ A is a subset of attributes and X ⊆ U is a subset of objects then the sets: {x ∈ U : IC (x) ⊆ X}, {x ∈ U : IC (x) ∩ X = ∅} are called C-lower and C-upper approximations of X, denoted by CX and CX, respectively. The set BNC (X) = CX − CX is called the C-boundary of X. A decision rule r describing class Kj is represented in the following form: if P then Q, where P = w1 ∧ w2 ∧ . . . wp is a condition part of the rule and Q is decision part of the rule indicating that example satisfying P should be assigned to class Kj . The elementary condition of the rule r is deﬁned as (ai (x) rel vai ), where rel is a relational operator from the set {=, , ≥} and vai is a constant being a value of attribute ai . Let us present some deﬁnitions of basic rule properties. [P ] is a cover of the condition part of rule r in DT , i.e. it is a set of examples, which description satisfy elementary conditions in P . Let B be a set of examples belonging to decision concept (class Kj or its appropriate rough approximation in case of inconsistencies). The rule r is discriminant if it distinguishes positive examples of B from its negative examples, i.e. [P ] = [wi ] ⊆ B. P should be a minimal conjunction of elementary conditions satisfying this requirement. The set of decision rules R completely describes examples of class Kj , if each example is covered by at least one decision rules. Discriminant rules are typically considered in the rough sets literature. However, we can also construct partially discriminant rules that besides positive examples could cover a limited number of negative ones. Such rules are characterized by the accuracy measure being a ratio covered positive examples to all examples covered by the rule, i.e. [P ∩ B]/[P ]. 2.2

Rule Generation

If decision tables contain inconsistent examples, decision rules could be generated from rough approximations of decision classes. This special way of treating inconsistencies in the input data is the main point where the concept of the rough sets theory is used in the rules induction phase. As a consequence of using the approximations, induced decision rules are categorized into certain (discriminant in the sense of the previous deﬁnition) and possible ones, depending on the used lower and upper approximations, respectively. Moreover, let us mention other rough sets approaches that use information on class distribution inside boundary and assign to lower approximation these inconsistent elementary sets where the majority of examples belong to the given class. This is handled in the Variable Precision Model introduced by Ziarko

On Combined Classiﬁers, Rule Induction and Rough Sets

333

[47] or Variable Consistency Model proposed by Greco et al. [10] – both are a subject of many extensions, see e.g. [31]. Rules induced from such variable lower approximations are not certain but partly discriminant ones. A number of various algorithms have been already proposed to induce decision rules – for some reviews see e.g. [1,11,14,20,28,34,37]. In fact, there is no unique ”rough set approach” to rule induction as elements of rough sets can be used on diﬀerent stages of the process of induction and data pre-processing. In general, we can distinguish approaches producing minimal set of rules (i.e. covering input objects using the minimum number of necessary rules) and approaches generating more extensive rule sets. A good example for the ﬁrst category is LEM2, MODLEM and similar algorithms [11,35]. The second approaches are nicely exempliﬁed by Boolean reasoning [28,29,1]. There are also speciﬁc algorithms inducing the set of decision rules which satisfy user’s requirements given a priori, e.g. the threshold value for a minimum number of examples covered by a rule or its accuracy. An example of such algorithms is Explore described in [42]. Let us comment that this algorithm could be further extended to handle imbalanced data (i.e. data set where one class – being particularly important – is under-represented comparing to cardinalities of other classes), see e.g. studies in [15,43].

3

Exemplary Rule Classiﬁer

In our study we will use the algorithm, called MODLEM, introduced by Stefanowski in [35]. We have chosen it because of several reasons. First of all, the union of rules induced by this algorithm with a classiﬁcation strategy proved to provide eﬃcient single classiﬁers, [14,41,37]. Next, it is designed to handle various data properties not included in the classical rough sets approach, as e.g. numerical attributes without its pre-discretization. Finally, it produces the set of rules with reasonable computational costs – what is important property for using it as a component inside combined classiﬁers. 3.1

MODLEM Algorithm

The general schema of the MODLEM algorithm is brieﬂy presented below. More detailed description could be found in [14,35,37]. This algorithm is based on the idea of a sequential covering and it generates a minimal set of decision rules for every decision concept (decision class or its rough approximation in case of inconsistent examples). Such a minimal set of rules (also called local covering [11]) attempts to cover all positive examples of the given decision concept, further denoted as B, and not to cover any negative examples (i.e. U \ B). The main procedure for rule induction scheme starts from creating a ﬁrst rule by choosing sequentially the ‘best’ elementary conditions according to chosen criteria (see the function Find best condition). When the rule is stored, all learning positive examples that match this rule are removed from consideration. The process

334

J. Stefanowski

is repeated while some positive examples of the decision concept remain still uncovered. Then, the procedure is sequentially repeated for each set of examples from a succeeding decision concept. In the MODLEM algorithm numerical attributes are handled during rule induction while elementary conditions of rules are created. These conditions are represented as either (a < va ) or (a ≥ va ), where a denotes an attribute and va is its value. If the same attribute is chosen twice while building a single rule, one may also obtain the condition (a = [v1 , v2 )) that results from an intersection of two conditions (a < v2 ) and (a ≥ v1 ) such that v1 < v2 . For nominal attributes, these conditions are (a = va ) or could be extended to the set of values. Procedure MODLEM (input B - a set of positive examples from a given decision concept; criterion - an evaluation measure; output T – single local covering of B, treated here as rule condition parts) begin G := B; {A temporary set of rules covered by generated rules} T := ∅; while G = ∅ do {look for rules until some examples remain uncovered} begin T := ∅; {a candidate for a rule condition part} S := U ; {a set of objects currently covered by T } while (T = ∅) or (not([T ] ⊆ B)) do {stop condition for accepting a rule} begin t := ∅; {a candidate for an elementary condition} for each attribute q ∈ C do {looking for the best elementary condition} begin new t :=Find best condition(q, S); if Better(new t, t, criterion) then t := new t; {evaluate if a new condition is better than previous one according to the chosen evaluation measure} end; T := T ∪ {t}; {add the best condition to the candidate rule} S := S ∩ [t]; {focus on examples covered by the candidate} end; { while not([T ] ⊆ B } for each elementary condition t ∈ T do if [T − t] ⊆ B then T := T − {t}; {test a rule minimality} T := T ∪ {T }; {store a rule} G := B − T ∈T [T ] ; {remove already covered examples} end; { while G = ∅ } for each T ∈ T do if T ∈T −T [T ] = B then T := T − T {test minimality of the rule set} end {procedure}

Ë

Ë

function Find best condition (input c - given attribute; S - set of examples; output best t - bestcondition) begin best t := ∅; if c is a numerical attribute then

On Combined Classiﬁers, Rule Induction and Rough Sets

335

begin H:=list of sorted values for attribute c and objects from S; { H(i) - ith unique value in the list } for i:=1 to length(H)-1 do if object class assignments for H(i) and H(i + 1) are diﬀerent then begin v := (H(i) + H(i + 1))/2; create a new t as either (c < v) or (c ≥ v); if Better(new t, best t, criterion) then best t := new t ; end end else { attribute is nominal } begin for each value v of attribute c do if Better((c = v), best t, criterion) then best t := (c = v) ; end end {function}.

For the evaluation measure (i.e. a function Better) indicating the best condition, one can use either class entropy measure or Laplacian accuracy. For their deﬁnitions see [14] or [23]. It is also possible to consider a lexicographic order of two criteria measuring the rule positive cover and, then, its conditional probability (originally considered by Grzymala in his LEM2 algorithm or its last, quite interesting modiﬁcation called MLEM). In all experiments, presented further in this paper, we will use the entropy as an evaluation measure. Having the best cut-point we choose a condition (a < v) or (a ≥ v) that covers more positive examples from the concept B. In a case of nominal attributes it is also possible to use another option of Find best condition function, where a single attribute value in the elementary condition (a = vi ) is extended to a multi-valued set (a ∈ Wa ), where Wa is a subset of values from the attribute domain. This set is constructed in the similar way as in techniques for inducing binary classiﬁcation trees. Moreover, the author created MODLEM version with another version of rule stop condition. Let us notice that in the above schema the candidate T is accepted to become a rule if [T ] ⊆ B, i.e. a rule should cover learning examples belonging to an appropriate approximation of the given class Kj . For some data sets – in particular noisy ones – using this stop condition may produce too speciﬁc rules (i.e. containing many elementary conditions and covering too few examples). In such situations the user may accept partially discriminating rules with high enough accuracy – this could be done by applying another stop condition ([T ∩ B]/[T ] ≥ α. An alternative is to induce all, even too speciﬁc rules and to post-process them – which is somehow similar to pruning of decision trees. Finally we can illustrate the use of MODLEM by a simple example. The data table contains examples of 17 decision concerning classiﬁcation of some customers into three classes coded as d, p, r. All examples are described by 5 qualitative and numerical attributes.

336

J. Stefanowski Table 1. A data table containing examples of customer classiﬁcation Age Job Period Income Purpose Decision m sr m st sr m sr m sr st m m sr m st m m

u p p p p u b p p e u b p b p p b

0 2 4 16 14 0 0 3 11 0 0 0 17 0 21 5 0

500 1400 2600 2300 1600 700 600 1400 1600 1100 1500 1000 2500 700 5000 3700 800

K S M D M W D D W D D M S D S M K

r r d d p r r p d p p r p r d d r

This data table is consistent, so lower and upper approximations are the same. The use of MODLEM results in the following set of certain rules (square brackets contain the number of learning examples covered by the rule): rule rule rule rule rule rule rule

1. 2. 3. 4. 5. 6. 7.

if (Income < 1050) then (Dec = r) [6] if (Age = sr) ∧ (P eriod < 2.5) then (Dec = r) [2] if (P eriod ∈ [3.5, 12.5)) then (Dec = d) [2] if (Age = st) ∧ (Job = p) then (Dec = d) [3] if (Age = m) ∧ (Income ∈ [1050, 2550)) then (Dec = p) [2] if (Job = e) then (Dec = p) [1] if (Age = sr) ∧ (P eriod ≥ 12.5) then (Dec = p) [2]

Due to the purpose and page limits of this paper we do not show details of MODLEM working steps while looking for a single rule - the reader is referred to the earlier author’s papers devoted to this topic only. 3.2

Classiﬁcation Strategies

Using rule sets to predict class assignment for an unseen object is based on matching the object description to condition parts of decision rules. This may result in unique matching to rules from the single class. However two other ambiguous cases are possible: matching to more rules indicating diﬀerent classes or the object description does not match any of the rules. In these cases, it is necessary to apply proper strategies to solve these conﬂict cases. Review of diﬀerent strategies is given in [37]

On Combined Classiﬁers, Rule Induction and Rough Sets

337

In this paper we employ two classiﬁcation strategies. The ﬁrst was introduced by Grzymala in LERS [12]. The decision to which class an object belongs to is made on the basis of the following factors: strength and support. The Strength is the total number of learning examples correctly classiﬁed by the rule during training. The support is deﬁned as the sum of scores of all matching rules from the class. The class Kj for which the support, i.e., the following expression Strength f actor(R) matching rules R describing Ki

is the largest is the winner and the object is assigned to Kj . If complete matching is impossible, all partially matching rules are identiﬁed. These are rules with at least one elementary condition matching the corresponding object description. For any partially matching rule R, the factor, called Matching factor (R), deﬁned as a ratio of matching conditions to all conditions in the rule, is computed. In partial matching, the concept Kj for which the following expression is the largest M atching f actor(R) ∗ Strength f actor(R) partially matching rules R

is the winner and the object is classiﬁed as being a member of Kj . The other strategy was introduced in [32]. The main diﬀerence is in solving no matching case. It is proposed to consider, so called, nearest rules instead of partially matched ones. These are rules nearest to the object description in the sense of chosen distance measure. In [32] a weighted heterogeneous metric DR is used which aggregates a normalized distance measure for numerical attributes and {0;1} diﬀerences for nominal attributes. Let r be a nearest matched rule, e denotes a classiﬁed object. Then DR(r, e) is deﬁned as: Dr(r, e) =

1 p 1/p ( da ) m a∈P

where p is a coeﬃcient equal to 1 or 2, m is the number of elementary conditions in P – a condition part of rule r. A distance da for numerical attributes is equal to |a(e) − vai |/|va−max − va−min |, where vai is the threshold value occurring in this elementary condition and va−max , va−min are maximal and minimal values in the domain of this attribute. For nominal attributes present in the elementary condition, distance da is equal to 0 if the description of the classiﬁed object e satisﬁes this condition or 1 otherwise. The coeﬃcient expressing rule similarity (complement of the calculated distance, i.e. 1−DR(r, e)) is used instead of matching factor in the above formula and again the strongest decision Kj wins. While computing this formula we can use also heuristic of choosing the ﬁrst k nearest rules only. More details on this strategy the reader can ﬁnd in papers [32,33,37]. Let us consider a simple example of classifying two objects e1 = {(Age = m), (Job = p), (P eriod = 6), (Income = 3000), (P urpose = K)} and e2 = {(Age = m), (Job = p), (P eriod = 2), (Income = 2600), (P urpose = M )}. The

338

J. Stefanowski

ﬁrst object is completely matched by to one rule no. 3. So, this object is be assigned to class d. The other object does not satisfy condition part of any rules. If we use the ﬁrst strategy for solving no matching case, we can notice that object e2 is partially matched to rules no. 2, 4 and 5. The support for class r is equal to 0.5·2 = 1. The support for class d is equal to 0.5·2 + 0.5·2 = 2. So, the object is assigned to class d. 3.3

Summarizing Experience with Single MODLEM Classiﬁers

Let us shortly summarize the results of studies, where we evaluated the classiﬁcation performance of the single rule classiﬁer induced by MODLEM. There are some options of using this algorithm. First of all one can choose as decision concepts either lower or upper approximations. We have carried out several experimental studies on benchmark data sets from ML Irvine repository [3]. Due to the limited size of this paper, we do not give precise tables but conclude that generally none of approximations was better. The diﬀerences of classiﬁcation accuracies were usually not signiﬁcant or depended on the particular data at hand. This observation is consistent with previous experiments on using certain or possible rules in the framework of LEM2 algorithm [13]. We also noticed that using classiﬁcation strategies while solving ambiguous matching was necessary for all data sets. Again the diﬀerence of applied strategies in case of non-matching (either Grzymala’s proposal or nearest rules) were not signiﬁcant. Moreover, in [14] we performed a comparative study of using MODLEM and LEM2 algorithms on numerical data. LEM2 was used with preprocessing phase with the good discretization algorithm. The results showed that MODLEM can achieved good classiﬁcation accuracy comparable to best pre-discretization and LEM2 rules. Here, we could comment that elements of rough sets are mainly used in MODLEM as a kind of preprocessing, i.e. approximations are decision concepts. Then, the main procedure of this algorithm follows rather the general inductive principle which is common aspect with many machine learning algorithms – see e.g. a discussion of rule induction presented in [23]. Moreover, the idea of handling numerical attributes is somehow consistent with solutions also already present in classiﬁcation tree generation. In this sense, other rule generation algorithms popular in rough sets community, as e.g. based on Boolean reasoning, are more connected with rough sets theory. It is natural to compare performance of MODLEM induced rules against standard machine learning systems. Such a comparative study was carried out in [37,41] and showed that generally the results obtained by MODLEM (with nearest rules strategies) were very similar to ones obtained by C4.5 decision tree.

4

Combined Classiﬁers – General Issues

In the next sections we will study the use of MOLDEM in the framework of the combined classiﬁers. Previous theoretical research (see, e.g., their summary in [6,45]) indicated that combining several classiﬁers is eﬀective only if there is

On Combined Classiﬁers, Rule Induction and Rough Sets

339

a substantial level of disagreement among them, i.e. if they make errors independently with respect to one another. In other words, if they make errors for a given object they should indicate diﬀerent class assignments. Diversiﬁed base classiﬁers can be generated in many ways, for some review see, e.g. [6,36,45]. In general, either homogeneous or heterogeneous classiﬁers are constructed. In the ﬁrst category, the same learning algorithm is used over diﬀerent samples of the data set. The best-known examples are either bagging and boosting techniques which manipulate set of examples by including or weighting particular examples, or methods that manipulate set of attributes, e.g. randomly choosing several attribute subsets. Moreover, multiple classiﬁers could be trained over diﬀerent samples or partitions of data sets. In the second category, diﬀerent learning algorithms are applied to the same data set, and the diversity of results comes from heterogeneous knowledge representations or diﬀerent evaluation criteria used to construct them. The stacked generalization or meta-learning belong to this category. In section 7 we study the combiner as one of these methods. Combining classiﬁcation predictions from single classiﬁers is usually done by group or specialized decision making. In the ﬁrst method all base classiﬁers are consulted to classify a new object while the other method chooses only these classiﬁers whose are expertised for this object. Voting is the most common method used to combine single classiﬁers. The vote of each classiﬁer may be weighted, e.g., by an evaluation of its classiﬁcation performance. Moreover, looking into the rough sets literature one can notice a growing research interest in constructing more complex classiﬁcation system. First works concerned rather an intelligent integration of diﬀerent algorithms into hybrid system. For instance, some researchers tried to reﬁne rule classiﬁers by analysing relationships with neural networks [44]. More related works included an integration of k - nearest neighbor with rough sets rule generation, see e.g. RIONA system, which oﬀered good classiﬁcation performance [8]. Yet another approach comprises two level knowledge representation: rules induced by Explore representing general patterns in data and case base representing exceptions [36], which worked quite well for the diﬃcult task of credit risk prediction [43]. Recently Skowron and his co-operators have been developing hierarchical classiﬁers which attempt at approximating more complex concepts [2]. Classiﬁers on different hierarchy level correspond to diﬀerent levels of pattern generalization and seems to be a speciﬁc combination of multiple models, which could be obtained in various ways, e.g. using a special lattice theory [46] or leveled rule generation. Nguyen et al. described in [24] an application concerning detecting sunspots where hierarchical classiﬁer is constructed with a domain knowledge containing an ontology of considered concepts.

5

Using MODLEM Inside the Bagging

Firstly, we consider the use of MODLEM induced classiﬁer inside the most popular homogeneous multiple classiﬁers [38].

340

J. Stefanowski

This approach was originally introduced by Breiman [4]. It aggregates classiﬁers generated from diﬀerent bootstrap samples. The bootstrap sample is obtained by uniformly sampling with replacement objects from the training set. Each sample has the same size as the original set, however, some examples do not appear in it, while others may appear more than once. For a training set with m examples, the probability of an example being selected at least once is 1−(1−1/m)m. For a large m, this is about 1 - 1/e. Given the parameter R which is the number of repetitions, R bootstrap samples S1 , S2 , . . . , SR are generated. From each sample Si a classiﬁer Ci is induced by the same learning algorithm and the ﬁnal classiﬁer C ∗ is formed by aggregating these R classiﬁers. A ﬁnal classiﬁcation of object x is built by a uniform voting scheme on C1 , C2 , . . . , CR , i.e. is assigned to the class predicted most often by these sub-classiﬁers, with ties broken arbitrarily. For more details and theoretical justiﬁcation see e.g. [4]. Table 2. Comparison of classiﬁcation accuracies [%] obtained by the single MODLEM based classiﬁer and the bagging approach; R denotes the number of component classiﬁers inside bagging Name of data set

Single classiﬁer

Bagging

R

bank buses zoo hepatitis hsv iris automobile segmentation glass bricks vote bupa election urology german crx pima

93.81 ± 0.94 97.20 ± 0.94 94.64 ± 0.67 78.62 ± 0.93 54.52 ± 1.05 94.93 ± 0.5 85.23 ± 1.1 85.71 ± 0.71 72.41 ± 1.23 90.32* ± 0.82 92.67 ± 0.38 65.77 ± 0.6 88.96± 0.54 63.80 ± 0.73 72.16 ± 0.27 84.64 ± 0.35 73.57 ± 0.67

95.22 ± 1.02 99.54 ± 1.09 93.89* ± 0.71 84.05 ± 1.1 64.78 ± 0.57 95.06* ± 0.53 83.00 ±0.99 87.62 ± 0.55 76.09 ± 0.68 91.21* ± 0.48 96.01 ± 0.29 76.28 ± 0.44 91.66 ± 0.34 67.40 ± 0.46 76.2 ± 0.34 89.42 ± 0.44 77.87 ± 0.39

7 5 7 5 7 5 5 7 10 7 10 5 7 7 5 10 7

In this paper we shortly summarize main results obtained in the extensive computational study [38]. The MODLEM algorithm was applied to generate base classiﬁers in the bagging combined classiﬁer. In table 2 we present the comparison of the classiﬁcation accuracy obtained for the best variant of the bagging against the single rule classiﬁer (also induced by MODLEM). The experiments were carried out on several data sets coming mainly from ML Irvine repository [3]. For each data set, we show the classiﬁcation accuracy obtained by a single classiﬁer over the 10 cross-validation loops. A standard deviation is also given. An asterisk

On Combined Classiﬁers, Rule Induction and Rough Sets

341

indicates that the diﬀerence for these compared classiﬁers and a given data set is not statistically signiﬁcant (according to two-paired t-Student test with α=0.05). The last column presents the number of R component classiﬁers inside the bagging - more details on tuning this value are described in [38]. We conclude that results of this experiment showed that the bagging significantly outperformed the single classiﬁer on 14 data sets out of total 18 ones. The diﬀerence between classiﬁers were non-signiﬁcant on 3 data sets (those which were rather easy to learn as, e.g. iris and bricks - which were characterized by a linear separation between classes). Moreover, we noticed the slightly worse performance of the bagging for quite small data (e.g. buses, zoo - which seemed to be too small for sampling), and signiﬁcantly better for data sets containing a higher number of examples. For some of these data sets we observed an substantial increase of predictive accuracy, e.g. for hsv – over 10%, bupa – around 10% and hepatitis – 5.43%. However, we should admit that this good performance was expected as we know that there are many previous reports on successful use of decision trees in bagging or boosting.

6

On Solving Multiclass Problems with the n2 -Classiﬁer

One can say the bagging experiment has been just a variant of a standard approach. Now we will move to more original approach, called the n2 -classiﬁer, which was introduced by Jelonek and author in [18,36]. This kind of a multiple classiﬁer is a specialized approach to solve multiple class learning problems. The n2 -classiﬁer is composed of (n2 − n)/2 base binary classiﬁers (where n is a number of decision classes; n > 2). The main idea is to discriminate each pair of the classes: (i, j), i, j ∈ [1..n], i = j , by an independent binary classiﬁer Cij . Each base binary classiﬁer Cij corresponds to a pair of two classes i and j only. Therefore, the speciﬁcity of the training of each base classiﬁer Cij consists in presenting to it a subset of the entire learning set that contains only examples coming from classes i and j. The classiﬁer Cij yields a binary classiﬁcation indicating whether a new example x belongs to class i or to class j. Let us denote by Cij (x) the classiﬁcation of an example x by the base classiﬁer Cij . The complementary classiﬁers: Cij and Cji (where i, j ∈ < 1 . . . n >; i = j) solve the same classiﬁcation problem – a discrimination between class i-th and j-th. So, they are equivalent (Cij ≡ Cji ) and it is suﬃcient to use only (n2 - n)/2 classiﬁers Cij (i < j), which correspond to all combinations of pairs of n classes. An algorithm providing the ﬁnal classiﬁcation assumes that a new example x is applied to all base classiﬁers Cij . As a result, their binary predictions Cij (x) are computed. The ﬁnal classiﬁcation is obtained by an aggregation rule, which is based on ﬁnding a class that wins the most pairwise comparisons. The more sophisticated approach includes a weighted majority voting rules, where the vote of each classiﬁer is modiﬁed by its credibility, which is calculated as its classiﬁcation performance during learning phase; more details in [18].

342

J. Stefanowski

We have to remark that the similar approach was independently studied by Friedman [7] and by Hastie and Tibshirani [17] – they called it classiﬁcation by pairwise coupling. The experimental studies, e.g. [7,17,18], have shown that such multiple classiﬁers performed usually better than the standard classiﬁers. Previously the author and J.Jelonek have also examined the inﬂuence of a learning algorithm on the classiﬁcation performance of the n2 -classiﬁer.

Table 3. Comparison of classiﬁcation accuracies [%] and computation times [s] for the single MODLEM based classiﬁer and the n2 -classiﬁer also based on decision rules induced by MODLEM algorithm Accuracy of Name of single data set MODLEM (%) automobile 85.25 ± 1.3 cooc 55.57 ± 2.0 ecoli 79.63 ± 0.8 glass 72.07 ± 1.2 hist 69.36 ± 1.1 meta-data 47.2 ± 1.3 iris 94.2 ± 0.6 soybean-large 91.09 ± 0.9 vowel 81.81 ± 0.5 yeast 54.12 ± 0.7 zoo 94.64 ± 0.5

Accuracy of n2M ODLEM (%) 87.96 ± 1.5 59.30 ± 1.4 81.34 ± 1.7 74.82 ± 1.4 73.10 ± 1.4 49.83 ± 1.9 95.53* ± 1.2 91.99* ± 0.8 83.79 ± 1.2 55.74 ± 0.9 94.46* ± 0.8

Time of comput. MODLEM 15.88 ± 0.4 4148,7 ± 48.8 27.53 ± 0.5 45.29 ± 1.1 3563.79 ± 116.1 252.59 ± 78.9 0.71 ± 0.04 26.38 ± 0.3 3750.57 ± 30.4 1544.3 ± 13.2 0.30 ± 0.02

Time of comput. n2M ODLEM 5.22 ± 0.3 431.51 ± 1.6 11.25 ± 0.7 13.88 ± 0.4 333.96 ± 0.8 276.71 ± 5.21 0.39 ± 0.04 107.5 ± 5.7 250.63 ± 0.7 673.82 ± 9.4 0.34 ± 0.12

Here, we summarize these of our previous results, where the MODLEM was applied to generate base classiﬁers inside the n2 -classiﬁer [38]. In table 3 we present classiﬁcation accuracies obtained by the n2 -classiﬁer and compare them against the single rule classiﬁer induced by MODLEM on 11 data sets, all concerning multiple-class learning problems, with a number of classes varied from 3 up to 14. The second and third columns are presented in a similar way as in Table 2. These results showed that the n2 -classiﬁer signiﬁcantly (again in the sense of paired t test with a signiﬁcance level α = 0.05) outperformed the single classiﬁer on 7 out of 11 problems, e.g. for hist – over 3.7%, glass – around 2.7%, automobile – 2.5% and meta-data – 2.6%. These improvements were not so high as in the bagging but still they occurred for many diﬃcult multi-class problems. Again, the multiple classiﬁer was not useful for easier problems (e.g. iris). Moreover, we noticed that its performance was better for data sets with a higher number of examples. Coming back to our previous results for the n2 -classiﬁer [18] we can again remark that the comparable classiﬁcation improvements were observed for the case of using decision trees. Then, let us focus our attention on interesting phenomena concerning computation costs of using the MODLEM in a construction of the n2 -classiﬁer. Table 3 (two last columns) contains computation times (in seconds calculated as average

On Combined Classiﬁers, Rule Induction and Rough Sets

343

values over 10 folds with standard deviations). We can notice that generally constructing a combined classiﬁers does not increase the computation time. What is even more astonishing, for some data sets constructing the n2 -classiﬁer requires even less time than training the standard single classiﬁer. Here, we have to stress that in our previous works [18,37] we noticed that the increase of classiﬁcation accuracy (for other learning algorithms as e.g. decision trees, k-nearest neighbor or neural networks) was burden with increasing the computational costs (sometimes quite high). In [38] we attempted to explain the good performance of MODLEM inside the n2 -classiﬁer. Shortly speaking, the n2 -classiﬁer should be rather applied to solving diﬃcult (”complex”) classiﬁcation tasks, where examples of decision classes are separated by non-linear decision borders – these are often diﬃcult concepts to be learned by standard classiﬁers, while pairwise decision boundaries between each pair of classes may be simpler and easier to be learned with using a smaller number of attributes. Here, MODLEM could gain its performance thanks to his sequential covering and greedy heuristic search. It generates rules distinguishing smaller number of learning examples (from two classes only) than in the multiple class case and, above all, testing a smaller number of elementary conditions. To verify hypothesis we inspect syntax of rule sets induced by the single classiﬁer and the n2 -classiﬁer. Rules for binary classiﬁers were using less attributes and covered more learning example on average than rules from the single set generated in the standard way [38].

7

Combining Predictions of Heterogeneous Classiﬁers

In two previous sections we described the use of MODLEM based classiﬁers inside the architecture of homogeneous classiﬁers. In these solutions, the MODLEM was the only algorithm applied to create base classiﬁers inside multiple classiﬁers and could directly inﬂuence their ﬁnal performance. Diversiﬁcation of base classiﬁers is one of the conditions for improving classiﬁcation performance of the ﬁnal system. Let us repeat that in previously considered solutions it was achieved by changing the distribution of examples in the input data. Another method to obtain component classiﬁer diversity is constructing, so called, heterogeneous classiﬁers. They are generated from the same input data by diﬀerent learning algorithms which use diﬀerent representation language and search strategies. These base classiﬁers could be put inside a layered architecture. At the ﬁrst level base classiﬁers receive the original data as input. Their predictions are then aggregated at the second level into the ﬁnal prediction of the system. This could be done in various ways. In one of our studies we used a solution coming from Chan & Stolfo [5], called a combiner. The combiner is based on an idea of merging predictions of base classiﬁers by an additional classiﬁer, called meta-classiﬁer. This is constructed in an extra meta-learning step, i.e. ﬁrst base classiﬁers are learned, then their predictions made on a set of extra validation examples, together with correct decision labels, form a meta-level training set. An extra learning algorithm is applied to this set to discover how to merge base classiﬁer predictions into a ﬁnal decision.

344

J. Stefanowski Table 4. Classiﬁcation accuracies [%] for diﬀerent multiple classiﬁers Data set

Bagging n2 -classiﬁer Combiner

Automobile Bank Bupa Ecoli Glass HSV Meta-data Pima Voting Yeast Zoo

83.00 95.22 76.28 85.70 74.82 64.75 48.11 75.78 93.33 58.18 93.89

87.90 – – 81.34 74.82 – 49.80 – – 55.74 94.46

84.90 95.45 69.12 85.42 71.50 59.02 51.33 74.78 94.67 58.36 95.05

In [41] we performed a comparative study of using a combiner approach against the single classiﬁers learned by these algorithms which were applied to create its component classiﬁers. In this study base classiﬁers were induced by k-NN, C4.5 and MODLEM. The meta-classiﬁer was a Naive Bayes. This comparative study was performed on 15 data sets. However, the obtained results showed that the combiner did not improve classiﬁcation accuracy in so many cases as previously studied homogeneous classiﬁers. Only in 33% data we observed a signiﬁcant improvement comparing against single classiﬁers. In table 4 we present only some of these results concerning the ﬁnal evaluation of the combiner compared also against the previous multiple classiﬁers. However, while comparing these classiﬁers we should be cautious as the number of the results on common data sets was limited. Moreover, MODLEM is only one of three component classiﬁers inside the combiner that inﬂuences the ﬁnal result. We could also ask a question about other elements of the architecture of heterogeneous classiﬁer, e.g. number of component classiﬁers or the aggregation techniques. In recent experiments we focus our interest on testing two other techniques instead of the meta-combiner: – a simple aggregation performed by means of a majority voting rule (denoted as MV in table 4), – using a quite sophisticated approach – SCANN; It was introduced by Merz [22] and uses a mechanism of the correspondence analysis to discover hidden relationships between the learning examples and the classiﬁcation done by the component classiﬁers. Results from ongoing experiments are given in Table 5. There is also a diﬀerence to previous architecture, i.e. adding an additional, forth component classiﬁers Naive Bayesian at the ﬁrst level. We can remark that the more advanced aggregation technique could slightly increase the classiﬁcation accuracy comparing to simpler one. On the other hand they are much time consuming.

On Combined Classiﬁers, Rule Induction and Rough Sets

345

Table 5. Comparison of diﬀerent methods producing the ﬁnal decision inside the heterogeneous classiﬁers - classiﬁcation accuracies [%] Data set credit-a 86.2 glass 68.5 ecoli 86.1 zoo 95

8

MV ± 0.6 ± 0.3 ± 0.9 ± 0.9

SCANN Combiner 87 70.1 81.5 92.2

± 0.7 ± 0.2 ± 0.8 ± 0.7

86.6 70.5 84.5 95.1

± 0.4 ± 0.6 ± 0.5 ± 0.4

Discussion of Results and Final Remarks

As Professor Zdzislaw Pawlak wrote in the introductory chapter of his book on rough sets [26] knowledge of human beings and other species is strictly connected with their ability to classify objects. Finding classiﬁcation patterns of sensor signals or data form fundamental mechanisms for very living being. In his point of view it was then connected with a partition (classiﬁcation) operation leading to basic blocks for constructing knowledge. Many researchers followed the Pawlak’s idea. One of the main research directions includes constructing approximations of knowledge from tables containing examples of decisions on object classiﬁcation. Rules were often induced as the most popular knowledge representation. They could be used either to describe the characteristics of available data or as the basis for supporting classiﬁcation decisions concerning new objects. Up to now several eﬃcient rule classiﬁers have been introduced. In this study we have attempted to brieﬂy describe the current experience with using the author’s rule induction algorithm MODLEM, which induces either certain or possible rules from appropriate rough approximations. This is the main point where elements of the rough sets theory is applied in this algorithm. Given as an input learning examples from approximations, the rule generation phase follows the general idea of sequential covering, which is somehow in common with machine learning paradigms. The MODLEM produces a minimal set of rules covering examples from rough approximations. This rule sets should be joined with classiﬁcation strategies for solving ambiguous matching of the new object description to condition parts of rules. An extra property of this algorithm is it ability to handle directly numerical attributes without prior discretization. The current experience with comparative studies on benchmark data sets and real life applications showed that the classiﬁcation performance of this approach was comparable to other symbolic classiﬁers, in particular to decision trees. Although the MODLEM classiﬁer and other machine learning approaches are eﬃcient for many classiﬁcation problems, they do not always lead to satisfactory classiﬁcation accuracy for more complex and diﬃcult problems. This is our motivation to consider new approaches for increasing classiﬁcation accuracy by combining several classiﬁers into an integrated system. Several proposals of

346

J. Stefanowski

constructing such multiple classiﬁers are already proposed. Most of them are general approaches, where many diﬀerent algorithms could be applied to induce the component classiﬁers. Thus, our main research interest in this study is to summarize our experiments with using MODLEM induced rule classiﬁers inside the framework of three diﬀerent multiple classiﬁers, namely the bagging, the n2 -classiﬁer and the combiner. A classiﬁcation accuracy for the multiple classiﬁer has been compared against the standard classiﬁers – also induced by MODLEM. These results and their detailed discussion has been given in the previous sections. Firstly we could notice that using MODLEM inside the bagging was quite eﬀective. However, it was a kind of standard approach and we could expect such good performance as MODLEM performs similarly to decision trees (which have been extensively studied in the bagging) and could be seen as unstable learning algorithm - i.e. an algorithm whose output classiﬁer undergoes changes in response to small changes in the training data. This kind of algorithm may produce base classiﬁers diversiﬁed enough (but not too much, see e.g. discussion of experimental study by Kuncheva and Whitaker [21]) which is a necessary condition for their eﬀective aggregation. Following the same arguments we also suspect that MODLEM should nicely work inside the boosting classiﬁer. Further on, we could hypothesize that slightly worse improvements of the classiﬁcation accuracy in the combiner approach may result from insuﬃcient diversiﬁcation of component heterogeneous classiﬁers. This has been veriﬁed by analysing distributions of wrong decisions for base classiﬁers, presented in [41]. It showed the correlation of errors for some data sets, where ﬁnally we did not notice the improvement of the classiﬁcation accuracy. The most original methodological approach is Jelonek and author’s proposal of the n2 -classiﬁer which is in fact a specialized approach to learning multiple class problems. The n2 -classiﬁer is particularly well suited for multiple class data where exist ”simpler” pairwise decision boundaries between pair of classes. MODLEM seems to be a good choice to be used inside this framework as it leads to an improvement of classiﬁcation performance and does not increase computational costs - reasons for this have been discussed in section 7. Let us notice that using other learning algorithms inside the n2 -classiﬁer and applying MODLEM in two other multiple classiﬁer requires an extra computation eﬀorts comparing to learning the single, standard classiﬁer [38]. Comparing results of all together multiple classiﬁers ”head to head” we should be cautious as we had a limited number of common data sets. It seems that the n2 -classiﬁer is slightly better for these data. While the standard multiple classiﬁers, as bagging or combiner, are quite eﬃcient for simpler data and are easier to be implemented. To sum up, the results of our experiments have shown that the MODLEM algorithm can be eﬃciently used within the framework of three multiple classiﬁers for data sets concerning more ”complex” decision concepts. However, the relative merits of these new approaches depends on the speciﬁes of particular problems and a training sample size.

On Combined Classiﬁers, Rule Induction and Rough Sets

347

Let us notice that there is a disadvantage of the multiple classiﬁers - loosing a simple and easy interpretable structure of knowledge represented in a form decision rules. These are ensembles of diversiﬁed rule sets specialized for predictive aims not one set of rules in a form for a human inspection. As to future research directions we could consider yet another way of obtaining diversiﬁed data – i.e. selecting diﬀerent subsets of attributes for each component classiﬁers. The author has already started research on extending bootstrap samples inside the bagging by applying additionally attribute selection [39,40]. In this way each bootstrap is replicated few times, each of them using diﬀerent subset of attributes. We have considered the use of diﬀerent selection techniques and observed that besides random choice or wrapper model, techniques which use either entropy based measures or correlation merits are quite useful. The results of comparative experiments carried out in [40] have showed that the classiﬁcation accuracy of such a new extended bagging is higher than for standard one. In this context one could come back to the classical rough sets topic of reducts, which relates to ﬁnding an ensemble of few attribute subsets covering diﬀerent data properties and constructing in this way a set of diversiﬁed examples for an integrated system. However, we are not limited to ”classical” meaning of pure rough sets reducts but rather to approximate ones, where the entropy measure is also considered [30]. Acknowledgment. The author would like to thank his colleagues Jacek Jelonek, Slawomir Nowaczyk and his M.Sc. students Michal Bro´ nczyk, Ryszard Gizelski, Maciej L uszczy´ nski who have worked with him on the software implementations of the classiﬁers or took part in some experiments.

References 1. Bazan J.: A comparison of dynamic and non-dynamic rough set methods for extracting laws from decision tables. In Polkowski L., Skowron A. (eds.), Rough Sets in Data Mining and Knowledge Discovery vol. 1, Physica-Verlag, 1998, 321–365. 2. Bazan J., Nguyen Hung Son, Skowron A.: Rough sets methods in approximation of hierarchical concepts. In Proc. of the Conference on Rough Sets and New Trends in Computing, RSCTC – 2004, LNAI 2066, Springer Verlag, 2004, 346–355. 3. Blake C., Koegh E., Mertz C.J.: Repository of Machine Learning, University of California at Irvine (1999). 4. Breiman L.: Bagging predictors. Machine Learning, 24 (2), 1996, 123–140. 5. Chan P.K., Stolfo S.: On the accuracy of meta-learning for scalable data mining. Journal of Intelligent Information Systems, 8, (1), 1997, 5-28. 6. Dietrich T.G.: Ensemble methods in machine learning. In Proc. of 1st Int. Workshop on Multiple Classiﬁer Systems, 2000, 1–15. 7. Friedman J.: Another approach to polychotomous classiﬁcation, Technical Report, Stanford University, 1996. 8. G´ ora G., Wojna A.: RIONA: a new classiﬁcation system combining rule induction and instance based learning. Fundamenta Informaticae 51 (4), 2002, 369-390. 9. Greco S., Matarazzo B., Slowi´ nski R.: The use of rough sets and fuzzy sets in MCDM. In Gal T., Stewart T., Hanne T. (eds), Advances in Multiple Criteria Decision Making, Kluwer, chapter 14, 1999, pp. 14.1-14.59.

348

J. Stefanowski

10. Greco S., Matarazzo B., Slowi´ nski R., Stefanowski J.: Variable consistency model of dominance-based rough set approach. In Proc. 2nd Int. Conference on Rough Sets and New Trends in Computing, RSCTC – 2000, LNAI 2005, Springer Verlag, 2001,170–181. 11. Grzymala-Busse J.W. LERS - a system for learning from examples based on rough sets. In Slowinski R. (ed.), Intelligent Decision Support, Kluwer Academic Publishers, 1992, 3–18. 12. Grzymala-Busse J.W.: Managing uncertainty in machine learning from examples. In Proc. 3rd Int. Symp. in Intelligent Systems, Wigry, Poland, IPI PAN Press, 1994, 70–84. 13. Grzymala-Busse J.W. Zou X.: Classiﬁcation strategies using certain and possible rules. In Proceedings of the 1th Rough Sets and Current Trends in Computing Conference, RSCTC–98 , LNAI 1424, Springer Verlag, 1998, 37-44. 14. Grzymala-Busse J.W., Stefanowski J.: Three approaches to numerical attribute discretization for rule induction. International Journal of Intelligent Systems, 16 (1), (2001) 29–38. 15. Grzymala-Busse J.W., Stefanowski J. Wilk Sz.: A comparison of two approaches to data mining from imbalanced data. In Proc. of the KES 2004 - 8-th Int. Conf. on Knowledge-based Intelligent Information & Engineering Systems, Springer LNCS vol. 3213, 2004, 757-763. 16. Han J., Kamber M.: Data mining: Concepts and techniques, San Francisco, Morgan Kaufmann, 2000. 17. Hastie T., Tibshirani R.: Classiﬁcation by pairwise coupling. In Jordan M.I. (ed.) Advances in Neural Information Processing Systems: 10 (NIPS-97), MIT Press, 1998, 507-513. 18. Jelonek J., Stefanowski J.: Experiments on solving multiclass learning problems by the n2 -classiﬁer. In Proceedings of 10th European Conference on Machine Learning ECML 98, Springer LNAI no. 1398, 1998, 172–177. ˙ 19. Klosgen W., Zytkow J.M. (eds.): Handbook of Data Mining and Knowledge Discovery, Oxford Press, 2002. 20. Komorowski J., Pawlak Z., Polkowski L. Skowron A.: Rough Sets: tutorial. In Pal S.K., Skowron A. (eds) Rough Fuzzy Hybridization. A new trend in decision making, Springer Verlag, Singapore, 1999, 3–98. 21. Kuncheva L., Whitaker C.J.: Measures of diversity in classiﬁer ensembles and their relationship with the ensemble accuracy. Machine Learning, 51, 2003, 181–207. 22. Merz C.: Using correspondence analysis to combine classiﬁers. Machine Learning, 36 (1/2), 1999, 33–58. 23. Mitchell Tom M.: Machine learning, McGraw Hill, 1997. 24. Nguyen Sinh Hoa, Trung Tham Nguyen, Nguyen Hung Son: Rough sets approach to sunspot classiﬁcation problem. In Proc. of the Conference RSFDGrC – 2005, vol 2, LNAI 3642, Springer Verlag, 2005, 263-272. 25. Pawlak Z.: Rough sets. Int. J. Computer and Information Sci., 11, 1982, 341–356. 26. Pawlak Z.: Rough sets. Theoretical aspects of reasoning about data. Kluwer Academic Publishers, Dordrecht, 1991. 27. Pawlak Z., Grzymala-Busse J., Slowinski R., Ziarko W.: Rough sets. Communications of the ACM, vol. 38, no. 11, 1995, 89-95. 28. Skowron A.: Boolean reasoning for decision rules generation. In Komorowski J., Ras Z. (des.) Methodologies for Intelligent Systems, LNAI 689, Springer-Verlag, 1993, 295–305.

On Combined Classiﬁers, Rule Induction and Rough Sets

349

29. Skowron A., Rauszer C.: The discernibility matrices and functions in information systems. In Slowinski R. (ed.), Intelligent Decision Support. Handbook of Applications and Advances of Rough Set Theory. Kluwer Academic Publishers, 1992, 331–362. 30. Slezak D.: Approximate entropy reducts. Fundamenta Informaticae 53 (3/4), 2002, 365-387. 31. Slowinski R., Greco S.: Inducing Robust Decision Rules from Rough Approximations of a Preference Relation. In Rutkowski L. et al. (eds): Artiﬀcial Intelligence and Soft Computing, LNAI 3070, Springer-Verlag, 2004, 118-132. 32. Stefanowski J.: Classiﬁcation support based on the rough sets. Foundations of Computing and Decision Sciences, vol. 18, no. 3-4, 1993, 371-380. 33. Stefanowski J.: Using valued closeness relation in classiﬁcation support of new objects. In Lin T. Y., Wildberger (eds) Soft computing: rough sets, fuzzy logic, neural networks uncertainty management, knowledge discovery, Simulation Councils Inc., San Diego CA, 1995, 324–327. 34. Stefanowski J.: On rough set based approaches to induction of decision rules. In Polkowski L., Skowron A. (eds), Rough Sets in Data Mining and Knowledge Discovery, vol. 1, Physica-Verlag, 1998, 500–529. 35. Stefanowski J.: The rough set based rule induction technique for classiﬁcation problems. In Proceedings of 6th European Conference on Intelligent Techniques and Soft Computing EUFIT 98, Aachen 7-10 Sept., 1998, 109–113. 36. Stefanowski J.: Multiple and hybrid classiﬁers. In Polkowski L. (ed.) Formal Methods and Intelligent Techniques in Control, Decision Making, Multimedia and Robotics, Post-Proceedings of 2nd Int. Conference, Warszawa, 2001, 174–188. 37. Stefanowski J.: Algorithims of rule induction for knowledge discovery. (In Polish), Habilitation Thesis published as Series Rozprawy no. 361, Poznan Univeristy of Technology Press, Poznan (2001). 38. Stefanowski J.: The bagging and n2-classiﬁers based on rules induced by MODLEM. In Proceedings of the 4th Int. Conference Rough Sets and Current Trends in Computing, RSCTC – 2004, LNAI 3066, Springer-Verlag, 2004, 488-497. 39. Stefanowski J.: An experimental study of methods combining multiple classiﬁers diversiﬁed both by feature selection and bootstrap sampling. In K.T. Atanassov, J. Kacprzyk, M. Krawczak, E. Szmidt (eds), Issues in the Representation and Processing of Uncertain and Imprecise Information, Akademicka Oﬁcyna Wydawnicza EXIT, Warszawa, 2005, 337-354. 40. Stefanowski J., Kaczmarek M.: Integrating attribute selection to improve accuracy of bagging classiﬁers. In Proc. of the AI-METH 2004. Recent Developments in Artiﬁcial Intelligence Methods, Gliwice, 2004, 263-268. 41. Stefanowski J., Nowaczyk S.: On using rule induction in multiple classiﬁers with a combiner aggregation strategy. In Proc. of the 5th Int. Conference on Intelligent Systems Design and Applications - ISDA 2005, IEEE Press, 432-437. 42. Stefanowski J., Vanderpooten D.: Induction of decision rules in classiﬁcation and discovery-oriented perspectives. International Journal of Intelligent Systems 16 (1), 2001, 13–28. 43. Stefanowski J., Wilk S.: Evaluating business credit risk by means of approach integrating decision rules and case based learning. International Journal of Intelligent Systems in Accounting, Finance and Management 10 (2001) 97–114. 44. Szczuka M: Reﬁning classiﬁers with neural networks. International Journal of Intelligent Systems 16 (1), 2001, 39–56.

350

J. Stefanowski

45. Valentini G., Masuli F.: Ensambles of learning machines. In R. Tagliaferri, M. Marinaro (eds), Neural Nets WIRN Vietri-2002, Springer-Verlag LNCS, vol. 2486, 2002 , 3–19. 46. Wang H., Duntsch I., Gediga G., Skowron A.: Hyperrelations in version space. International Journal of Approximate Reasoning, 23, 2000, 111–136. 47. Ziarko W.: Variable precision rough sets model. Journal of Computer and Systems Sciences, vol. 46. no. 1, 1993, 39–59.

Approximation Spaces in Multi Relational Knowledge Discovery Jaroslaw Stepaniuk Department of Computer Science, Bialystok University of Technology Wiejska 45a, 15-351 Bialystok, Poland jstepan@ii.pb.bialystok.pl

Abstract. Pawlak introduced approximation spaces in his seminal work on rough sets more than two decades ago. In this paper, we show that approximation spaces are basic structures for knowledge discovery from multi-relational data. The utility of approximation spaces as fundamental objects constructed for concept approximation is emphasized. Examples of basic concepts are given throughout this paper to illustrate how approximation spaces can be beneﬁcially used in many settings. The contribution of this paper is the presentation of an approximation space-based framework for doing research in various forms of knowledge discovery in multi relational data. Keywords: rough sets, approximation spaces, multi-relational data mining, rough inclusion, uncertainty function.

1

Introduction

Approximation spaces are fundamental structures for the rough set approach [7,8,10]. In this paper we present a generalization of the original approximation space model. Using such approximation spaces we show how the rough set approach can be used for approximation of concepts assuming that only partial information on approximation spaces is available. Hence, searching for concept approximation, i.e., the basic task in machine learning and pattern recognition can be formulated as searching for relevant approximation spaces. Rough set approach has been used in a lot of applications aimed at description of concepts. In most cases, only approximate descriptions of concepts can be constructed because of incomplete information about them. In learning approximations of concepts, there is a need to choose a description language. This choice may limit the domains to which a given algorithm can be applied. There are at least two basic types of objects: structured and unstructured. An unstructured object is usually described by attribute-value pairs. For objects having an internal structure ﬁrst order logic language is often used. Attribute-value languages have the expressive power of propositional logic. These languages sometimes do not allow for proper representation of complex structured objects and relations among objects or their components. The background knowledge that can be J.F. Peters et al. (Eds.): Transactions on Rough Sets VI, LNCS 4374, pp. 351–365, 2007. c Springer-Verlag Berlin Heidelberg 2007

352

J. Stepaniuk

used in the discovery process is of a restricted form and other relations from the database cannot be used in the discovery process. Using ﬁrst-order logic (or FOL for short) has some advantages over propositional logic [1,2,4]. First order logic provides a uniform and very expressive means of representation. The background knowledge and the examples, as well as the induced patterns, can all be represented as formulas in a ﬁrst order language. Unlike propositional learning systems, the ﬁrst order approaches do not require that the relevant data be composed into a single relation but, rather they can take into account data organized in several database relations with various connections existing among them. The paper is organized as follows. In Section 2 we recall the deﬁnition of approximation spaces. Next, we describe a constructive approach for computing values of uncertainty and rough inclusion functions. These functions are the basic components of approximation spaces. Parameters of the uncertainty and rough inclusion functions are tuned in searching for relevant approximation spaces. Among such parameters we distinguish sensory environments and their extensions. These parameters are used for constructive deﬁnition of uncertainty and rough inclusion functions. In Section 3 we discuss notions of relational learning. In Sections 4 and 5 we consider application of rough set methods to discovery of interesting patterns expressed in a ﬁrst order language. In Section 4 rough set methodology is used in the process of translating ﬁrst–order data into attribute– value data. Some properties of this algorithm were presented in [13]. In Section 5 rough set methodology is used in the process of selecting literals which may be a part of a rule. The criterion of selecting a literal is as follows: only such a literal is selected which added to the rule makes the rule discerning most of the examples which were indiscernible so far. Some properties of this algorithm were presented in [14,15].

2

Approximation Spaces

In this section we recall the deﬁnition of an approximation space from [10,13,11]. Deﬁnition 1. A parameterized approximation space is a system AS#,$ = (U, I# , ν$ ), where – U is a non-empty set of objects, – I# : U → P (U ) is an uncertainty function, where P (U ) denotes the power set of U , – ν$ : P (U ) × P (U ) → [0, 1] is a rough inclusion function, and #, $ denote vectors of parameters (the indexes #, $ will be omitted if it does not lead to misunderstanding). 2.1

Uncertainty Function

The uncertainty function deﬁnes for every object x, a set of objects described similarly to x. The set I(x) is called the neighborhood of x (see, e.g., [8,10]).

Approximation Spaces in Multi Relational Knowledge Discovery

353

We assume that the values of the uncertainty function are deﬁned using a sensory environment [11], i.e., a pair (Σ, · U ), where Σ is a set of formulas, called the sensory formulas, and ·U : Σ −→ P (U ) is the sensory semantics. We assume that for any sensory formula α and any object x ∈ U the information whether x ∈ αU holds is available. The set {α : x ∈ αU } is called the signature of x in AS and is denoted by InfAS (x). For any x ∈ U, the set NAS (x) of neighborhoods of x in AS is deﬁned by {αU : x ∈ αU } and from this set the neighborhood I(x) is constructed. For example, I(x) is deﬁned by selecting an element from the set {αU : x ∈ αU } or by I(x) = NAS (x). Observe that any sensory environment (Σ, · U ) can be treated as a parameter of I from the vector # (see Deﬁnition 1). Let us consider two examples. Any decision table DT = (U, A, d) [8] deﬁnes an approximation space ASDT = (U, IA , νSRI ), where, as we will see, IA (x) = {y ∈ U : a(y) = a(x) for all a ∈ A}. Any sensory formula is a descriptor, i.e., a formula of the form a = v where a ∈ A and v ∈ Va with the standard semantics a = vU = {x ∈ U : a(x) = v}. Then, for any x ∈ U its signature Inf ASDT (x) is equal to {a = a(x) : a ∈ A} and the neighborhood IA (x) is equal to NASDT (x). Another example can be obtained assuming that for any a ∈ A there is given a tolerance relation τa ⊆ Va × Va (see, e.g., [10]). Let τ = {τa }a∈A . Then, one can consider a tolerance decision table DTτ = (U, A, d, τ ) with tolerance descriptors a =τa v and their semantics a =τa vU = {x ∈ U : vτa a(x)}. Any such tolerance decision table DTτ = (U, A, d, τ ) deﬁnes the approximation signature InfASDTτ (x) = {a =τa a(x) : a ∈ A} and the space ASDTτ with the neighborhood IA (x) = NASDTτ (x) for any x ∈ U . The fusion of NASDTτ (x) for computing the neighborhood of x can have many diﬀerent forms, the intersection is only an example. For example, to compute the value of I(x) some subfamilies of NAS (x) may ﬁrst be selected and the family consisting of intersection of each such a subfamily is next taken as the value of I(x). 2.2

Rough Inclusion Function

One can consider general constraints which the rough inclusion functions should satisfy. Searching for such constraints initiated investigations resulting in creation and development of rough mereology (see, the bibliography in [9]). In this subsection, we present some examples of rough inclusion functions only. The rough inclusion function ν$ : P (U ) × P (U ) → [0, 1] deﬁnes the degree of inclusion of X in Y , where X, Y ⊆ U . In the simplest case it can be deﬁned by (see, e.g., [10,8]): card(X∩Y ) card(X) if X = ∅ νSRI (X, Y ) = 1 if X = ∅. This measure is widely used by the data mining and rough set communities. It is worth mentioning that Jan L ukasiewicz [3] was the ﬁrst one who used this idea

354

J. Stepaniuk

to estimate the probability of implications. However, rough inclusion can have a much more general form than inclusion of sets to a degree (see, e.g., [9]). Another example of rough inclusion is used for relation approximation [12] and in the variable precision rough set approach [16]. 2.3

Lower and Upper Approximations

The lower and the upper approximations of subsets of U are deﬁned as follows. Deﬁnition 2. For any approximation space AS#,$ = (U, I# , ν$ ) and any subset X ⊆ U , the lower and upper approximations are deﬁned by LOW AS#,$ , X = {x ∈ U : ν$ (I# (x) , X) = 1} , U P P AS#,$ , X = {x ∈ U : ν$ (I# (x) , X) > 0}, respectively. The lower approximation of a set X with respect to the approximation space AS#,$ is the set of all objects, which can be classiﬁed with certainty as objects of X with respect to AS#,$ . The upper approximation of a set X with respect to the approximation space AS#,$ is the set of all objects which can possibly be classiﬁed as objects of X with respect to AS#,$ . Several known approaches to concept approximation can be covered using the approximation spaces discussed here, e.g., the approach given in [8] or tolerance (similarity) rough set approximations (see, e.g., references in [10]). We recall the notions of the positive region and the quality of approximation of classiﬁcation in the case of generalized approximation spaces [13]. Deﬁnition 3. Let AS#,$ = (U, I# , ν$ ) be an approximation space and let r > 1 be a given naturalnumber and let {X1 , . . . , Xr } be a classiﬁcation of objects (i.e. X1 , . . . , Xr ⊆ U , ri=1 Xi = U and Xi ∩ Xj = ∅ for i = j, where i, j = 1, . . . , r). 1. The positive region of the classiﬁcation {X1 , . . . , Xr } with respect to the approximation space AS#,$ is deﬁned by P OS AS#,$ , {X1 , . . . , Xr } = ri=1 LOW AS#,$ , Xi . 2. The quality of approximation of the classiﬁcation {X1 , . . . , Xr } in the approximation space AS#,$ is deﬁned by card(P OS (AS#,$ ,{X1 ,...,Xr })) γ AS#,$ , {X1 , . . . , Xr } = . card(U) The quality of approximation of the classiﬁcation coeﬃcient expresses the ratio of the number of all AS#,$ -correctly classiﬁed objects to the number of all objects in the data table.

3

Relational Data Mining

Knowledge discovery is the process of discovering particular patterns over data. In this context data is typically stored in a database. Approaches using ﬁrst order logic (FOL, for short) languages for the description of such patterns oﬀer data mining the opportunity of discovering more complex regularities which may be out of reach for attribute-value languages.

Approximation Spaces in Multi Relational Knowledge Discovery

3.1

355

Didactic Example

In this section we present an example inspired by [2]. Example 1. There are two information systems: ISCustomer = (UCustomer , ACustomer ) where the set of objects UCustomer = {x1 , . . . , x7 }, and the set of attributes ACustomer = {N ame, Gender, Income, BigSpender} (see Table 1) and ISMarriedT o = (UMarriedT o , AMarriedT o ) where UMarriedT o = {y1 , y2 , y3 }, and AMarriedT o = {Spouse1, Spouse2} (see Table 2). Table 1. An Information System ISCustomer UCustomer x1 x2 x3 x4 x5 x6 x7

Name Mary Eve Kate Meg Jim Tom Henry

Gender Female Female Female Female Male Male Male

Income BigSpender 70000 yes 120000 yes 80000 no 80000 yes 100000 yes 100000 yes 60000 no

Table 2. An Information System ISM arriedT o UM arriedT o Spouse1 Spouse2 y1 M ary Jim y2 M eg T om y3 Kate Henry

Using attribute–value language we obtain for example the following decision rules: if Income ≥ 100000 then BigSpender = yes if Income ≤ 75000 then BigSpender = yes (May be this rule is not intuitive.) if N ame = M eg then BigSpender = yes (This rule is generally not applicable to new objects.) Using ﬁrst order language one can obtain the following two rules: BigSpender(var1 , var2 , var3 ) ← var3 ≥ 100000 BigSpender(var1 , var3 , var3 ) ← M arriedT o(var1 , var1 ) and Customer(var1 , var2 , var3 , var4 ) and var3 ≥ 100000 which involve the predicates Customer and M arriedT o. It predicts a person to be a big spender if the person is married to somebody with high income (compare this to the rule that states a person is a big spender if he/she has high

356

J. Stepaniuk

income, listed above the relational rules). Note that the two persons var1 and var1 are connected through the relation MarriedTo. Relational patterns are typically expressed in subsets of ﬁrst-order logic (also called predicate or relational logic). Essentials of predicate logic include predicates (M arriedT o) and variables (var1 , var1 ), which are not present in propositional logic (attribute–value language). Relational patterns are thus more expressive than the propositional ones. Knowledge discovery based on FOL has other advantages as well. Complex background knowledge provided by experts can be encoded as ﬁrst order formulas and be used in the discovery task. The expressiveness of FOL enables the discovered patterns to be described in a concise way, which in most cases increases readability of the output. Multiple relations can be naturally handled without explicit (and expensive) joins. 3.2

Relational Learning

Before moving on to the algorithm for learning of a set of rules, let us introduce some basic terminology from relational learning. Relational learning algorithms learn classiﬁcation rules for a concept [2] (for relational methods and their applications in computer science see also [5]). The program typically receives a large collection of positive and negative examples from real-world databases as well as background knowledge in the form of relations. Let p be a target predicate of arity m and r1 , . . . , rl be background predicates, where m, l > 0 are given natural numbers. We denote the constants by con1 , . . . , conn , where n > 0. A term is either a variable or a constant. An atomic formula is of the form p (t1 , . . . , tm ) or ri (t1 , . . .) where the t s are terms and i = 1, . . . , l. A literal is an atomic formula or its negation. If a literal contains a negation symbol (¬), we call it a negative literal, otherwise it is a positive literal. A clause is any disjunction of literals, where all variables are assumed to be universally quantiﬁed. The learning task for relational learning systems is as follows: Input + − of positive and a set Xtarget of negative training examples (exa set Xtarget pressed by literals without variables) for the target relation, background knowledge (or BK for short) expressed by literals without variables and not including the target predicate. Output p ) a set of ξ ← λ rules, where ξ is an atomic formula of the form p (var1p , . . . , varm with the target predicate p and λ is a conjunction of literals over background predicates r1 , . . . , rl , such that the set of rules satisﬁes the positive examples relatively to background knowledge. Example 2. Let us consider the data set related to document understanding. The learning task involves identifying the purposes served by components of single-page letters such as that in Figure 1.

Approximation Spaces in Multi Relational Knowledge Discovery

357

Background predicates describe properties of components such as their width and height, and relationships such as horizontal and vertical alignment with other components. Target predicates describe whether a block is one of the ﬁve predetermined types: sender, receiver, logo, reference, and date. For example, for letter presented in Figure 1, we obtain the following predicate data: date (c8 ), logo (c3 ), receiver (c21 ), on top (c8 , c21 ), on top (c21 , c14 ), on top (c5 , c24 ), on top (c3 , c5 ), aligned only lef t col (c1 , c3 ), aligned only right col (c5 , c21 ), . . . We consider generation of rules of the form: sender (var1 ) ← on top (var1 , var2 ) and logo (var2 ). We will adopt the lower and the upper approximations for subsets of the set of target examples. First, we deﬁne the coverage of a rule. Deﬁnition 4. The coverage of Rule, written Coverage(Rule), is the set of examples such that there exists a substitution giving values to all variables appearing in the rule and all literals of the rule are satisﬁed for this substitution. The set of the positive (negative) examples covered by Rule is denoted by Coverage+ (Rule), Coverage− (Rule), respectively. Remark 1. For any literal L, we obtain Coverage(h ← b) = Coverage(h ← b ∧ L) ∪ Coverage(h ← b ∧ ¬L). + − ∪ Xtarget and Rule Set = {Rule1, . . . , Rulen}. Let U = Xtarget

Deﬁnition 5. For the set of rules Rule Set and any example x ∈ U the uncertainty function is deﬁned by IRule

Set (x)

= {x} ∪

n

{Coverage(Rulei ) : x ∈ Coverage(Rulei )} .

i=1

The lower and upper approximations may be deﬁned as earlier but in this case they are equal to the forms presented in Remark 2. Remark 2. For an approximation space ASRule Set = (U, IRule Set , νSRI ) and any subset X ⊆ U the lower and the upper approximations are deﬁned by LOW (ASRule Set , X) = {x ∈ U : IRule U P P (ASRule respectively.

Set , X)

= {x ∈ U : IRule

Set (x)

Set (x)

⊆ X} ,

∩ X = ∅} ,

358

J. Stepaniuk

c1 (sender)

c8 (date)

c21 (receiver)

c3 (logo) c14 (reference)

c5

c7

c24

Fig. 1. Sample Letter Showing Components

4

Translating First–Order Data into Attribute–Value Form

In this section we discuss the approach based on two steps. First, the data is transformed from ﬁrst-order logic into decision table format by the iterative checking whether a new attribute adds any relevant information to the decision table. Next, the reducts and rules from reducts [8,10,13] are computed from the decision table obtained. Data represented as a set of formulas can be transformed into attribute– value form. The idea of translation was inspired by LINUS and DINUS systems

Approximation Spaces in Multi Relational Knowledge Discovery

359

(see, e.g., [2]). We start with a decision table directly derived from the positive and negative examples of the target relation. Assuming that we have mary target predicate, the set U of objects in the decision table is a subset of {con1 , . . . , conn }m . Decision attribute dp : U → {+, −} is deﬁned by the target predicate with possible values ” + ” or ” − ”. All positive and negative examples of the target predicate are now put into the decision table. Each example forms a separate row in the table. Then background knowledge is applied to the decision table. We determine all the possible applications of the background predicates to the arguments of the target relation. Each such application introduces a new Boolean attribute. To analyze the complexity of the obtained data table, let us consider the number of condition attributes. Let Ari be a set of attributes constructed for every predicate symbol ri , where i = 1, . . . , l. The number of condition attributes l in constructed data table is equal to i=1 card (Ari ) resulting from the possible applications of the l background predicates on the variables of the target relation. The cardinality of Ari depends on the number of arguments of target predicate p (denoted by m) and the arity of ri . Namely, card (Ari ) is equal to mar(ri ) , where ar (ri ) is the arity of the predicate ri . The number of condition attributes in obtained data table is polynomial in the arity m of the target predicate p and the number l of background knowledge predicates, but its size is usually so large that its processing is unfeasible. Therefore, one can check interactively if a new attribute is relevant, i.e., if it adds any information to the decision table and, next we add to the decision table only relevant attributes. Two conditions for testing if a new attribute a is relevant are proposed: 1. γ ASB∪{a} , {X+ , X− } > γ (ASB , {X+ , X− }) , where X+ and X− denote the decision classes corresponding to the target concept. An attribute a is added to the decision table if this results in a growth of the positive region with respect to the attributes selected previously. 2. QDIS (a) = νSRI (X+ × X− , {(x, y) ∈ X+ × X− : a (x) = a (y)}) ≥ θ, where θ ∈ [0, 1] is a given real number. An attribute a is added to the decision table if it introduces some discernibility between objects belonging to diﬀerent non-empty classes X+ and X− . Each of these conditions can be applied to a single attribute before it is introduced to the decision table. If this attribute does not meet a condition, it should not be included into the decision table. The received data table is then analyzed by a rough set based systems. First, reducts are computed. Next, decision rules are generated. Example 3. The problem with three binary predicates r1 , r3 , p and one unary predicate r2 can be used to demonstrate the transformation of relational learning problem into attribute–value form. Suppose that there are the following positive and negative examples of a target predicate p : + − = {p(1, 2), p(4, 1), p(4, 2)}, Xtarget = {¬p(6, 2), ¬p(3, 5), ¬p(1, 4)}. Xtarget

360

J. Stepaniuk

Consider the background knowledge about relations, r1 , r2 , and r3 : r1 (5, 1), r1 (1, 2), r1 (1, 4), r1 (4, 1), r1 (3, 1), r1 (2, 6), r1 (3, 5), r1 (4, 2), r2 (1), r2 (2), r2 (3), r2 (4), r2 (6), r3 (2, 1), r3 (1, 4), r3 (2, 4), r3 (2, 5), r3 (3, 2), r3 (3, 5), r3 (5, 1), r3 (5, 3), r3 (2, 6), r3 (4, 2). We then transform the data into attribute–value form (decision table). In Table 3, a quality index QDIS of potential attributes is presented. Table 3. Quality QDIS of Potential Attributes Symbol Attribute QDIS (•) a1 r2 (var1 ) 0 a2 r2 (var2 ) 0.33 a3 r1 (var1 , var1 ) 0 a4 r1 (var1 , var2 ) 0.33 a5 r1 (var2 , var1 ) 0.56 a6 r1 (var2 , var2 ) 0 a7 r3 (var1 , var1 ) 0 a8 r3 (var1 , var2 ) 0.56 a9 r3 (var2 , var1 ) 0.33 a10 r3 (var2 , var2 ) 0

Using conditions introduced in this section some attributes will not be included in the resulting decision table. For example, the second condition with QDIS (•) ≥ θ = 0.3 would permit the following attribute set into the decision table: A0.3 = {a2 , a4 , a5 , a8 , a9 }. Therefore, DT0.3 = (U, A0.3 ∪ {d}) ﬁnally. We obtain two decision classes: X+ = {(1, 2) , (4, 1) , (4, 2)} and X− = {(6, 2) , (3, 5) , (1, 4)} . For the obtained decision table we construct an approximation space ASA0.3 = (U, IA0.3 , νSRI ) such that the uncertainty function and the rough inclusion are deﬁned in Table 4. Then, we can compute reducts and decision rules.

5

The Rough Set Relational Learning Algorithm

In this section we introduce and investigate the RSRL (Rough Set Relational Learning) algorithm. Some preliminary versions of this algorithm were presented in [14,15]. 5.1

RSRL Algorithm

To select the most promising literal from the candidates generated at each step, RSRL considers the performance of the rule over the training data. The evaluation function card(R(L, N ewRule)) used by RSRL to estimate the utility of adding a new literal is based on the numbers of discernible positive and negative examples before and after adding the new literal (see, Figure 2).

Approximation Spaces in Multi Relational Knowledge Discovery

361

Table 4. Resulting Decision Table DT0.3 , Uncertainty Function and Rough Inclusion (var1 , var2 ) (1, 2) (4, 1) (4, 2) (6, 2) (3, 5) (1, 4)

a2 true true true true false true

a4 true true true false true true

a5 false true false true false true

a8 false false true false true true

a9 true true true true true false

dp + + + -

IA0.3 (•) νSRI (•, X+ ) νSRI (•, X− ) {(1, 2)} 1 0 {(4, 1)} 1 0 {(4, 2)} 1 0 {(6, 2)} 0 1 {(3, 5)} 0 1 {(1, 4)} 0 1

Some modiﬁcation of the algorithm RSRL were presented in [15]. The modiﬁed algorithm generates rules as the original RSRL but its complexity is lower because it performs operations on the cardinalities of sets without computing the sets. 5.2

Illustrative Example

Let us illustrate the RSRL algorithm on a simple problem of learning a relation. Example 4. The task is to deﬁne the target relation p(var1 , var2 ) in terms of the background knowledge relations r1 and r3 . Let BK = {r1 (1, 2), r1 (1, 3), r1 (2, 4), r3 (5, 2), r3 (5, 3), r3 (4, 6), r3 (4, 7)}. There are two positive and three negative examples of the target relation: + − = {e1 , e2 } and Xtarget = {e3 , e4 , e5 }, where Xtarget

e1 = p(1, 4), e2 = p(2, 6), e3 = ¬p(5, 4), e4 = ¬p(5, 3) and e5 = ¬p(1, 2). Let us see how the algorithm generates rules for h = p(var1 , var2 ), app = lower. The successive steps of the algorithm: P os = {e1 , e2 }, Learned rules = ∅. P os = ∅. R = {(e1 , e3 ), (e1 , e4 ), (e1 , e5 ), (e2 , e3 ), (e2 , e4 ), (e2 , e5 )}. R = ∅. We obtain the following candidates: ri (var1 , var1 ), ri (var1 , var2 ), ri (var1 , var3 ), ri (var2 , var1 ), ri (var2 , var2 ), ri (var2 , var3 ), ri (var3 , var1 ), ri (var3 , var2 ), where i = 1, 3. For ever1. candidate, we compute R(L, N ewRule) and we obtain the best result for r3 (var1 , var3 ). In the ﬁrst step, every example is covered either by the rule p(var1 , var2 ) ← r3 (var1 , var3 ) or by p(var1 , var2 ) ← ¬r3 (var1 , var3 ). We obtain: e1 , e2 , e5 ∈ Coverage+ (h ← ¬Best literal) ∪ Coverage− (h ← ¬Best literal), e3 , e4 ∈ Coverage+ (h ← Best literal) ∪ Coverage− (h ← Best literal). From the intersection of R and the set (Coverage+ (h ← Best literal) × Coverage− (h ← ¬Best literal))∪

362

J. Stepaniuk

Coverage − (h ← b ∧ L )

Coverage − (h ← b ∧ ¬L )

Coverage + (h ← b ∧ ¬L )

Coverage + (h ← b ∧ L )

Fig. 2. The set R(L, h ← b) is equal to the union of two Cartesian products

(Coverage+ (h ← ¬Best literal) × Coverage− (h ← Best literal)), we obtain R(Best literal, N ewRule) = {(e1 , e3 ), (e1 , e4 ), (e2 , e3 ), (e2 , e4 )} = ∅. Since the value of the coverage of p(var1 , var2 ) ← ¬r3 (var1 , var3 ) is greater than the value of the coverage of p(var1 , var2 ) ← r3 (var1 , var3 ), Best literal = ¬r3 (var1 , var3 ). app = upper. b = ¬r3 (var1 , var3 ). Coverage− (N ewRule) = ∅.

Approximation Spaces in Multi Relational Knowledge Discovery

363

We ﬁnd new R considering a general case: b = b1 = ¬r3 (var1 , var3 ), L1 = ¬r3 (var1 , var3 ). R1 = Coverage+ (h ← b1 ) × Coverage− (h ← b1 ) ∪ S1 . Coverage+ (h ← b1 ) = {e1 , e2 }, Coverage− (h ← b1 ) = {e5 }, Coverage+ (h ← b0 ∧ ¬L1 ) = ∅, Coverage− (h ← b0 ∧ ¬L1 ) = {e3 , e4 }, S1 = Coverage+ (h ← b0 ∧ ¬L1 ) × Coverage− (h ← b0 ∧ ¬L1 ) = ∅. Hence, we obtain R = R1 = {(e1 , e5 ), (e2 , e5 )}. The second step in the second loop: R = ∅. Algorithm 1. RSRL Algorithm + − input : T arget predicate, BK, Xtarget ∪ Xtarget , app //where T arget predicate + is a target predicate with a set Xtarget of positive examples and a set − Xtarget of negative examples, BK is a background knowledge, app is a type of approximation (app ∈ {lower, upper}). output: Learned rules //where Learned rules is a set of rules for ”positive decision class”. + ; P os ←− Xtarget Learned rules ←− ∅; while P os = ∅ do Learn a N ewRule; N ewRule ←− most general rule possible; − ; R ←− P os × Xtarget while R = ∅ do Candidate literals ←− generated candidates; // RSRL generates candidate specializations of N ewRule by considering a new literal L that ﬁts one of the following forms: – r(var1 , . . . , vars ), where at least one of the variable vari in the created literal must already exist in the positive literals of the rule; – the negation of the above form of literal;

Best literal ←− arg max L∈Candidate literals card(R(L, N ewRule)); // (the explanation of R(L, Rule) is in Figure 2 given) if R(Best literal, N ewRule) = ∅ or (app = upper and (N ewRule = most general rule possible) and Coverage+ (N ewRule) = Coverage+ (N ewRule ∧ Best literal)) then exit while; end Add Best literal to N ewRule preconditions; //Add a new literal to specialize N ewRule; if Coverage−(N ewRule) = ∅ then exit while; end R := R \ R(Best literal, N ewRule); end Learned rules ←− Learned rules ∪ {N ewRule} ; P os ←− P os \ Coverage+ (N ewRule); end

364

J. Stepaniuk

We generate new candidates. We obtain the best result for the candidate r1 (var1 , var2 ) thus Best literal = r1 (var1 , var2 ). Now b = b2 = ¬r3 (var1 , var3 ) ∧ r1 (var1 , var2 ), L2 = r1 (var1 , var2 ). We compute the following sets: Coverage+ (h ← b2 ) = ∅, Coverage− (h ← b2 ) = {e5 }, Coverage+ (h ← b1 ∧ ¬L2 ) = {e1 , e2 }, Coverage− (h ← b1 ∧ ¬L2 ) = ∅. Coverage+ (h ← b2 ) × Coverage− (h ← b1 ∧ ¬L2 )∪ Coverage+ (h ← b1 ∧ ¬L2 ) × Coverage− (h ← b2 ) = {(e1 , e5 ), (e2 , e5 )}. We obtain R(Best literal, N ewRule) = {(e1 , e5 ), (e2 , e5 )} = ∅. Since the value of the coverage of p(var1 , var2 ) ← ¬r3 (var1 , var3 ) ∧ ¬r1 (var1 , var2 ) is greater than the value of the coverage of p(var1 , var2 ) ← ¬r3 (var1 , var3 ) ∧ r1 (var1 , var2 ) then Best literal = ¬r1 (var1 , var2 ). app = upper. b = ¬r3 (var1 , var3 ) ∧ ¬r1 (var1 , var2 ). Coverage− (N ewRule) = ∅. The end of the second loop. Learned rules = {p(var1 , var2 ) ← ¬r3 (var1 , var3 ) ∧ ¬r1 (var1 , var2 )}. Coverage+ (p(var1 , var2 ) ← ¬r3 (var1 , var3 ) ∧ ¬r1 (var1 , var2 )) = {e1 , e2 } = P os, hence P os = P os \ Coverage+ (N ewRule) = ∅. The end of the algorithm proceeding. In each step of the algorithm we obtain Coverage+ (N ewRule) = Coverage+ (N ewRule ∧ Best literal). Hence, if app = upper then we obtain the same rules as for app = lower. Hence, the lower and the upper approximations + of Xtarget are equal in our example. Let us compute the above sets to compare them. We have Rule Set = {p(var1 , var2 ) ← ¬r3 (var1 , var3 ) ∧ r1 (var1 , var2 ), + p(var1 , var2 ) ← ¬r3 (var1 , var3 ) ∧ ¬r1 (var1 , var2 )} and Xtarget = {e1 , e2 }. We obtain the uncertainty function IRule Set (e1 ) = IRule Set (e2 ) = {e1 , e2 }, IRule Set (e3 ) = {e3 }, IRule Set (e4 ) = {e4 } and IRule Set(e5 ) = {e5 }. + + Hence, LOW ASRule Set , Xtarget = {e1 , e2 } = U P P ASRule Set , Xtarget .

6

Conclusions

The ﬁrst approach presented in this paper transforms input ﬁrst-order logic formulas into decision table form, then uses reducts to select only meaningful data. The second approach is based on the algorithm RSRL for the ﬁrst order rules generation. We showed that approximation spaces are basic structures for knowledge discovery from multi-relational data. Furthermore, our approach can be treated as a step towards the understanding of rough set methods in the ﬁrst order rules generation.

Acknowledgements The author wishes to thank the anonymous reviewers for their many helpful comments.

Approximation Spaces in Multi Relational Knowledge Discovery

365

References 1. Bonchi, F., Boulicaut, J. F. (Eds.): Knowledge Discovery in Inductive Databases, Lecture Notes in Computer Science 3933, Springer–Verlag, Berlin Heidelberg, 2006. 2. Dzeroski, S., Lavrac, N. (Eds.): Relational Data Mining, Springer-Verlag, Berlin, 2001. 3. L ukasiewicz, J.: Die logischen Grundlagen der Wahrscheinlichkeitsrechnung, Krak´ ow 1913. In: Borkowski, L. (ed.), Jan L ukasiewicz - Selected Works. North Holland, Amstardam, Polish Scientiﬁc Publishers, Warsaw, 1970. 4. Milton, R. S., Maheswari V. U., Siromoney A.: Rough Sets and Relational Learning, Transactions on Rough Sets I, Lecture Notes in Computer Science 3100, Springer, 2004, 321–337. 5. Orlowska, E., Szalas, A. (Eds.): Relational Methods for Computer Science Applications, Physica–Verlag, Heidelberg, 2001. 6. Pal, S.K., Polkowski, L., Skowron, A. (Eds.): Rough-Neural Computing: Techniques for Computing with Words. Springer-Verlag, Berlin, 2004. 7. Pawlak, Z.: Rough sets, International J. Comp. Inform. Science 11, 1982, 341–356. 8. Pawlak, Z.: Rough Sets. Theoretical Aspects of Reasoning about Data, Kluwer Academic Publishers, Dordrecht, 1991. 9. Polkowski, L., Skowron, A. (Eds.): Rough Sets in Knowledge Discovery 1 and 2. Physica-Verlag, Heidelberg, 1998. 10. Skowron, A., Stepaniuk, J.: Tolerance Approximation Spaces, Fundamenta Informaticae, 27, 1996, 245–253. 11. Skowron A., Stepaniuk J., Peters J. F., Swiniarski R.: Calculi of Approximation Spaces, Fundamenta Informaticae vol. 72(1–3), 2006, 363–378. 12. Stepaniuk, J.: Rough relations and logics. In: L. Polkowski, A. Skowron (Eds.), Rough Sets in Knowledge Discovery 1. Methodology and Applications, Physica Verlag, Heidelberg, 1998, 248–260. 13. Stepaniuk, J.: Knowledge Discovery by Application of Rough Set Models, L. Polkowski, S. Tsumoto, T.Y. Lin, (Eds.) Rough Set Methods and Applications. New Developments in Knowledge Discovery in Information Systems, Physica– Verlag, Heidelberg, 2000, 137–233. 14. Stepaniuk, J., G´ oralczuk, L.: An Algorithm Generating First Order Rules Based on Rough Set Methods, (ed.) J. Stepaniuk, Zeszyty Naukowe Politechniki Bialostockiej Informatyka nr 1, 2002, 235–250. [in Polish] 15. Stepaniuk, J., Honko, P.: Learning First–Order Rules: A Rough Set Approach Fundamenta Informaticae, 61(2), 2004, 139–157. 16. Ziarko, W., Variable precision rough set model, Journal of Computer and System Sciences 46, 1993, 39–59.

Finding Relevant Attributes in High Dimensional Data: A Distributed Computing Hybrid Data Mining Strategy Julio J. Vald´es and Alan J. Barton National Research Council Canada, M50, 1200 Montreal Rd., Ottawa, ON K1A 0R6 julio.valdes@nrc-cnrc.gc.ca, alan.barton@nrc-cnrc.gc.ca http://iit-iti.nrc-cnrc.gc.ca

Abstract. In many domains the data objects are described in terms of a large number of features (e.g. microarray experiments, or spectral characterizations of organic and inorganic samples). A pipelined approach using two clustering algorithms in combination with Rough Sets is investigated for the purpose of discovering important combinations of attributes in high dimensional data. The Leader and several k-means algorithms are used as fast procedures for attribute set simplification of the information systems presented to the rough sets algorithms. The data described in terms of these fewer features are then discretized with respect to the decision attribute according to different rough set based schemes. From them, the reducts and their derived rules are extracted, which are applied to test data in order to evaluate the resulting classification accuracy in crossvalidation experiments. The data mining process is implemented within a high throughput distributed computing environment. Nonlinear transformation of attribute subsets preserving the similarity structure of the data were also investigated. Their classification ability, and that of subsets of attributes obtained after the mining process were described in terms of analytic functions obtained by genetic programming (gene expression programming), and simplified using computer algebra systems. Visual data mining techniques using virtual reality were used for inspecting results. An exploration of this approach (using Leukemia, Colon cancer and Breast cancer gene expression data) was conducted in a series of experiments. They led to small subsets of genes with high discrimination power.

1 Introduction As a consequence of the information explosion and the development of sensor and observation technologies, it is now common in many domains to have data objects characterized by an increasingly larger number of attributes, leading to high dimensional databases in terms of the set of fields. A typical example is a gene expression experiment, where the genetic content of samples of tissues are obtained with high throughput technologies (microchips). Usually, thousands of genes are investigated in such experiments. In other bio-medical research contexts, the samples are characterized by infrared, ultraviolet, and other kinds of spectra, where the absorption properties, with respect to a large number of wavelengths, are investigated. The same situation occurs in other domains, and the common denominator is to have a set of data objects of a very high dimensional nature. J.F. Peters et al. (Eds.): Transactions on Rough Sets VI, LNCS 4374, pp. 366–396, 2007. c Springer-Verlag Berlin Heidelberg 2007

Finding Relevant Attributes in High Dimensional Data

367

This paper investigates one, of the possibly many approaches to the problem of finding relevant attributes in high dimensional datasets. The approach is based on a combination of clustering and rough sets techniques in a high throughput distributed computing environment, with low dimensional virtual reality data representations aiding data analysis understanding. The goals are: i) to investigate the behavior of the combination of these techniques in a knowledge discovery process ii) to perform preliminary comparisons of the experimental results from the point of view of the discovered relevant attributes, applied to the example problem of finding relevant genes

2 Datasets In this study publicly available datasets were considered. They result from gene expression experiments in genomics, and appear in numerous studies about data mining and machine learning in bioinformatics. All of them share a feature typical of that kind of information: the data consist of a relatively small number of samples, described in terms of a large collection of attributes. Besides genomics, this situation is found in other fields as well, like experimental physics and astronomy. When infrared, ultraviolet or other spectral properties are used to describe the sampled objects, hundreds or thousands of energy intensity values for radiation emission or absorption at different wavelengths are used as sample attributes. The techniques investigated here are of a general nature, that is, not specific or tailored to any particular domain. The datasets considered for this study were: – Leukemia ALL/AML dataset: (72 samples described in terms of 7129 genes [15]). – Breast Cancer (24 samples described in terms of 12, 625 genes [7]). – Colon Cancer: (62 samples described in terms of 2000 genes [1]). The Leukemia dataset is that of [15], and consists of 7129 genes, where patients are separated into i) a training set containing 38 bone marrow samples: 27 acute lymphoblastic leukemia (ALL) and 11 acute myeloid leukemia (AML), obtained from patients at the time of diagnosis, and ii) a testing set containing 34 samples (24 bone marrow and 10 peripheral blood samples), where 20 are ALL and 14 AML. The test set contains a much broader range of biological samples, including those from peripheral blood rather than bone marrow, from childhood AML patients, and from different reference laboratories that used different sample preparation protocols. In the present study, however, the dataset will not be divided into training and test samples, because crossvalidation is used, as explained below. The breast cancer data selected [7] was that provided by the Gene Expression Omnibus (GEO) (See www.ncbi.nlm.nih.gov/projects/geo/gds/ gds browse.cgi?gds=360). It consists of 24 core biopsies taken from patients found to be resistant (greater than 25% residual tumor volume, of which there are 14) or sensitive (less than 25% residual tumor volume, of which there are 10) to

368

J.J. Vald´es and A.J. Barton

docetaxel treatment. The number of genes (probes) placed onto (and measured from) the microarray is 12, 625, and two classes are recognized: resistant and sensitive. The Colon cancer data correspond to tumor and normal colon tissues probed by oligonucleotide arrays [1].

3 Foundational Concepts 3.1 Clustering Methods Clustering with classical partition methods constructs crisp (non overlapping) subpopulations of objects or attributes. Two such classical algorithms were used in this study: the Leader algorithm [17], and several variants of k-means [2]. Leader Algorithm. The leader algorithm operates with a dissimilarity or similarity measure and a preset threshold. A single pass is made through the data objects, assigning each object to the first cluster whose leader (i.e. representative) is close enough (or similar enough) to the current object w.r.t. the specified measure and threshold. If no such matching leader is found, then the algorithm will set the current object to be a new leader; forming a new cluster. This technique is very fast; however, it has several negative properties: i) the first data object always defines a cluster and therefore, appears as a leader. ii) the partition formed is not invariant under a permutation of the data objects. iii) the algorithm is biased, as the first clusters tend to be larger than the later ones since they get first chance at “absorbing” each object as it is allocated. Variants of this algorithm with the purpose of reducing bias include: a) reversing the order of presentation of a data object to the list of currently formed leaders. b) selecting the absolute best leader found (thus making the object presentation order irrelevant). The highest quality is obtained using b), but at a higher computational cost because the set of leaders (whose cardinality increases as the process progresses), has to be completely explored for every data object. Nevertheless, even with this extra computational overhead, the technique is still very fast, and large datasets can be clustered very quickly. Usually the partitions generated by this method are used as initial approximations to more elaborated methods. K-Means. The k-means algorithm is actually a family of techniques based on the concept of data reallocation. A dissimilarity or similarity measure is supplied, together with an initial partition of the data, and the goal is to alter cluster membership so as to obtain a better partition w.r.t. the chosen measure. The modification of membership is performed by reallocating the data objects to a different group w.r.t. the one in which it was a member. Different variants very often give different partition results. However,

Finding Relevant Attributes in High Dimensional Data

369

in papers dealing with gene expression analysis, very seldom are the specificities of the k-means algorithm described. For the purposes of this study, the following k-means variants were used: Forgy’s, Jancey’s, convergent, and MacQueen’s [13], [20], [24], [2]. Let nc be the number of clusters desired. The definition of an initial partition follows basically two schemes: i) direct specification of a set of nc initial centroids (seed points), or ii) specification of nc initial disjoint groups such that they cover the entire dataset, and compute from them initial centroids to start the process. There are many variations of these two schemes, and the following variants for defining an initial partition were considered in this paper: 1. Select nc data objects and use them as initial centroids. 2. Divide the total number of objects into nc consecutive clusters, compute the centroid of each, and use them as initial centroids. 3. Arbitrary nc centroids are given externally. 4. Take the first nc data objects and use them as initial centroids. The classical k-means clustering is a simple algorithm with the following sequence of steps. 1. Allocate each data unit to the cluster with the nearest seed point (if a dissimilarity measure is used), or to the cluster with the most similar seed point (if a similarity measure is used). 2. Compute new seed points as the centroids of the newly formed clusters 3. If (termination criteria = true) then stop else goto 2 Several termination criteria (or a combination of them) can be established, which provide better control on the conditions under which a k-means process concludes. Among them are the folowing: 1. 2. 3. 4.

A preset number of object reallocations have been performed. A preset number of iterations has been reached. A partition quality measure has been reached. The partition quality measure does not change in subsequent steps.

There are several variants of the general k-means scheme. That is why it is necessary to specify explicitly the specific variant applied. In this paper, several of them were used. K-Means: Forgy’s Variant. The classical Forgy’s k-means algorithm [13] consists of the following steps: i) Begin with any desired initial configuration. Go to (ii) if beginning with a set of seed objects, or go to (iii) if beginning with a partition of the dataset. ii) Allocate each object to the cluster with the nearest (most similar) seed object (centroid). The seed objects remain fixed for a full cycle through the entire dataset. iii) Compute new centroids of the clusters. iv) Alternate (ii) and (iii) until the process converges (that is, until no objects change their cluster membership).

370

J.J. Vald´es and A.J. Barton

K-Means: Jancey’s Variant. In Jancey’s variant [20], the process is similar to Forgy’s, but the first set of cluster seed objects is either given, or computed as the centroids of clusters in the initial partition. Then, at all succeeding stages, each new seed point is found by reflecting the old one through the new centroid for the cluster (a heuristic which tries to approximate the direction of the gradient of the error function). K-Means: MacQueen’s Variant. MacQueen’s method [24] is another popular member of the k-means family, and is composed of the following steps: i) Take the first k data units as clusters of one member each. ii) Assign each of the remaining objects to the cluster with the nearest (most similar) centroid. After each assignment, recompute the centroid of the gaining cluster. iii) After all objects have been assigned in step ii), take the existing cluster centroids as fixed points and make one more pass through the dataset assigning each object to the nearest (most similar) seed object. K-Means: Convergent Variant. The so called convergent k-means [2] is a variant defined by the following steps: i)

Begin with an initial partition like in Forgy’s and Jancey’s methods (or the output of MacQueen’s method). ii) Take each object in sequence and compute the distances (similarities) to all cluster centroids; if the nearest (most similar) is not that of the object’s parent cluster, reassign the object and update the centroids of the losing and gaining clusters. iii) Repeat steps ii) and iii) until convergence is achieved (that is, until there is no change in cluster membership). Similarity Measure. The Leader and the k-means algorithms were used with a similarity measure rather than with a distance. In particular Gower’s general coefficient was used [16], where the similarity between objects i and j is given by Eq-1: Sij =

p k=1

sijk /

p

wijk ,

(1)

k=1

where the weight of the attribute (wijk ) is set equal to 0 or 1 depending on whether the comparison is considered valid for attribute k. If vk (i), vk (j) are the values of attribute k for objects i and j respectively, an invalid comparison occurs when at least one them is missing. In this situation wijk is set to 0. For quantitative attributes (like the ones in the datasets used in this paper), the scores sijk are assigned as in Eq-2: sijk = 1 − |Xik − Xjk |/Rk ,

(2)

where Xik is the value of attribute k for object i (similarly for object j), and Rk is the range of attribute k. For symbolic attributes (nominal), the scores sijk are assigned as in Eq-3 1 if Xik = Xjk sijk = (3) 0 otherwise .

Finding Relevant Attributes in High Dimensional Data

371

3.2 Rough Sets The Rough Set Theory [31] bears on the assumption that in order to define a set, some knowledge about the elements of the dataset is needed. This is in contrast to the classical approach where a set is uniquely defined by its elements. In the Rough Set Theory, some elements may be indiscernible from the point of view of the available information and it turns out that vagueness and uncertainty are strongly related to indiscernibility. Within this theory, knowledge is understood to be the ability of characterizing all classes of the classification. More specifically, an information system is a pair A = (U, A) where U is a non-empty finite set called the universe and A is a non-empty finite set of attributes such that a : U → Va for every a ∈ A . The set Va is called the value set of a. For example, a decision table is any information system of the form A = (U, A ∪ {d}), where d ∈ A is the decision attribute and the elements of A are the condition attributes. Implicants. It has been described [28] that an m-variable function f : Bm → B is called a Boolean function if and only if it can be expressed by a Boolean formula. An implicant of a Boolean function f is a term p such that p f , where is a partial order called the inclusion relation. A prime implicant is an implicant of f that ceases to be so if any of its literals are removed. An implicant p of f is a prime implicant of f in case, for any term q, the implication of Eq-4 holds. pqf ⇒p=q .

(4)

General Boolean Reasoning Solution Scheme. It has been described [28] that following the presentation of earlier work, the general scheme of applying Boolean reasoning to solve a problem P can be formulated as follows: 1. Encode problem P as a system of simultaneously-asserted Boolean equations as in Eq-5, where the gi and hi are Boolean functions on B. ⎧ ⎪ ⎨ g 1 = h1 .. P = . (5) . ⎪ ⎩ g k = hk 2. Reduce the system to a single Boolean equation (e.g. fp = 0) as in Eq-6. fp =

k g i · hi + g i · hi .

(6)

i=1

3. Compute Blake’s Canonical Form (BCF (fp )), the prime implicants of fp . 4. Solutions to P are then obtained by interpreting the prime implicants of fp . Discernibility Matrices. An information system A defines a matrix MA called a discernibility matrix. Each entry MA (x, y) ⊆ A consists of the set of attributes that can be used to discern between objects x, y ∈ U according to Eq-7. MA (x, y) = {a ∈ A : discerns (a, x, y)} . Where, discerns (a, x, y) may be tailored to the application at hand.

(7)

372

J.J. Vald´es and A.J. Barton

Indiscernibility Relations and Graphs. A discernibility matrix MA defines a binary relation RA ⊆ U 2 . The relation RA is called an indiscernibility relation [28] (See Eq-8) with respect to A, and expresses which pairs of objects that we cannot discern between. xRA y ⇔ MA (x, y) = ∅ .

(8)

An alternative way to represent RA is via an indiscernibility graph (IDG), which is a graph GA = (U, RA ) with vertex set U and edge set RA . It has been stated [28] that GA is normally only interesting to consider when RA is a tolerance relation, in which case GA may be used for the purpose of clustering or unsupervised learning. Discernibility Functions. A discernibility function [28] is a function that expresses how an object or a set of objects can be discerned from a certain subset of the full universe of objects. It can be constructed relative to an object x ∈ U from a discernibility matrix MA according to Eq-9.

(9) fA (x) = a∗ : a ∈MA (x, y) and MA (x, y) = ∅ . y∈U

The function fA (x) contains |A| Boolean variables, where variable a∗ corresponds to attribute a. Each conjunction of fA (x) stems from an object y ∈ U from which x can be discerned and each term within that conjunction represents an attribute that discerns between those objects. The prime implicants of fA (x) reveal the minimal subsets of A that are needed to discern object x from the objects in U that are not members of RA (x). In addition to defining discernibility relative to a particular object, discernibility can also be defined for the information system A as a whole. The full discernibility function gA (U ) (See Eq-10) expresses how all objects in U can be discerned from each other. The prime implicants of gA (U ) reveal the minimal subsets of A we need to discern all distinct objects in U from each other.

gA (U ) = fA (x) . (10) x∈U

Reducts. If an attribute subset B ⊆ A preserves the indiscernibility relation RA then the attributes A\B are said to be dispensable. An information system may have many such attribute subsets B. All such subsets that are minimal (i.e. that do not contain any dispensable attributes) are called reducts. The set of all reducts of an information system A is denoted RED(A). In particular, minimum reducts (those with a small number of attributes), are extremely important, as decision rules can be constructed from them [4]. However, the problem of reduct computation is NP-hard, and several heuristics have been proposed [43]. Rough Clustering. Based on the concept of a rough set, modifications to the classical family of k-means algorithms have been introduced in [22] and [23] observing that

Finding Relevant Attributes in High Dimensional Data

373

in data mining it is not possible to provide an exact representation of each class in the partition. For example, an approximation image classification method has been reported in [32]. Rough sets enable such representation using upper and lower bounds. In the case of rough k-means clustering, the centroids of the clusters have to be modified to include the effects of lower and upper bounds. The modified centroid calculations for a distance-based clustering would be as shown in Eq-11 [23]: ⎧ ⎨ wlower × ν∈A(x) vj + wupper × ν∈(A(x)−A(x)) vj if A(x) − A(x) = φ |A(x)| |A(x)−A(x)| x= , ⎩ wlower × ν∈A(x) vj otherwise

|A(x)|

(11) where 1 ≤ j ≤ m (the number of clusters). The parameters wlower and wupper control the importance of the lower and upper bonds. Equation 11 generalizes the corresponding k-means centroids update. If the lower and upper bounds are equal, conventional crisp clusters would be obtained (the boundary region A(x) − A(x) is empty). The object membership w.r.t. the lower or upper bound of a cluster is determined in the following way: Let v be an object and xi , xj the centroids of clusters Xi , Xj respectively, where xi is the closest centroid to object v, and xj an arbitrary other centroid. Let d(v, xi ), d(v, xj ) be the distances from object v to the corresponding centroids, and let T be a threshold value. If d(v, xi ) − d(v, xj ) ≤ T , then v ∈ A(xi ), and v ∈ A(xj ) (i.e. v is not part of any lower bound). Otherwise, v ∈ A(xi ) and clearly v ∈ A(xi ). This algorithm depends on three parameters wlower , wupper , and T . 3.3 Virtual Reality Representation of Relational Structures The role of visualization techniques in the knowledge discovery process is well known. Several reasons make Virtual Reality (VR) a suitable paradigm: Virtual Reality is flexible, in the sense that it allows the choice of different representation models to better accommodate different human perception preferences. In other words, allows the construction of different virtual worlds representing the same underlying information, but with a different look and feel. Thus, the user can choose the particular representation that is most appealing. VR allows immersion. VR creates a living experience. The user is not merely a passive observer or an outsider, but an actor in the world. VR is broad and deep. The user may see the VR world as a whole, and/or concentrate the focus of attention on specific details of the world. Of no less importance is the fact that in order to interact with a Virtual World, no mathematical knowledge is required, but only minimal computer skills. A virtual reality based visual data mining technique, extending the concept of 3D modeling to relational structures, was introduced [40], [41] (see also http://www.hybridstrategies.com). It is oriented to the understanding of large heterogeneous, incomplete and imprecise data, as well as symbolic knowledge. The notion of data is not restricted to databases, but includes logical relations and other forms of both structured and non-structured knowledge. In this approach, the data objects are considered as tuples from a heterogeneous space [39]. Different

374

J.J. Vald´es and A.J. Barton

information sources are associated with the attributes, relations and functions, and these sources are associated with the nature of what is observed (e.g. point measurements, signals, documents, images, directed graphs, etc). They are described by mathematical sets of the appropriate kind called source sets (Ψi ), constructed according to the nature of the information source to represent. Source sets also account for incomplete information. A heterogeneous domain is a Cartesian product of a collection of source sets: ˆ n = Ψ1 × · · · × Ψn , where n > 0 is the number of information sources. H A virtual reality space is the tuple Υ =< O, G, B, m , go , l, gr , b, r >, where O is a relational structure (O =< O, Γ v > , the O is a finite set of objects, and Γ v is a set of relations), G is a non-empty set of geometries representing the different objects and relations. B is a non-empty set of behaviors of the objects in the virtual world. is the set of real numbers and m ⊂ Rm is a metric space of dimension m (Euclidean or not) which will be the actual virtual reality geometric space. The other elements are mappings: go : O → G, l : O → m , gr : Γ v → G, b : O → B. Of particular importance is the mapping l. If the objects are in a heterogeneous space, ˆ n → m . Several desiderata can be considered for building a VR-space. One may l:H be to preserve one or more properties from the original space as much as possible (for example, the similarity structure of the data [6]). From an unsupervised perspective, the role of l could be to maximize some metric/non-metric structure preservation criteria [5], or to minimize some measure of information loss. From a supervised point of view l could be chosen as to emphasize some measure of class separability over the objects in O [41]. Hybrid requirements are also possible. For example, if δij is a dissimilarity measure between any two i, j ∈ U (i, j ∈ [1, N ], where n is the number of objects), and ζiv j v is another dissimilarity measure defined on objects iv , j v ∈ O from Υ (iv = ξ(i), j v = ξ(j), they are in one-to-one correspondence). An error measure frequently used is shown in Eq-12 [35]:

2 1 i<j (δij − ζij ) . (12) Sammon error = δij i<j δij Typically, classical algorithms have been used for directly optimizing measures of this type, like Steepest descent, Conjugate gradient, Fletcher-Reeves, Powell, LevenbergMarquardt, and others. The l mappings within this paper were obtained using the method of Fletcher-Reeves [33]. The new nonlinear features are a form of dimensionality reduction and new attribute creation. 3.4 Gene Expression Programming Pattern matching and function approximation are very important operations within data mining and data analysis. Typical examples of general function approximators are neural networks and fuzzy systems. While their performance is unquestioned, their interpretation is still awkward, sometimes extremely difficult in human terms. In the case of a neural network, the understanding of its performance is obscured by the intricacies of its architecture and its weights, some times very many. In the case of a fuzzy system,

Finding Relevant Attributes in High Dimensional Data

375

the set of fuzzy rules might be large in number and complexity. Moreover, the number of linguistic variables required and the collection of membership functions, might be large as well. Therefore, either a neural network or a fuzzy model may have an excellent performance, but interpretability issues might make a human user reluctant to use them. Analytic functions, have a relation with physical systems in general, which has a long history in science. They are easier to understand by humans, the preferred building blocks of modeling, and a highly condensed form of knowledge. Regression is an example where the family of functions is restricted to a few (typically just one), and the problem reduces to finding a set of parameters or coefficients which makes the function fulfill some desirable approximation property (for example, minimizing a least square error or other model quality measure). However, direct discovery of general analytic functions poses enormous challenges because of the (in principle) infinite size of the search space. This important knowledge discovery problem can be approached from a computational intelligence perspective via evolutionary computation, and the solutions obtained are relevant to a large number of disciplines and domains. In particular genetic programming techniques aim at evolving computer programs, which ultimately are functions. Among this subfield of evolutionary computation, gene expression programming (GEP) is appealing [12]. Gene expression programming (GEP), like genetic algorithms (GAs), evolution strategies (ES) and genetic programming (GP), is an evolutionary algorithm as it uses populations of individuals, selects them according to fitness, and introduces genetic variation using one or more genetic operators. The fundamental difference between these techniques resides in the nature of the individuals. Different from GA, ES and GP, GEP individuals are nonlinear entities of different sizes and shapes (expression trees) encoded as strings of fixed length. For the interplay of the GEP chromosomes and the expression trees (ET), GEP uses a translation system to transfer the chromosomes into expression trees and vice versa [12]. The set of operators applied to GEP chromosomes always produces valid ETs. The chromosomes in GEP itself are composed of genes structurally organized into a head and a tail [11]. The head contains symbols that represent both functions (from a function set F) and terminals (from a terminal set T), whereas the tail contains only terminals. Two different alphabets occur at different regions within a gene. For each problem, the length of the head h is chosen, whereas the length of the tail t is a function of h and the number of arguments of the function with the largest arity. As an example, consider a gene composed of the function set F={Q, +, −, ∗, /}, where Q represents the square root function, and the terminal set T={a, b}. Such a gene (the tail is shown in bold) is: *Q-b++a/-bbaabaaabaab, and encodes the ET which corresponds to the mathematical √equation √ f (a, b) = b · a + ab − ((a − b) + b) , which simplifies to f (a, b) = b·a b . GEP chromosomes are usually composed of more than one gene of equal length. For each problem the number of genes as well as the length of the head has to be chosen. Each gene encodes a sub-ET and the sub-ETs interact with one another forming more complex multi-subunit ETs through a connection function. To evaluate GEP chromosomes, different fitness functions can be used.

376

J.J. Vald´es and A.J. Barton

3.5 Distributed Computing and the Grid Distributed computing can be defined in different ways, and there is no universally accepted formulation of the concept. It can be understood as an environment where idle CPU cycles and storage space of tens, hundreds, or thousands of networked systems can be harnessed to work together on a particular processing-intensive problem. The growth of such processing models has been limited, however, due to a lack of compelling applications and by bandwidth bottlenecks, combined with significant security, management, and standardization challenges. However, in the last years the interest has grown to the extent of making the technology an emergent fact. Increasing desktop CPU power and communications bandwidth have also helped to make distributed computing a more practical approach. The numbers of real applications are still somewhat limited, and the challenges (particularly standardization) are significant. Grid computing is a form of distributed computing that involves coordinating and sharing computing, application, data, storage, or network resources across dynamic and geographically dispersed organizations. As previously stated, there is no universally accepted definition, but a consensus exists in that a Grid is a type of parallel and distributed system that enables the sharing, selection, and aggregation of geographically distributed “autonomous” resources dynamically at runtime depending on their availability, capability, performance, cost, and users’ quality-of-service requirements. Grid technologies promise to change the way complex computational problems are approached and solved. However, the vision of large scale resource sharing is not yet a reality in many areas. Grid computing is an evolving area of computing, where standards and technology are still being developed to enable this new paradigm. The grid computing concept aims to promote the development and advancement of technologies that provide seamless and scalable access to wide-area distributed resources. Computational Grids enable the sharing, selection, and aggregation of a wide variety of geographically distributed computational resources (such as supercomputers, compute clusters, storage systems, data sources, instruments, people) and presents them as a single, unified resource for solving large-scale compute and data intensive computing applications (e.g, molecular modelling for drug design, brain activity analysis, and high energy physics). The idea is analogous to electric power networks (grids) where power generators are distributed, but the users are able to access electric power without bothering about the source of energy and its location. Grids aim at exploiting synergies that result from cooperation–ablity to share and agreegrate distributed computational capabilities and deliver them as service. The use of grid technologies for data mining is an obvious choice for many exploratory data analysis tasks within the knowledge discovery process. The identification of the research issues and their potential priorities for the years 2003-2010, as well as the formulation of proposal of suitable means for implementation, has been addressed by several groups of experts [14], [3]. Among distributed computing systems for delivering high throughput computing, the Condor system stands out [9], [36], [37], [38],(http://www.cs.wisc.edu/ condor/). Condor is a specialized workload management system for computeintensive jobs in a distributed computing environment, developed by the Condor

Finding Relevant Attributes in High Dimensional Data

377

Research Project at the University of Wisconsin-Madison (UW-Madison). Like other full-featured batch systems, Condor provides a job queueing mechanism, scheduling policy, priority scheme, resource monitoring, and resource management. Users submit their serial or parallel jobs to Condor, Condor places them into a queue, chooses when and where to run the jobs based upon a policy, carefully monitors their progress, and ultimately informs the user upon completion. While providing functionality similar to that of a more traditional batch queueing system, Condor’s novel architecture allows it to succeed in areas where traditional scheduling systems fail. Condor can be used to manage a cluster of dedicated compute nodes (such as a “Beowulf” cluster), possibly mixed with individual nodes. In addition, unique mechanisms enable Condor to effectively harness wasted CPU power from otherwise idle desktop workstations. For instance, Condor can be configured to only use desktop machines where the keyboard and mouse are idle. Should Condor detect that a machine is no longer available (such as a key press detected), in many circumstances Condor is able to transparently produce a checkpoint and migrate a job to a different machine which would otherwise be idle. Condor does not require a shared file system across machines - if no shared file system is available, Condor can transfer the job’s data files on behalf of the user, or Condor may be able to transparently redirect all of the job’s I/O requests back to the submit machine. As a result, Condor can be used to seamlessly combine all of an organization’s computational power into one resource. 3.6 Implementation A detailed perspective of data mining procedures provides insight into additional important issues to consider (e.g. storage/memory/communication/management/time/etc) when evaluating a computational methodology consisting of combined techniques. This study presents one possible implementation, from which more software development may occur in order to integrate better and/or different tools. In addition, all of these issues become even more pronounced when, as in this study, a complex problem is investigated. The implementation of the distributed pipeline is shown in Alg.1. It consists of two pieces; a sequential portion, and a distributed portion. For the sequential portion, a specific machine (usually the local host) is used to perform some preliminary processing on the data (as it only needs to be performed once) and then distributes the data via a specific distribution mechanism to a set of waiting computing nodes, which may include the distributing machine. Once all of the computations have completed, the sequential portion of the pipeline may then proceed to collect the results from all of the files that have been placed onto the distributing machine (again via the distribution mechanism, but this time from compute node (e.g. remote host) to distributing host). The resultant databases may then be queried for the purpose of analysis. The specific distribution mechanism used, is a high throughput pipeline (Fig. 2) consisting of many co-operating programs. Such a pipeline structure is generated automatically in order to ease the proper configuration of each participating program within the

378

J.J. Vald´es and A.J. Barton

Algorithm 1. Abstract Conceptualization of the Distributed Pipeline Input : A Data Matrix, DInput Output: A Set of Relevant Attributes. From a Specific Host, Sequentially do GenerateAndConfigurePipeline() ; DRandom ←− ShuffleObjects(i)(Opt(i) , DInput ) ; DistributeToComputeNodes(j)(Opt(j) , DRandom ) ; StartPipelineExecution() ; BlockedWaitForResults() ; // Monitor Each Job’s Progress // Store All Completed Job Results Locally ResultsRules ←− ReformatAndCollectRules(k)() ; ResultsStatistics ←− ReformatAndCollectStats(l)() ; AnalyzeResults(DRandom, ResultsRules , ResultsStatistics) ; end On Each Compute Node Run A Job And do DLeaders ←− ConstructLeaders(m)(Opt(m) , DRandom ) ; DSubsets ←− SubsetSelection(n)(Opt(n) , DLeaders ) ; // e.g. create DSubsets by 10-fold cross-validation forall (DTi r , DTi e ) ∈ DSubsets do r (DTi,discr , CutsTi r ) ←− Discretize(o)(OptT(o)r , DTi r ) ; r ); ReductsTi r ←− FormReducts(p) (OptT(p)r , DTi,discr Tr Tr Rulesi ←− GenerateRules(q)(Opt(q) , ReductsTi r ) ; e ←− Discretize(o)(OptT(o)e , DTi e , CutsTi r ) ; DTi,discr e ); RuleSetMeriti ←− Classify(r) (Opt(r) , DTi,discr Record(RuleSetMeriti) end end

pipeline. In this paper, the automatically generated pipeline was facilitated via i) a file generation program (written in Python and running on the local host) and ii) the Condor tool described in section 3.5. The initial preprocessing stage of the pipeline, occurring on the distributing host after generation of files, involves shuffling the input data records as described previously and in Fig.1. The shuffled data is stored on the distributing host’s disk, in order to provide the same randomized data to the next stage of processing, which occurs on the computing hosts (Fig.2). A Condor submission program, which was also automatically generated, is used to specify all of the data and configuration files for the programs that will execute on the remote host. The submission process enables Condor to: i) ii) iii) iv)

schedule jobs for execution check point them (put a job on hold) transfer all data to the remote host transfer all generated data back to the local host (submitting machine)

Finding Relevant Attributes in High Dimensional Data

379

Fig. 1. Data processing strategy combining clustering, Rough Sets analysis and crossvalidation

The final postprocessing stage of the pipeline involves collecting all of the results (parsing the files) and reporting them in a database.

4 Experimental Methodology The datasets consist of information systems with an attribute set composed of ratio and interval variables, and a nominal or ordinal decision attribute. More general information systems have been described in [39]. The general idea is to construct subsets of relatively similar attributes, such that a simplified representation of the data objects is obtained by using the corresponding attribute subset representatives. The attributes of these simplified information systems are explored from the point of view of their reducts. From them, rules are learned and applied systematically to testing data subsets not involved in the learning process (Fig.1). The whole procedure can be seen as a pipeline. In a first step, the objects in the dataset are shuffled using a randomized approach in order to reduce the possible biases introduced within the learning process by data chunks sharing the same decision attribute. Then, the attributes of the shuffled dataset are clustered using the two families of fast clustering algorithms described in previous sections (the leader, and k-means). Each of the formed clusters of attributes is

380

J.J. Vald´es and A.J. Barton

Fig. 2. Automatically generated high throughput pipeline oriented towards the Condor distributed computing environment

represented by exactly one of the original data attributes. By the nature of the leader algorithm, the representative is the leader (called an l-leader), whereas for a k-means algorithm, a cluster is represented by the most similar object w.r.t. the centroid of the corresponding cluster (called a k-leader). This operation can be seen as a filtering of the attribute set of the original information system. As a next step, the filtered information system undergoes a segmentation with the purpose of learning classification rules, and testing their generalization ability in a cross-validation framework. N-folds are used as training sets; where the numeric attributes present are converted into nominal attributes via a discretization process (many possibilities exist), and from them, reducts are constructed. Finally, classification rules are built from the reducts, and applied to a discretized version of the test fold (according to the cuts obtained previously), from which the generalization ability of the generated rules can be evaluated. Cross-validation is used in order to create a statistically meaningful estimate of the classification accuracy of the rules generated from the reducts for a particular experiment. The final database of experimental results was then sorted by minimum mean classification accuracy. The best mean accuracy experiments were then selected in order to extract the attributes from within the computed reducts for further analysis. Besides the numeric descriptors associated with the application of classification rules to data, use of visual data mining techniques, like the virtual reality representation (section 3.3), enables structural understanding of the data described in terms of the selected

Finding Relevant Attributes in High Dimensional Data

381

subset of attributes and/or the rules learned from them. This technique can be applied at a pre- and/or post-processing stage. In this paper, all of the applications were made in the unsupervised mode. That is, the existing class information was not used during the process, and is incorporated within the resulting visualization only to enable thecomparison between the data structure provided by the predictor variables with the class distribution that is known to exist. Each stage of the process feeds its results to the next stage of processing, yielding a pipelined data analysis stream, with partial outputs that can be used for other kinds of analysis. 4.1 ROSETTA The ROSETTA Software [28], [30] was used within this study with the algorithms that are described within the following sections. Discretization: NaiveScaler. The heuristic implemented by Rosetta [29] was used and was described under the assumption that all condition attributes A are numerical. For each condition attribute a, sort its value set Va to obtain the ordering indicated by Eq-13. va1 < · · · vai < · · · < va|Va | . (13) Then let Ca denote the set of all cuts for attribute a generated in a naive fashion according to equations Eq-14, Eq-15, and Eq-16. (14) Xai = x ∈ U : a (x) = vai , Δia = v ∈ Vd : ∃x ∈ Xai such that d (x) = v , and Ca =

vai + vai+1 i > 1 or Δia = Δi+1 : Δa > 1 or Δi+1 a a 2

(15) .

(16)

The set Ca consists of all cuts midway between two observed attribute values, except for the cuts that are clearly not needed due to the fact that the objects have the same decision value. Hence, such a cut would not discern the objects. If no cuts are found for an attribute, NaiveScaler leaves the attribute unprocessed. Missing values are ignored in the search for cuts. In the worst case, each observed value is assigned its own interval. Discretization: SemiNaiveScaler. The discretization algorithm as implemented within Rosetta [29] is similar to the NaiveScaler but has more logic to handle the case where value-neighboring objects belong to different decision classes. This algorithm typically results in fewer cuts than the simpler NaiveScaler, but may still produce more cuts than are desired. In Eq-17, the set Dai collects the dominating decision values for the objects in Xai . If there are no ties, Dai is a singleton. The rationale used within Rosetta for not adding a cut if the sets of dominating decisions define an inclusion is that it is hoped (although it is stated that the implementation does not check)

382

J.J. Vald´es and A.J. Barton

that a cut will be added for another attribute (different from a) such that the objects in Xai and Xai+1 can be discerned. i i . (17) Da = v ∈ Vd : v = argmax x ∈ Xa : d (x) = v Ca =

v

vai + vai+1 : Dai Dai+1 and Dai+1 Dai 2

.

(18)

Discretization: RSESOrthogonalScaler. This algorithm is an efficient implementation [25] of the Boolean reasoning algorithm [27] within the Rough Set Exploration System (RSES) (See http://logic.mimuw.edu.pl/˜rses/). It is mentioned [29] that this algorithm is functionally similar to BROrthogonalScaler but much faster. Approximate solutions are not supported. If a(x) is missing, object x is not excluded from consideration when processing attribute a, but is instead treated as an infinitely large positive value. If no cuts are found for an attribute, all entries for that attribute are set to 0. Discretization: BROrthogonalScaler. The Rosetta implementation [29] of a previously outlined algorithm [27] was used, which is based on the combination of the NaiveScaler algorithm previously presented and a Boolean reasoning procedure for discarding all but a small subset of the generated cuts. Construct set of candidate cuts Ca according to Eq-16. Then construct a boolean function f from the set of candidate cuts according to Eq-19.

c∗ : c ∈Ca and a (x) < c < a (y) and ∂A (x) = ∂A (y) . (19) f= (x,y) a

Then compute the prime implicant of f using a greedy algorithm [21] (see JohnsonReducer). This Boolean reasoning approach to discretization may result in no cuts being deemed necessary (because they do not aid discernibility) for some attributes. The Rosetta implementation does not alter such attributes. Discretization: EntropyScaler. The Rosetta implementation [29] of the algorithm [8] is based on recursively partitioning the value set of each attribute so that a local measure of entropy is optimized. The minimum description length principle defines a stopping criterion for the partitioning process. Rosetta ignores missing values in the search for cuts and Rosetta does not alter attributes for which no cuts were found. Reduct Computation: RSESExhaustiveReducer. The RSES algorithm included within Rosetta [29] computes all reducts by brute force. Computing reducts is NP-hard, so information systems of moderate size are suggested to be used within Rosetta. Reduct Computation: Holte1RReducer. Rosetta’s [29] algorithm creates all singleton attribute sets, which was inspired by a paper in Machine Learning [18]. The set of all 1R rules, (i.e. univariate decision rules) are thus directly constructed from the attribute sets.

Finding Relevant Attributes in High Dimensional Data

383

Reduct Computation: RSESJohnsonReducer. Rosetta [29] invokes the RSES implementation of the greedy algorithm [21] for reduct computation. No support is provided for IDGs, boundary region thinning or approximate solutions. Reduct Computation: JohnsonReducer. Rosetta [29] invokes a variation of a greedy algorithm to compute a single reduct [21]. The algorithm (See Alg.2) has a natural bias towards finding a single prime implicant of minimal length. The reduct R is found by executing the following algorithm, where w(X) denotes a weight for set X ∈ S that is computed from the data. Support for computing approximate solutions is provided by aborting the loop when enough sets have been removed from S, instead of requiring that S has to be fully emptied.

Algorithm 2. Johnson Reducer Input : A Data Matrix, DInput Output: One Reduct R←∅; // Reduct has no attributes within it S ← {S1 , S2 , . . . , Sn } ; repeat // A contains all attributes that maximizes w(X), // where the sum is taken over all sets X ∈ S that // contain a. w(X)}} ; A ← {a : maximal { {X∈S:a∈X}

// The Rosetta implementation resolves ties arbitrarily a ← RandomElementFromSet (A) ; R ← R ∪ {a} ; // Add attribute to growing reduct S ← {X ∈ S : a ∈ / X} ; // Stop considering sets containing a until S = ∅ ; // No more attributes left for consideration return R ;

Rule Generation: RSESRuleGenerator. Rosetta [29] invokes the RSES implementation of an algorithm to generate rules from a set of reducts. Conceptually performed by overlaying each reduct in the reduct set over the reduct set’s parent decision table and reading off the values.

5 Results 5.1 Leukemia Gene Expression Data The example high dimensional dataset selected is that of [15], and consists of 7129 genes where patients are separated into i) a training set containing 38 bone marrow samples: 27 acute lymphoblastic leukemia (ALL) and 11 acute myeloid leukemia (AML),

384

J.J. Vald´es and A.J. Barton

obtained from patients at the time of diagnosis, and ii) a testing set containing 34 samples (24 bone marrow and 10 peripheral blood samples), where 20 are ALL and 14 AML. The test set contains a much broader range of biological samples, including those from peripheral blood rather than bone marrow, from childhood AML patients,and from different reference laboratories that used different sample preparation protocols. Further, the dataset is known to have two types of ALL, namely B-cell and T-cell. For the purposes of investigation, only the AML and ALL distinction was made. The dataset distributed by [15] contains preprocessed intensity values, which were obtained by rescaling such that overall intensities for each chip are equivalent (A linear regression model using all genes was fit to the data). In this paper no explicit preprocessing of the data was performed, in order to not introduce bias and to be able to expose the behavior of the data processing strategy, the methods used, and their robustness. That is, no background subtraction, deletions, filtering, or averaging of samples/genes were applied, as is typically done in gene expression experiments. In a preprocessing stage, a virtual reality representation of the opriginal dataset in a 3-dimensional space as described in section 3.3 was computed. Gower similarity was used for the original space, and normalized Euclidean distance for the target space. Steepest descent was used for optimizing Sammon’s error. The purpose was to appreciate the relationship of the structure of the existing classes and the collection of original attributes. As shown in (Fig.3, the two Leukemia classes appear completely mixed, as approximated with the original set of attributes. Noisy attributes do not allow a resolution of the classes. The pipeline (Fig.1) was investigated through the generation of 480 k-leader and 160 l-leader for a total of 640 experiments (Table-1). The discretization, reduct computation and rule generation algorithms are those included in the Rosetta system [30]. This approach leads to the generation of 74 files per experiment, with 10-fold cross-validation. From the experiments completed so far, one was chosen which illustrates the kind of results obtained with the explored methodology. It corresponds to a leader clustering algorithm with a similarity threshold of 0.99 (leading to 766 l-leader attributes), used as input to the data processing pipeline containing 38 samples. The results of the best 10 fold cross-validated experiment has a mean accuracy of 0.925 and a standard deviation Table 1. The set of parameters and values used in the experiments using the distributed pipeline environment Algorithm/Parameter Leader Leader Similarity Threshold K-Means Cross-validation Discretization Reduct Computation Rule Generation

Values ReverseSearch, ClosestSearch 0.7, 0.8, 0.9, 0.95, 0.99, 0.999, 0.9999, 0.99999 Forgy, Jancey, Convergent, MacQueen 10 folds BROrthogonalScaler, EntropyScaler, NaiveScaler, SemiNaiveScaler JohnsonReducer, Holte1RReducer RSESRuleGenerator

Finding Relevant Attributes in High Dimensional Data

385

Fig. 3. Snapshot of the Virtual Reality representation of the original Leukemia data (training set with 38 samples + test set with 34, both with 7129 genes). Dark objects= ALL class, Light objects=AML class. Spheres = training, Cubes = test. Representation error = 0.143, Sammon error = 3.56e − 6.

of 0.168. This experiment led to 766 reducts (all of them singleton attributes), which was consistent across each of the 10 folds. The obtained classification accuracy represents a slight improvement over those results reported in [42] (0.912). It was conjectured in that study that the introduction of a cross-validated methodology could improve the obtained classification accuracies, which is indeed the case. It is interesting to observe that all of the 7 relevant attributes (genes) reported in [42] are contained (subsumed) within the single experiment mentioned above. Moreover, they were collectively found using both the leader and k-means algorithms, with different dissimilarity thresholds and number of clusters, whereas with the present approach, a single leader clustering input was required to get the better result. Among the relevant attributes (genes) obtained, many coincide with those reported by [15], [10], and [42]. At a post-processing stage, a virtual reality space representation of the above mentioned experiment is shown in Fig.4. Due to the limitations of representing an interactive virtual world on static media, a snapshot from an appropriate perspective is presented. Sammon’s error [35] was used as criteria for computing the virtual reality space, and Gower’s similarity was used for characterizing the data in the space of the 766 selected genes. After 200 iterations a satisfactory error level of 0.0998 was obtained. It is interesting to see that the ALL and AML classes can be clearly differentiated.

386

J.J. Vald´es and A.J. Barton

Fig. 4. Snapshot of the Virtual Reality representation of the union of all of the reducts obtained from 10 fold cross-validation input (38 samples with 766 genes) for the Leukemia data. The leader clustering algorithm was used with a similarity threshold of 0.99. The ALL and the AML classes are perfectly separated. Representation error = 0.0998.

5.2 Breast Cancer Data A visual representation of the original data in terms of the 12, 625 genes obtained using Gower similarity in the original space, Sammon error, and steepest descent optimization [34] is shown in Fig.5. The sensitive and resistant classes are shown for comparison, and semi-transparent convex hulls wrap the objects from the corresponding classes. There is a little overlap between the two sets, indicating the classification potential of the whole set of attributes, but complete class resolution is not obtained with the nonlinear coordinates of the VR space. Rough k-means [23] was used to cluster the 24 samples into 2 groups using the whole set of original attributes (genes) in order to illustrate the difficulties involved in using all of the original attributes. The particular algorithm parameters (wlower = 0.9, wupper = 0.1, distanceT hreshold = 1). The rough k-means result for the 24 samples using the 12, 625 original attributes, and requesting 2 classes is shown in Fig.6. In the VR space, 2 classes are clearly well differentiated, but one of them contains 5 objects and the other contains the rest. Moreover, when the smaller class is investigated, it contains a mixture of samples from the resistant and sensitive class. Therefore, even the more elaborated rough k-means clustering can not resolve the two known classes from the point of view of all of the original attributes used at the same time. It is also interesting, that for this dataset that no boundary cases were obtained with the clustering parameters used.

Finding Relevant Attributes in High Dimensional Data

387

Fig. 5. Visual representation of 24 breast cancer samples with 12, 625 genes. Convex hulls wrap the resistant(size= 14) and sensitive(size= 10) classes. Absolute Error = 7.33 · 10−2 . Relative Mapping Error = 1.22 · 10−4 .

The experimental settings used in the investigation of breast cancer data with the distributed pipeline are reported in Table 2. For each experiment, the discretization, reduct computation and rule generation algorithms are those included in the Rosetta system [30]. The leader algorithm variants were described in section 3.1. This approach leads to the generation of 84 files per experiment, with 10-fold cross-validation. From the series of l-leader Breast Cancer experiments performed, 4 experiments (Exp-81, Exp-82, Exp-145, Exp-146) were found to be equivalent when analyzing the mean (0.73), median (0.67), standard deviation (0.25), minimum (0.5) and maximum (1.0) of the 10-fold cross validated classification accuracy of the produced rules. For the l-leader algorithm a similarity threshold of 0.7 was used by all experiments, with Exp81 and Exp-145 using closest placement criteria and Exp-82 and Exp-146 using reverse search criteria. The discretization algorithm as provided by the Rosetta system was the RSESOrthogonalScaler for Exp-81 and Exp-82 and the BROrthogonalScaler for Exp145 and Exp-146. The reduct algorithm (RSESExhaustiveReducer) was the same for all 4 experiments with full discernibility and all selection. The rule generation algorithm (RSESRuleGenerator), was also the same for all 4 experiments. In a postprocessing stage, the gene expression programming was applied to selected pipeline results. The idea was to try a simple function set (F = {+, −, ∗}) without the use of numeric constants, in order to reduce the complexity of the assembled functions as much as possible. In particular, experiments 81, 82, 145 and 146 all found a subset

388

J.J. Vald´es and A.J. Barton

Fig. 6. Visual representation of 24 breast cancer samples with 12, 625 genes. Convex hulls wrap the RC1(size= 19) and RC2(size= 5) classes built by rough set based k-means. Absolute Error = 7.06 · 10−2 . Relative Mapping Error = 5.21 · 10−5 . RC1, RC2 stand for the two rough clusters obtained.

of only 3 l-leader attributes with a mean crossvalidation error of 0.73. The subset of attributes was {2, 139, 222}, corresponding to genes {31307 at, 31444 s at, 31527 at}. They represent {12359, 27, 239} data objects respectively. Accordingly, the terminal set defined by the attributes found by the pipeline was set to T = {v2 , v139 , v222 }. The resulting class membership function emerging from the GEP process is as shown in Eq-20: f (v2 , v139 , v222 ) = ((v222 + v139 ) + v139 ) +

(20)

(((v222 − v139 ) ∗ (v2 ∗ v2 )) ∗ ((v139 + v139 ) − v139 )) + (v222 − (((v2 − v222 ) ∗ (v139 − v2 )) ∗ ((v222 + v139 ) + v2 ))) . This analytic expression was simplified with the Yacas computer algebra system http://www.xs4all.nl/˜apinkus/yacas.html which resulted in the expression in Eq-21: 2 2 2 f (v2 , v139 , v222 ) = v222 ∗ v139 − v222 ∗ v2 + v222 ∗ v139 +

v222 ∗ v139 ∗ v22 − v222 ∗ v139 ∗ v2 + 2 ∗ v222 + 2 2 ∗ v2 + v139 ∗ v22 ) + v23 . 2 ∗ v139 − (v139

(21)

Finding Relevant Attributes in High Dimensional Data

389

Table 2. The set of parameters and values used in the experiments with the Breast Cancer dataset using the distributed pipeline environment Algorithm/Parameter Leader Leader Similarity Threshold Cross-validation Discretization Reduct Computation Rule Generation

Values ReverseSearch, ClosestSearch 0.7, 0.8, 0.9, 0.95, 0.99, 0.999, 0.9999, 0.99999 10 folds BROrthogonalScaler, EntropyScaler, NaiveScaler, RSESOrthogonalScaler, SemiNaiveScaler RSESExhaustiveReducer, RSESJohnsonReducer RSESRuleGenerator

The classification rule associated with Eq-21 is shown in Eq-22. The classification accuracy on the original dataset was 91.67%, and it should be noted that only 3 genes out of the 12, 625 original ones are used. Moreover, the resulting model is relatively simple. IF f (v2 , v139 , v222 ) ≥ 0.5) −→ class = sensitive

(22)

otherwise −→ class = resistant . 5.3 Colon Cancer Data A virtual reality space representation of the dataset in terms of the original 2000 attributes was computed for an initial assesment of the structure of the data. Sammon error was used as structure measure, with normalized Euclidean distance as dissimilarity, and Powell’s method for error optimization [34]. In 50 iterations an extremely low mapping error obtained (1.067x10−6) is shown in Fig.7.

Fig. 7. Visual representation (3 dimensions) of 62 colon cancer samples with 2000 genes. Darker objects belong to the tumor class, and lighter objects to the normal class. After 50 iterations: Absolute Error = 1.067 · 10−6 . Relative Mapping Error = 0.0488.

390

J.J. Vald´es and A.J. Barton

Table 3. Two Breast Cancer dataset experiments and their associated reducts for each of the 10 cross-validated folds. The GEP encodings are also reported. Experiment

81

82

Fold 1 2 3 4 5 6 7 8 9 10 1 2 3 4 5 6 7 8 9 10

Reducts {31307 at, 31444 s at, 31527 {31307 at, 31527 at} {31307 at, 31527 at} {31307 at, 31444 s at, 31527 {31307 at, 31444 s at, 31527 {31307 at, 31444 s at, 31527 {31307 at, 31444 s at, 31527 {31307 at, 31444 s at, 31527 {31307 at, 31444 s at, 31527 {31307 at, 31527 at} {31307 at, 31444 s at, 31527 {31444 s at, 31527 at} {31307 at, 31527 at} {31307 at, 31444 s at, 31527 {31307 at, 31444 s at, 31527 {31307 at, 31444 s at, 31527 {31307 at, 31527 at} {31307 at, 31444 s at, 31527 {31307 at, 31527 at} {31307 at, 31527 at}

at}

at} at} at} at} at} at} at}

at} at} at} at}

GEP Encoding {v2 , v139 , v222 } {v2 , v222 } {v2 , v222 } {v2 , v139 , v222 } {v2 , v139 , v222 } {v2 , v139 , v222 } {v2 , v139 , v222 } {v2 , v139 , v222 } {v2 , v139 , v222 } {v2 , v222 } {v2 , v139 , v222 } {v139 , v222 } {v2 , v222 } {v2 , v139 , v222 } {v2 , v139 , v222 } {v2 , v139 , v222 } {v2 , v222 } {v2 , v139 , v222 } {v2 , v222 } {v2 , v222 }

Such a small mapping error indicates that the VR space is a very accurate portrait of the 2000 dimensional original space. The most interesting feature in the VR space is the existence of an intrinsic unidimensionality in the data from the point of view of preserving the distance structure. Although the right hand side of the projected data line in Fig.7 predominantly contains objects of the tumor class, and the left half objects of the normal classes, they are mixed, and therefore, the space does not resolve the classes. Nevertheless, this result is an indication about the potential of finding relatively small subsets of attributes with reasonable classification power and about the large redundancy within the original attributes. The experimental settings used in the investigation of colon cancer data with the distributed pipeline are reported in Table 4.For each experiment, the discretization, reduct computation and rule generation algorithms are those included in the Rosetta system [30]. The leader algorithm variants were described in section 3.1. This approach leads to the generation of 84 files per experiment, with 10-fold cross-validation. From the series of 320 l-leader Colon Cancer experiments, 5 are selected for illustration: Exp-113, Exp-304, Exp-195, Exp-180, and Exp-178. They were found to be equivalent when analyzing the mean (0.73), median (0.67), standard deviation (0.25), minimum (0.5), and maximum (1.0) of the 10-fold cross validated classification accuracy.From the series of l-leader experiments, 2 were selected for illustrating the number of created rules per fold. The i) rules and their respective reducts from which they were generated are shown in Table 6 and ii) the original attribute names as well as related genes as found

Finding Relevant Attributes in High Dimensional Data

391

Table 4. The set of parameters and values used in the experiments with the Colon Cancer dataset using the distributed pipeline environment Algorithm/Parameter Leader Leader Similarity Threshold Cross-validation Discretization Reduct Computation Rule Generation

Values ReverseSearch, ClosestSearch 0.7, 0.8, 0.9, 0.95, 0.99, 0.999, 0.9999, 0.99999 10 folds BROrthogonalScaler, EntropyScaler, NaiveScaler, RSESOrthogonalScaler, SemiNaiveScaler JohnsonReducer, Holte1RReducer, RSESExhaustiveReducer, RSESJohnsonReducer RSESRuleGenerator

Table 5. Selected l-leader Colon Cancer experiments sorted by minimum 10-fold cross-validated classification accuracy. The last column shows the resultant number of pipeline selected attributes (from 2, 000) for each experiment. Experiment Mean Median Standard Deviation Min. Max. Sim. No. Attr. 178 0.562 0.500 0.261 0.167 1.0 0.7 3 180 0.629 0.646 0.219 0.333 1.0 0.8 9

within the source reference [1] are shown in Table 7. In these cases, all reducts were composed of singleton attributes found from the original 2000. They are presented due to their cross-validated minimum and maximum accuracies of [0.167-1.0] and [0.3331.0] respectively. Experiment 178 contains the same 3 singleton reducts in each of the 10 folds, from which [59 − 73] rules were obtained. Whereas, Experiment 180 contains 9 singleton reducts in each of the 10 folds, from which [209 − 241] rules were obtained. For the found l-leaders in these experiments, it can be seen that perfect classification has been made for some of the folds which is a sign of interestingness. In a postprocessing stage, gene expression programming was applied to selected pipeline results. The idea was to try a simple function set (F = {+, −, ∗, sin, cos, log}) in order to reduce the complexity of the assembled nonlinear functions as much as possible. In particular, experiment 178 found a subset of only 3 l-leader attributes. The subset of attributes was {1, 2, 12}, corresponding to genes {H55933, R39465, H86060}. They represent {1, 11, 1988} data objects respectively. Accordingly, the terminal set defined by the attributes found by the pipeline was set to T = {v1 , v2 , v12 }. The resulting class membership function emerging from the GEP process is shown in Eq-23: f (v1 , v2 , v12 ) = ((v1 ∗ (v1 ∗ cos((v12 ∗ v12 )))) − v2 ) + (v1 ∗ (cos(log(v1 )) ∗ v2 )) + ((v2 ∗ (cos((v2 + v2 )) ∗ v2 )) + v1 ) .

(23)

392

J.J. Vald´es and A.J. Barton

When Eq-23 is simplified, the resultant equation is that as shown in Eq-24: 2 ) ∗ v12 + v1 ∗ v2 ∗ cos(log(v1 )) + f (v1 , v2 , v12 ) = cos(v12 2 v1 + v2 ∗ cos(2 ∗ v2 ) − v2 .

(24)

The classification rule associated to Eq-24 is shown in Eq-25, which has a classification accuracy on the original dataset of 88.7%, and it should be noted that only 3 genes out of the 2000 original ones were used. IF (f (v1 , v2 , v12 ) ≥ 0.5) −→ class = normal otherwise −→ class = tumor .

(25)

In a second application of the gene expression programming method, the attributes found by experiment 180 {1, 2, 4, 5, 6, 22, 27, 119, 878} were used (again l-leaders). The terminal set was allowed to have numeric constants. The resulting class membership function emerging from the GEP process is as shown in Eq-26: f (v1 , v2 , v4 , v5 , v6 , v22 , v27 , v119 , v878 ) = (((((v878 − v22 ) ∗ v5 ) ∗ (v4 ∗ v878 )) ∗ ((v6 + v5 ) + v5 )) + v1 ) + (((((v119 − v878 ) ∗ v119 ) ∗ v1 ) ∗ ((v27 + v878 ) ∗ (v1 + v22 ))) + v1 ) + (v27 − (v5 ∗ (k1 − (v119 + ((v119 + v1 ) + v5 ))))) + (k2 ∗ (((v1 − (v878 + v5 )) + v4 ) ∗ v6 )) ,

(26)

where k1 = 3.55777, k2 = −7.828919. After simplification, the resulting function is as shown in Eq-27: f (v1 , v2 , v4 , v5 , v6 , v22 , v27 , v119 , v878 ) = 2 2 ∗ v878 ∗ v52 ∗ v4 + 2 2 2 ∗ v5 ∗ v4 ∗ v6 − v878 ∗ v12 ∗ v119 − v878 ∗ v22 ∗ v1 ∗ v119 + v878

(−2) ∗ v878 ∗ v22 ∗ v52 ∗ v4 − v878 ∗ v22 ∗ v5 ∗ v4 ∗ v6 + 2 − v878 ∗ v22 ∗ v1 ∗ v119 ∗ v27 + v878 ∗ v22 ∗ v1 ∗ v119 2 v878 ∗ v12 ∗ v119 − v878 ∗ v12 ∗ v119 ∗ v27 − v878 ∗ v6 ∗ k2 + 2 ∗ v27 + v22 ∗ v1 ∗ v119

v52 + v5 ∗ v1 + 2 ∗ v5 ∗ v119 − v5 ∗ k1 − v5 ∗ v6 ∗ k2 + v4 ∗ v6 ∗ k2 + v6 ∗ v1 ∗ k2 + 2 ∗ v27 + v12 ∗ v119 2 ∗ v1 + v27 .

(27)

Finding Relevant Attributes in High Dimensional Data

393

The classification rule associated to Eq-27 is as shown in Eq-28: IF (f (v1 , v4 , v5 , v6 , v22 , v27 , v119 , v878 ) ≥ 0.5) −→ class = normal otherwise −→ class = tumor ,

(28)

Table 6. Two Colon Cancer dataset experiments. Exp. 178 has 3 reducts that are the same in all 10 folds. Exp. 180 has 9 reducts that are the same in all 10 folds. Cross Validation Fold No. Rules in Exp. 178 No. Rules in Exp. 180

1 2 3 4 5 6 7 8 9 10 67 72 73 72 68 59 66 72 68 68 230 236 233 234 227 209 216 241 228 227

Table 7. Discovered attributes for 2 Colon Cancer dataset experiments. Exp. 180 found one attribute (v6 =R02593) that was also previously reported. Exp. 178

Exp. 180

Encoding: Original: Encoding: Original: Encoding Original: Compared to [1]:

v1 H55933 v1 H55933 v4 R85482 (R85464)

v2 R39465 v2 R39465 v5 U14973 (U14971)

v12 H86060 v22 J02763 v6 R02593 (Same)

v27 v119 H86060 T72175 v878 M87789

and has a classification accuracy on the original dataset of 91.9%. From the point of view of classification accuracy, it is only slightly better than the one obtained with only 3 attributes. On the other hand, despite the fact that most of the individual terms are relatively simple (addition is the root of the expression tree), the expression as a whole is very complex, and certainly much more than the previous model. Likely such an expression is not an arguable replacement for a neural network or a set of fuzzy relations in terms of simplicity or understandability, and moreover, the situation can be even worse if the function set is extended with other nonlinear functions like ex , ln(x), transcendental functions, numeric constants, etc., as is required in complex function approximation tasks. However, despite these difficulties, genetic programming, and particularly GEP allows an explicit assessment of the role of predictor variables. It also provides analytic means to perform sensitivity analysis directly, through the study of the partial derivatives and the multidimensional gradient of the generated functions. The approach is promising, and new developments in genetic programming including meta-function approximation and the incorporation of more intelligent techniques may overcome the above mentioned difficulties. It is interesting to observe that gene 2 is not found in Eq-28, indicating that it was irrelevant (boldfaced in Eq-27). However, despite the increased complexity, this independent technique showed that the set of genes suggested by the data mining pipeline has an important classification power.

394

J.J. Vald´es and A.J. Barton

6 Conclusions Good results were obtained with the proposed high throughput pipeline based on the combination of clustering and rough sets techniques for the discovery of relevant attributes in high dimensional data. The use of several clustering and rough set analysis techniques, and their combination as a virtual data mining machine implemented in a grid and high throughput computing environment, proved to be a promising way to address complex knowledge discovery tasks. In particular, the introduction of a fast attribute reduction procedure aided rough set reduct discovery in terms of computational time, of which the former is further improvable via its amenability for parallel and distributed computing. Cross-validated experiments using three different sets of gene expression data demonstrate the possibilities of the proposed approach. With the hybrid methodology presented, in all cases it was possible to find subsets of attributes of size much smaller than the original set, retaining a high classification accuracy. The pre- and post-processing stages of visual data mining using multidimensional space mappings and genetic programming techniques like gene expression programming (in combination with computer algebra systems), are effective elements within the data processing strategy proposed. The analytic functional models obtained for evaluating class memberships and ultimately for classification via gene expression programming, allowed a better understanding of the role of the different attributes in the classification process, as well as an explicit explanation of their influence. More thorough studies are required to correctly evaluate the impact of the experimental settings on the data mining effectiveness, and further experiments with this approach are necessary.

Acknowledgements This research was conducted within the scope of the BioMine project (National Research Council Canada (NRC), Institute for Information Technology (IIT)). The authors would like to thank Robert Orchard and Marc Leveille from the Integrated Reasoning Group (NRC-IIT), and to Ratilal Haria and Roger Impey from the High Performance Computing Group (NRC-IIT).

References 1. Alon, U., Barkai, N., Notterman, D.A., Gish, K., Ybarra, S., Mack, D., Levine, A.J.: Broad patterns of gene expression revealed by clustering analysis of tumor and normal colon tissues probed by oligonucleotide arrays. In: Proceedings National Academy of Science. USA v96 (1999) 67456750 2. Anderberg, M.: Cluster Analysis for Applications. Academic Press (1973) 3. Bal, H., de Laat, C., Haridi, S., Labarta, J., Laforenza, D., Maccallum, P., Mass, J., Matyska, L., Priol, T., Reinefeld, A., Reuter, A., Riguidel, M., Snelling, D., van Steen, M.: Next Generation Grid(s) European Grid Research 2005 - 2010 Expert Group Report, (2003) 4. Bazan, J.G., Skowron, A., Synak, P.: Dynamic Reducts as a Tool for Extracting Laws from Decision Tables. In: Proceedings of the Symp. on Methodologies for Intelligent Systems. Charlotte, NC, Oct. 16-19 1994. Lecture Notes in Artificial Intelligence 869, Springer-Verlag (1994) 346-355

Finding Relevant Attributes in High Dimensional Data

395

5. Borg, I., Lingoes, J.: Multidimensional similarity structure analysis. Springer-Verlag, New York, NY (1987) 6. Chandon, J.L., Pinson, S.: Analyse typologique. Thorie et applications. Masson, Paris (1981) 7. Chang, J.C. et al.: Gene expression profiling for the prediction of therapeutic response to docetaxel in patients with breast cancer. Mechanisms of Disease. The Lancet, vol 362 (2003) 8. Dougherty, J., Kohavi, R., Sahami,M.: Supervised and unsupervised discretization of continuous features. In A. Prieditis and S. Russell, editors, Proc. Twelfth International Conference on Machine Learning, Morgan Kaufmann (1995) 194-202 9. Epema, D.H.J., Livny, M., van Dantzig, R., Evers, X., Pruyne, J.: A worldwide flock of Condors: Load sharing among workstation clusters. Journal of Future Generation Computer Systems, (1996) 53-65 10. Famili, F., Ouyang, J.: Data mining: understanding data and disease modeling. In: Proceedings of the 21st IASTED International Conference, Applied Informatics, Innsbruck, Austria, Feb. 10-13, (2003) 32-37 11. Ferreira C.: Gene Expression Programming: A New Adaptive Algorithm for Problem Solving. Journal of Complex Systems ˇ13, 2, (2001) 87-129 12. Ferreira C.: Gene Expression Programming: Mathematical Modeling by an Artificial Intelligence. Angra do Heroismo, Portugal (2002) 13. Forgy, E.W.: Cluster analysis of multivariate data: Efficiency versus interpretability of classifications. Biometric Soc. Meetings, Riverside, California (abstract in Biometrics, v.21, no. 3 (1965) 768 14. Foster, I., Kesselman, C., Tuecke, S.: The Anatomy of the Grid: Enabling Scalable Virtual Organizations. International. Journal of Supercomp. App., v.15(3):20 (2001) 222-237 15. Golub, T.R., et al.: Molecular classification of cancer: class discovery and class prediction by gene expression monitoring. Science, vol. 286 (1999) 531-537 16. Gower, J.C.: A general coefficient of similarity and some of its properties. Biometrics, v.1, no. 27, (1973) 857-871 17. Hartigan, J.: Clustering Algorithms. John Wiley & Sons (1975) 18. Holte, R.C.: Very simple classification rules perform well on most commonly used datasets. Machine Learning, 11(1) April (1993) 63-91 19. Jain, A.K., Mao, J.: Artificial Neural Networks for Nonlinear Projection of Multivariate Data. In: Proceedings 1992 IEEE Joint Conf. on Neural Networks (1992) 335-340 20. Jancey, R.C.: Multidimensional group analysis. Australian Journal of Botany, v.14, no. 1 (1966) 127-130 21. Johnson, D.S.: Approximation algorithms for combinatorial problems. Journal of Computer and System Sciences, 9 (1974) 256-278 22. Lingras, P.: Unsupervised Rough Classification using GAs. Journal of Intelligent Information Systems v16, 3 Springer-Verlag (2001) 215-228 23. Lingras, P., Yao, Y.: Time Complexity of Rough Clustering: GAs versus K-Means. Third. Int. Conf. on Rough Sets and Current Trends in Computing RSCTC 2002. Alpigini, Peters, Skowron, Zhong (Eds.) Lecture Notes in Computer Science (Lecture Notes in Artificial Intelligence Series) LNCS 2475. Springer-Verlag (2002) 279-288 24. MacQueen, J.B.: Some methods for classification and analysis of multivariate observations. In: Proceedings of the 5-th Symposium on Math. Statist. and Probability. Berkeley. AD669871 Univ. of California Press, Berkeley. v.1 (1967) 281–297 25. Nguyen, H.S., Nguyen, S.H.: Some efficient algorithms for rough set methods. In: Proceedings Fifth Conference on Information Processing and Management of Uncertainty in Knowledge-Based Systems (IPMU’96), Granada, Spain, July (1996) 1451–1456 26. Nguyen, H.S., Nguyen, S.H.: Discretization Methods in Data Mining. In: L. Polkowski, A. Skowron (eds.): Rough Sets in Knowledge Discovery. Physica-Verlag, Heidelberg (1998) 451-482

396

J.J. Vald´es and A.J. Barton

27. Nguyen, H.S., Skowron, A.: Quantization of real-valued attributes. In: Proceedings Second International Joint Conference on Information Sciences, Wrightsville Beach, NC, September (1995) 34-37 28. Øhrn, A.: Discernibility and Rough Sets in Medicine: Tools and Applications. PhD thesis, Norwegian University of Science and Technology, Department of Computer and Information Science, December NTNU report 1999:133. [http://www.idi.ntnu.no/˜aleks/thesis/] (1999) 29. Øhrn, A.: Rosetta Technical Reference Manual. Department of Computer and Information Science, Norwegian University of Science and Technology, Trondheim, Norway (2001) 30. Øhrn, A., Komorowski, J.: Rosetta- A Rough Set Toolkit for the Analysis of Data. In: Proceedings of Third Int. Join Conf. on Information Sciences (JCIS97), Durham, NC, USA, March 1-5 (1997) 403-407 31. Pawlak, Z.: Rough sets: Theoretical aspects of reasoning about data. Kluwer Academic Publishers, Dordrecht, Netherlands (1991) 32. Peters, J.F., Borkowski, M.: K-means Indiscernibility Relation over Pixels, Fourth. Int. Conf. on Rough Sets and Current Trends in Computing RSCTC 2004. Tsumoto, Slowinski, Komorowki, Grzymala-Busse (Eds.) Lecture Notes in Computer Science (Lecture Notes in Artificial Intelligence Series) LNAI 3066, Springer-Verlag (2004) 580-585 33. Press, W.H., Flannery, B.P., Teukolsky, S.A., Vetterling, W.T.: Numerical Recipes in C, Cambridge University Press, New York (1986) 34. Press, W.H., Teukolsky, S.A., Vetterling, W.T., Flannery, B.P.: Numerical Recipes in C. The Art of Scientific Computing. Cambridge Univ. Press (1992) 35. Sammon, J.W.: A non-linear mapping for data structure analysis. IEEE Trans. on Computers C18 (1969) 401-409 36. Tannenbaum, T., Wright, D., Miller, K., Livny, M.: Condor – A Distributed Job Scheduler. In Thomas Sterling (Ed.) Beowulf Cluster Computing with Linux. MIT Press (2001) 37. Thain, D., Tannenbaum, T., Livny, M.: Condor and the Grid. In Fran Berman and Geoffrey Fox and Tony Hey (Eds.) Grid Computing: Making the Global Infrastructure a Reality. John Wiley & Sons (2002) 38. Thain, D., Tannenbaum, T., Livny, M.: Distributed Computing in Practice: The Condor Experience. Journal of Concurrency and Computation: Practice and Experience (2004) 39. Vald´es, J.J.: Similarity-Based Heterog