Lecture Notes in Computer Science Edited by G. Goos, J. Hartmanis, and J. van Leeuwen
2561
3
Berlin Heidelberg New York Barcelona Hong Kong London Milan Paris Tokyo
Harrie C.M. de Swart (Ed.)
Relational Methods in Computer Science 6th International Conference, RelMiCS 2001 and 1st Workshop of COST Action 274 TARSKI Oisterwijk, The Netherlands, October 16-21, 2001 Revised Papers
13
Series Editors Gerhard Goos, Karlsruhe University, Germany Juris Hartmanis, Cornell University, NY, USA Jan van Leeuwen, Utrecht University, The Netherlands Volume Editor Harrie C.M. de Swart Tilburg University P.O. Box 90153 5000 LE Tilburg The Netherlands E-mail:
[email protected] Cataloging-in-Publication Data applied for A catalog record for this book is available from the Library of Congress. Bibliographic information published by Die Deutsche Bibliothek Die Deutsche Bibliothek lists this publication in the Deutsche Nationalbibliografie; detailed bibliographic data is available in the Internet at http://dnb.ddb.de
CR Subject Classification (1998): F.4, I.1, I.2.3, D.2.4 ISSN 0302-9743 ISBN 3-540-00315-0 Springer-Verlag Berlin Heidelberg New York This work is subject to copyright. All rights are reserved, whether the whole or part of the material is concerned, specifically the rights of translation, reprinting, re-use of illustrations, recitation, broadcasting, reproduction on microfilms or in any other way, and storage in data banks. Duplication of this publication or parts thereof is permitted only under the provisions of the German Copyright Law of September 9, 1965, in its current version, and permission for use must always be obtained from Springer-Verlag. Violations are liable for prosecution under the German Copyright Law. Springer-Verlag Berlin Heidelberg New York a member of BertelsmannSpringer Science+Business Media GmbH http://www.springer.de © Springer-Verlag Berlin Heidelberg 2002 Printed in Germany Typesetting: Camera-ready by author, data conversion by PTP Berlin, Stefan Sossna e. K. Printed on acid-free paper SPIN: 10871819 06/3142 543210
Preface
This volume contains the papers presented at RelMiCS 2001, the 6th International Conference on Relational Methods in Computer Science, and the First Workshop of COST Action 274 TARSKI, Theory and Application of Relational Structures as Knowledge Instruments. The conference was held in conference centre Boschoord, Oisterwijk near Tilburg, The Netherlands, from October 16 till October 21, 2001. The conference attracted interest from many parts of the world with contributions from many countries. This conference was a continuation of international conferences/workshops on Relational Methods in Computer Science held in: Schloss Dagstuhl, Germany, January 1994; Parati near Rio de Janeiro, September 1995; Hammamet, Tunisia, January 1997; the Stefan Banach Center, Warsaw, September 1998; and Quebec, Canada, January 2000. The purpose of these conferences/workshops is to bring together researchers from various subdisciplines of Computer Science, Mathematics, and Philosophy, all of whom use relational methods as a conceptual and methodological tool in their work. Topics include, but are not limited to: relational, cylindric, fork, and Kleene algebras; relational proof theory and decidability issues; relational representation theorems; relational semantics; applications to programming, databases, and analysis of language; and computer systems for relational knowledge representation. With respect to applications one can think of: relational specifications and modeling; relational software design and development techniques; programming with relations; and implementing relational algebra. The RelMiCS’6 conference had three invited lectures, one of which has been included in these Proceedings. In addition, three tutorials were presented: Ivo D¨ untsch and G¨ unther Gediga, Rough’ Sets: tools for non-invasive data analysis; Wolfram Kahl and Eric Offermann, Programming with and in relational categories; Gheorghe S ¸ tef˘anescu, An introduction to network algebra. After a thorough refereeing process the Program Committee selected 21 papers for inclusion in these Proceedings. The papers have been classified into the different Work Areas (WAs) of Cost Action TARSKI: WA 1, Algebraic and logical foundations of “real-world” relations; WA 2, Mechanization of relational reasoning; WA 3, Relational scaling and preferences.
September 2002
Harrie de Swart
Organization
Program Committee Harrie de Swart (Chair): Ewa Orlowska: Gunther Schmidt: Chris Brink: Ivo D¨ untsch:
Tilburg University (The Netherlands) Institute of Telecommunications (Warsaw) Universit¨ at der Bundeswehr (M¨ unchen) University of Wollongong (Australia) Brock University (St. Catharines, Canada)
Invited Speakers Michel Grabisch: Wendy MacCaull:
Universit´e Pierre et Marie Curie (Paris) St. Francis Xavier University (Antigonish, Canada) Fakult¨ at f¨ ur Informatik (UniBw, M¨ unchen)
Wolfram Kahl:
Referees Philippe Balbiani Wojchiech Buszkowski Anthony Cohn Stephane Demri Thorsten Ehm Melvin Fitting G¨ unther Gediga Wolfram Kahl Joachim Lambek
Wendy MacCaull Larisa Maksimova Bernhard M¨ oller Eric Offermann Ewa Orlowska Jochen Pfalzgraf Anna Radzikowska Anna Romanowska Marc Roubens
Gunther Schmidt Bruce Spencer Harrie de Swart Dimiter Vakarelov Michael Winter Andrzej Wronski
Sponsoring Institutions Faculty of Philosophy of Tilburg University COST Action 274: TARSKI Royal Netherlands Academy of Arts and Sciences Dutch Research School in Logic TUE-UvT Research Group on Logic and Information Systems
Table of Contents
Invited Lecture A Relation-Algebraic Approach to Graph Structure Transformation . . . . . Wolfram Kahl
1
Contributed Papers WA 1: Algebraic and Logical Foundations of ‘Real World’ Relations Emptiness Relations in Property Systems . . . . . . . . . . . . . . . . . . . . . . . . . . . Philippe Balbiani
15
Pregroups: Models and Grammars . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Wojciech Buszkowski
35
Algebraic Semantics of ER-Models in the Context of the Calculus of Relations. II: Dynamic View. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Ernst-Erich Doberkat, Eugenio G. Omodeo
50
Interpretability of First-Order Dynamic Logic in a Relational Calculus . . . Marcelo F. Frias, Gabriel A. Baum, Thomas S.E. Maibaum
66
Relations in GUHA Style Data Mining . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Petr H´ ajek
81
Groups in Allegories . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Yasuo Kawahara
88
Distributed Conceptual Structures . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 104 Robert E. Kent A Computer Algebra Approach to Relational Systems Using Gr¨ obner Bases . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 124 L.M. Laita, E. Roanes-Lozano, L. de Ledesma, T. Calvo, L. Goz´ alez-Sotos Fuzzy Relational Images in Computer Science . . . . . . . . . . . . . . . . . . . . . . . . 134 Mike Nachtegael, Martine De Cock, Dietrich Van der Weken, Etienne E. Kerre A Completeness Theorem for Extended Order Dependencies on Relational Attribute Models in Dedekind Categories . . . . . . . . . . . . . . . . . . . 152 Hitomi Okuma, Yasuo Kawahara
X
Table of Contents
Double Residuated Lattices and Their Applications . . . . . . . . . . . . . . . . . . . 171 Ewa Orlowska, Anna Maria Radzikowska Interval Bilattices and Some Other Simple Bilattices . . . . . . . . . . . . . . . . . . 190 Agata Pilitowska Interactive Systems: From Folklore to Mathematics . . . . . . . . . . . . . . . . . . . . 197 Gheorghe S ¸ tef˘ anescu Relational Constructions in Goguen Categories . . . . . . . . . . . . . . . . . . . . . . . 212 Michael Winter A Subintuitionistic Logic and Some of Its Methods . . . . . . . . . . . . . . . . . . . 228 Ernst Zimmermann WA 2: Mechanisation of Relational Reasoning Implementation of Relational Algebra Using Binary Decision Diagrams . . 241 Rudolf Berghammer, Barbara Leoniuk, Ulf Milanese Calculating a Relational Program for Transitive Reductions of Strongly Connected Graphs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 258 Rudolf Berghammer, Thorsten Hoffmann Calculating Church-Rosser Proofs in Kleene Algebra . . . . . . . . . . . . . . . . . . 276 Georg Struth WA 3: Relational Scaling and Preferences On the Definition and Representation of a Ranking . . . . . . . . . . . . . . . . . . . 291 Kim Cao-Van, Bernard De Baets Tangent Circle Algebras . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 300 Ivo D¨ untsch, Marc Roubens Author Index . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 315
A Relation-Algebraic Approach to Graph Structure Transformation (Invited Talk
Extended
Abstract)
Wolfram Kahl Dep~rtment of Computing a,nd Software, McM~ster University kahl@cas
1
.mcmaster.
ca
Introduction
Graph transformation is a rapidly expanding field of research, motivated by a wide range of applications. Individual graph transformation steps can be specified at different levels of abstraction. On one end, there are ~programmed" graph transformation systems with very fine control over manipulation of individual graph items. Different flavours of rule-based graph transformation systems match their patterns in more generic ways and therefore interact with tile graph structure at slightly higher levels. A rather dominant approach to graph transformation uses the abstractions of category theory to define matching and replacement almost on a black-box level, using easily understandable basic category-theoretic concepts, like pushouts and pullbacks. However, some details cannot be covered on this level, and most authors refrain from resorting to the much more advanced category-theoretic tools of topos theory that, are available for graphs, too topos theory is not felt to be an appropriate language for specifying graph transformation. We show that the language of relations is better suited for this purpose and can be used very naturally to cover all tile problems of the categoric approach to graph transformation. Although much of this follows from the well-known fact that every graph-structure category is a topos, very little of this power has been exploited before, and even surveys of the categoric approach to graph transformation state essential conditions outside the category-theoretic framework. One achievement is therefore the capability to provide descriptions of all graph transformation effects on a suitable level of abstraction from tile concrete choice of graph structures. * Most of this work was completed during the author's appointment at Institute of Software Technology, Universit~it der Bundeswehr M/inchen. H. de Swart (Ed.): RelMiCS 2001, LNCS 2561, pp. 1-14, 2002. 9 Springer-VerlagBerlin Heidelberg 2002
2
W. Kahl
Another important result is the definition of a graph rewriting approach where relational matchings can match rule parameters to arbitrary subgraphs, which then can be copied or deleted by rewriting. At the same time, the rules are still as intuitive as in the double-pushout approach, and there is no need to use complicated encodings as in the pullback approaches. In short: A natural way to achieve a double-pushout-like rewriting concept that incorporates some kind of "graph variable" matching and replication is to amalgamate pushouts and pullbacks, and the relation-algebraic approach offers appropriate abstractions that allow to formalise this in a fully component-free yet intuitively accessible manner. This extended abstract is a short overview of the central results of [Kah01]. The next section contains a summary of the conventional approaches to describing graph transformation in the setting of categories, which are a generalisation of total functions between sets. Section 3 moves to the abstract setting of Dedekind categories, one particular abstraction from the setting of relations between sets. We first present natural relational characterisations of the central ingredients of the category theoretic approaches, and then use the flexibility of the relational approach to amalgamate the relational versions of pushouts and pullbacks into the original pullout approach to graph structure transformation. 2
Conventional
Categoric
Graph
Transformation
The so-called "algebraic approach to graph transformation" is really a collection of approaches that essentially rely on category-theoretic abstractions. 1 A category consists of a collection of objects and a collection of morphisms, where every morphism has a source and a target; we introduce a morphism F with source A and target B by writing F : A -+ B. In addition, for every object A there is an identity morphism I[A : A -+ A usually just written I[, and the associative composition operator combines two morphisms F : A -+ B and G : B -+ C into their composition F;G, which is a morphism from A to C.
2.1
Pushout Rewriting
Historically, Ehrig, Pfender, and Schneider developed the double-pushout approach as a way to generalise Chomsky grammars from strings to graphs, using pushouts as "gluing construction" to play the rdle of concatenation on strings [EPS73], see also [CMR+97]. The name "algebraic approach" derives from the fact that graphs can be considered as a special kind of algebras, and that the pushout in the appropriate category of algebras was perceived more as a concept from universal algebra than from category theory. 1 An accessible introduction to category theory in computing science is [BW90]. However, in the context of the algebraic approach to graph transformation, usually only a very limited amount of category theory is employed, and most of the literature is accessible starting for example from the handbook chapters [CMR+97,EHK+97].
A Relation-Algebraic Approach to Graph Structure Transformation
3
The pushout for a span (i.e., a pair of morphisms starting from a common source)/3 ~ p A Q ~ C in some category consists of a pushout object 19 and two morphisms R and S with l? as target, such t h a t the resulting square commutes, i.e., P ~ R - Q~S, and t h a t for every other completion/3 R' ~ l) I ~ S' C to a commuting square, i.e. with P~R t = Q~S t, there is a unique morphism Y : l? --+ l? t such t h a t the candidate morphisms can be factovised via Y, t h a t is, /~ = R~Y and S t = S~Y.
Q A
,
C
l? t The following example in the category of unlabelled directed graphs shows a pushout t h a t "glues" together two directed graphs along a common interface consisting of two nodes connected by a single directed edge (for anchoring the edge component of graph homomorphisms, we draw edges as small squares with incoming and outgoing source and target "tentacles"). The pushout object can be considered as containing "copies" of the two g r a p h s / 3 and C, where the two "instances" of the "interface" A have been "identified".
In the double-pushout approach, a rewriting rule is a span s :#L ~ #R~ 7-4 of morphisms starting from the gluing object ~. As an example consider the following graph rewriting rule, which deletes a path of length two between the two interface nodes, and inserts a four-node cycle with one edge connecting the interface nodes.
4
W. Kahl
A redex for such a rule is a morphism XL : /2 --+ A from the rule's left-hand side /2 into some application graph A. In the following example, the redex is injective, although, under certain circumstances, non-injective redex morphisms are allowed, too.
PhiL
PhiR
Application of the rule has to establish a double-pushout diagram of the following shape:
XL
A
XR
.
g'L
~
g'R
,
B
Note that for the left-hand side pushout, the "wrong" arrows are given, so the completion to a pushout square is not a universally characterised categorical construction. This completion is called a pushout complement, which consists of a host object ~ and a host morphism ~ : ~ -+ ~ such that the pushout of the rule's left-hand side @L and the host morphism ~ recovers the application graph. Such a pushout complement exists iff the so-called gluing condition holds for items from left-hand side outside the image of the interface, the redex must not perform any identification or mapping to nodes incident with "dangling edges". This gluing condition is usually stated using the concrete language of graphs with nodes and edges; the only abstract formalisation given so far is that of Kawahara [Kaw901, using the language of toposes (categories with "internal logic"), which may be the reason why this seems not to have caught on.
A Relation-Algebraic Approach to Graph Structure Transformation
5
Once a host morphism is found, the right-hand-side pushout completes the rewriting step, which in our example then looks as follows"
X/
Note that in contrast with the first pushout example, there is no edge in the interface graph here; this produces a pushout object 13 that has two edges between the images of the interface nodes only the interface nodes themselves are forced to be identified here. As this example illustrates, application of double-pushout rules is quite easy to understand. This makes the double-pushout approach a useful tool also for specifyin9 all kinds of transition and transformation systems, with applications ranging from applied software systems and systems modelling to aspects of parallelism, distributed systems and synchronisation mechanisms, and operational semantics of programming languages from a wide range of paradigms, including functional, logic, and object-oriented. However, pushout rewriting does not easily lend itself to extensions that include some kind of 9raph variable that could be matched to whole subgraphs, such that rewriting might replicate or delete these subgraphs.
2.2
Pullback Rewriting
Bauderon proposed a different categorical setting for graph rewriting that overcomes the lack of replication capabilities in the pushout approaches. Starting from the fact that the most natural replication mechanism in category theory is the categorical product, and that, in graph-structure categories, pullbacks are subobjects of products, Bauderon and Jacquet introduced a setup that uses pullbacks in place of pushouts [Bau97,BJ01]. The pullback is the dual concept of the pushout, that is, defined with "all arrows turned around": Starting from two morphisms R and S with common target 7?, a pullback consists of a pullback object A and two morphisms P and Q starting from A such that the resulting square commutes, and such that each commuting candidate square can be uniquely factorised via A.
6
W. Kahl R
B
C
A' The following example shows how a simple single-pullback rule R can be used to specify duplication of application graphs.
In general, however, pullback rules tend to be less intuitive. The natural approach is to consider different parts of the target of the rule arrow R as standing for different "rdles" in the rewriting step. The rule morphism R decides the behaviour of each rdle, while the morphism S from the application graph into 7? assigns rdles to all parts of the application graph. For maximum flexibility in the treatment of these rdles, Bauderon and Jacquet therefore provide an alphabet graph that contains items that each represent a different treatment via rewriting, such as preservation, duplication, or deletion. The following shows the translation of a simple NLC rule ("node-labelcontrolled graph rewriting", one variant of vertex replacement [ER97]), where Z? is the alphabet graph for three labels, which are represented by the horizontally flattened three-node clique in the middle. The top node is the "context", and the bottom node is the "redex". The rule morphism R rewrites the redex by splitting it into two nodes connected by a single edge, redirecting incoming edges from one of the three kinds of neighbours to the source of that new edge, and redirecting the source tentacles of outgoing edges directed at the same kind of neighbours to the target of the new edge.
A Relation-Algebraic Approach to Graph Structure Transformation
7
Context:
Labels:
Redex: i ~
~
i ~ ~i i ii
Such a rule is applied by establishing what Bauderon and Jacquet call a "label" on an application graph, that is, a graph morphism from the application graph to the alphabet graph such that exactly one node is mapped to the "redex" node, precisely its neighbours are mapped to label nodes in the alphabet graph (which are the neighbours of the redex node), and all other nodes are mapped to the "context" node. Since just the rule above even sufficiently demonstrates that pullback rules are much less intuitive than pushout rules, we refrain from showing an application of this rule. Even though the expressivity of pullback rewriting encompasses most wellknown approaches to graph rewriting, including node replacement, (hyper-)edge replacement, and pushout rewriting, the lacking intuitive accessibility of its rewriting concept has severely limited interest in the pullback approach.
3
A m a l g a m a t i n g P u s h o u t s and P u l l b a c k s to P u l l o u t s
We are going to present a formalism that allows the abstract specification of graph transformation with "graph variables" and replication of their images. It will superficially be modelled at the double-pushout approach, but works in the setting of relational graph morphisms, and it fully incorporates both the pushout and the pullback approaches. For details, see [Kah01]. The key to the relation-algebraic approach to graph transformation is the fact the graph structures with relational graph-structure hornornorphisrns give rise to Dedekind categories. Dedekind categories are relational categories that differ from heterogeneous relation algebras only in that the lattice of relations between two objects need not be Boolean. Therefore, most of the laws of relation algebras that do not involve complement still hold. An even weaker, but in many circumstances still useful structure is an allegory [FS90], which is essentially a category with converse and intersection. Dedekind categories are also known as "locally complete distributive allegories". We call morphisms of allegories and Dedekind categories relations, and we introduce a relation R from object A to object 13 by writing R : A ++ 13; and for relations R, S : A ++ 13 we use converse R- (which is a relation from
8
W. Kahl
13 to A), union R u S, intersection R ~ S; the domain of definition of R is dom R = lrA ~ R;R~, and the range is ran R = dom (R~). For a homogeneous relation Q : A ++ A, we write Q* for the reflexive and transitive closure of Q. We consider a graph to be a unary algebra (Af, E, s,t) consisting of a node set Af, an edge set g, and two mappings s, t : g --4 Af which map edges to their source resp. target nodes. Given two graphs ~1 = (Afl,gl,Sl,tl) and ~2 (A/'2, g2, s2, t2), a relational graph homomorphism R : ~ ++ ~2 is a pair (RAt, Rc) of relations Rz; : H I ++ A/2 and Re : gl ++ g2 such that the following relational homomorphism condition holds: R$;82 E 81;RAf
Re;t2 E tl;-]~.N"
and
This definition is easily generalised to arbitrary many-sorted algebras. In general, algebras with such relational homomorphisms give rise to allegories, while unary algebras, also called graph structures, give rise to Dedekind categories. The following example of a relational graph homomorphism is neither total nor univalent:
In a Dedekind category of relational graph structure homomorphisms, conventional graph structure homomorphisms are recovered as the mappings in these Dedekind categories, that is, as total and univalent relational graph structure homomorphisms. Pushouts and pullbacks of mappings in Dedekind categories have nice relational characterisations; there are even useful generalisations that correspond to starting from the diagonals U := P - ; Q resp. V := R;S-. R
A
13
C
Dt
~4t
A Relation-Algebraic Approach to Graph Structure Transformation
9
The generalisation of pullbacks is called tabulation by Freyd and Scedrov [FS90, 2.14] and is defined already in allegories; we slightly rearrange the conditions" D e f i n i t i o n 3.1 In an allegory D, let V 9 13 ++ C be an arbitrary relation. The span 13 ~ P A Q ~ C in the allegory D (i.e., P and Q are not yet specified as mappings) is a direct tabulation for V iff the following equations hold"
P
;Q- v
P-;P - dom V
Q~;Q
-
-
ran V
P;P
n Q;Q
-
D
For pushouts, the following conditions have first been stated by K a w a h a r a [Kaw901, working directly with P~;Q instead of with U" D e f i n i t i o n 3.2 In a Dedekind category D, let U 9 13 ++ C be an a r b i t r a r y relation. The cospan 13 R 7? ~ s C in the Dedekind category D is a direct gluing for U iff the following equations hold"
R;s
-
A t a b u l a t i o n for a universal relation is a direct product, and a gluing for an e m p t y relation is a direct sum. Although the usual relational characterisation of direct sums is dual to t h a t of direct products, the gluing conditions cannot be obtained from the t a b u l a t i o n conditions via na'ive relational dualisation. T h e reason for this is t h a t for a cospan 13 R 7? ~ s C of mappings, the relation R;S~ is always difunctional (and this difunctionality is recovered via U;(U~;U)*, which is the difunctional closure of U), while for a span B~ P relation P~;Q can essentially be arbitrary.
A
Q~ C the
For our relational rewriting concept we are going to use a construction t h a t can be u n d e r s t o o d as a m a l g a m a t e d from a p u s h o u t and a pullback. This construction starts with a span 7~ ~ ~ ~ ~ of two relational morphisms, and in addition a partial identity u0 _E lrG designating the interface component of the gluing object ~. From this interface component, one may calculate the variable (or parameter) component vo as the smallest partial identity such t h a t u0 U v0 - lrG; we call this a semi-complement. An intuitive explanation of the a m a l g a m a t e d construction is to start with producing a p u s h o u t for the interface parts u0;q5 and u0;~, and a pullback for the converses of the p a r a m e t e r parts, i.e., for qS-;v0 and ~-;u0. T h e n the two result objects have to be glued together along a gluing relation B induced by the restrictions of the p u s h o u t and pullback morphisms to the borders between interface and p a r a m e t e r parts.
10
W. Kahl U0
c2
PO
C3
no V1Vo
q
9
v2
C1
PB
B
9
V0
Vl
V3
gluing
G3 One m a y observe t h a t with this c o n s t r u c t i o n , m o s t of t h e details a b o u t @ a n d are c o m p l e t e l y irrelevant, a n d we only need access to t h e resulting p a r a m e t e r a n d interface parts. In the spirit of t h e definitions of t a b u l a t i o n s a n d gluings, we therefore can a b s t r a c t away from @ a n d ~ , a n d s u b s t i t u t e U "-- ~5~';UO;~ ,
V " - ~5"-';vo "" .
W i t h a few interface preservation conditions t h a t we leave out here, this s e t u p allows to define t h a t a cospan G1 x ~ G3 ~ e G2 is called a glued tabulation f o r V along U on Cl and c2 (where Cl a n d c2 are p a r t i a l identities comprising t h e respective interface p a r t s a n d contexts) iff t h e following conditions hold: =
X;X'-"
=
*
u
c1; ( U ; U ~ ) * ;c1
II
k~;k~
=
c2;(U~;U)* ;c2
II
~.
=
( X ' - ' ; c 1 ; X I I ~t~;C2;~t )
II
v
dom V ran V (X'-';Vl;X
~ ~;V2;~)
0
S t a r t i n g with a span Tr ~ e ~ ~ ~ 7-{ a n d a p a r t i a l i d e n t i t y u0 _E 1[G d e s i g n a t i n g t h e interface c o m p o n e n t , a glued t a b u l a t i o n for V along U as defined above is called a pullout f o r ~ ~ ~
(~, uo)
~ ~ 7-{.
Obviously, t h e pullout conditions are an a m a l g a m a t i o n of t h e gluing a n d tabulation conditions. In addition, t h e c o n s t r u c t i o n sketched above satisfies these conditions. T h e ease with which this a m a l g a m a t i o n is possible is t h e essential a d v a n t a g e of t h e relational approach.
A Relation-Algebraic Approach to Graph Structure Transformation
11
Once we have pullouts, we naturally obtain a double-pullout approach to rewriting. A double-pullout rule consists of left- and right-hand side morphisms as usual, but these are now relational morphisms; and the rule has an additional component: a partial identity u0 on the gluing graph that indicates the interface. The following is an example rule, where the interface consists only of the top-most node:
i
"~---~PhiL'~'-"-"~"
hiR
Everything besides the interface is considered as parameter, and we see that the right-hand side of the example rule duplicates the parameter part. A redex is a relational graph morphism, too, and is restricted to be total and univalent besides the parameter part. (For simplicity, we ignore further more technical conditions.) It may map the parameter part to a larger subgraph, as in the following example, where the redex relation Xc is univalent only on the interface part:
The host construction can easily be expressed using relation-algebraic constructions. Here, the rule's left-hand side ~5c is univalent on the parameter part, which allows us to use essentially the same host construction that was identified by Kawahara [Kaw90] for the double-pushout approach. (If the rule's left-hand side ~5c is not univalent on the parameter part, then the rule can be applied only if the redex documents appropriate replication of the respective parameter parts,
12
W. Kahl
which is equivalent with the existence of a certain partial equivalence relation O on the application object A. In such cases a non-trivial pullback complement has to be constructed for the parameter part; this becomes easy once O has been identified.)
)
For the result, the parameter part is replicated according to the prescription of the rule's right-hand side:
This concludes the application of the double-pullout rule. Although we have oriented our nomenclature and the direction of the morphisms at the doublepushout approach, it should be obvious that this choice is really arbitrary since
A Relation-Algebraic Approach to Graph Structure Transformation
13
the double-pullback approach is embedded in our approach to an equal extent; the pullback component made the replication of the parameter possible while the pushout component guaranteed the rigid matching of the interface component.
In the next picture we show an example how one may use a single rule to achieve several effects at once. This time, we use directed hypergraphs: besides nodes (big octagons) and edges (big squares), we now also have s o u r c e - a n d targettentacles (small upwards and downwards triangles) going from nodes to edges and vice versa. Starting from the left-hand side of the rule, we see that the upper and lower "ridges" are preserved, but the vertical "bridge" in the middle of the left-hand side is deleted. Furthermore, the left and right parts of the left-hand side each contain a unit hypergraph and act as parameters; the left parameter is deleted. In the right-hand side, we see that the two nodes the left parameter was incident with are identified. Furthermore, the right parameter is duplicated, and the "new copy" is connected with the "original copy" via two new edges. In the example application of this rule, we only show the matching and result morphisms, and omit the host part for the sake of readability.
ii~i~ii~il ~ii; ~III~IIII ii
14
W. Kahl
4
Conclusion
We have introduced an abstract approach to graph structure transformation that is based on the language of relations and leads to a rewriting concept where rewriting rules can contain parameters as well as "rigid" parts, and can specify not only addition and deletion of the rigidly mapped parts, but also deletion or replication of the parts matched to parameters, thus giving rise to a useful "graph variable" concept. The important point of the presented work is that we are able to specify this kind of rewriting with fully abstract definitions that never need to mention edges or nodes. Nevertheless, we obtain a rewriting concept that unifies the intuitive elegance of the pushout approaches with the expressive and replicative power of the pullback approaches. This is made possible by working in a relation-algebraic setting, which, we claim, is the right level of abstraction for specifying graphstructure transformation.
References [Bau97]
Michel Bauderon. A uniform approach to graph rewriting: The pullback approach. In Manfred Nagl, editor, Graph Theoretic Concepts in Computer Science, WG '96, volume 1017 of LNCS, pages 101-115. Springer, 1997. [BJ01] Michel Bauderon and H61~ne Jacquet. Pullback as a generic graph rewriting mechanism. Applied Categorical Structures, 9(1):65-82, 2001. [BW90] Michael Barr and Charles Wells. Category Theory for Computing Science. Prentice Hall International Series in Computer Science. Prentice Hall, 1990. [CMR+97] Andrea Corradini, Ugo Montanari, Francesca Rossi, Hartmut Ehrig, Reiko Heckel, and Michael LSwe. Algebraic approaches to graph transformation, part I: Basic concepts and double pushout approach. In Rozenberg [Roz97], chapter 3, pages 163-245. [EHK+97] Hartmut Ehrig, Reiko Heckel, Martin Korff, Michael L5we, Leila Ribeiro, Annika Wagner, and Andrea Corradini. Algebraic approaches to graph transformation, part II: Single pushout approach and comparison with double pushout approach. In Rozenberg [Roz97], chapter 4, pages 247-312. [EPS73] Hartmut Ehrig, M. Pfender, and H. J. Schneider. Graph grammars: An algebraic approach. In Proc. IEEE Conf. on Automata and Switching Theory, SWAT '73, pages 167-180, 1973. [ER97] Joost Engelfriet and Grzegorz Rozenberg. Node replacement graph grammars. In Rozenberg [Roz97], chapter 1, pages 1-94. [FS90] Peter J. Freyd and Andre Scedrov. Categories, Allegories, volume 39 of North-Holland Mathematical Library. North-Holland, Amsterdam, 1990. [Kah01] Wolfram Kahl. A relation-algebraic approach to graph structure transformation, 2001. Habil. Thesis, Fakultgt ffir Informatik, Univ. der Bundeswehr Mfinchen, Techn. Bericht 2002-03. [Kaw90] Yasuo Kawahara. Pushout-complements and basic concepts of grammars in toposes. Theoretical Computer Science, 77:267-289, 1990. [Roz97] Grzegorz Rozenberg, editor. Handbook of Graph Grammars and Computing by Graph Transformation, Vol. 1: Foundations. World Scientific, Singapore, 1997.
Emptiness Relations in Property Systems Philippe Balbiani Institut de recherche en informatique de Toulouse 118 route de Narbonne, 31062 Toulouse Cedex 4, France
Abstract. In this paper, we carry out the modal analysis of emptiness through a modal logic which modalities correspond to emptiness relations in property systems. We mainly address the questions of axiomatization/completeness and decidability/complexity of our modal logic.
1
Introduction
The concept of informational relation has been introduced by Pawlak [10] within the context of attribute systems — knowledge-based systems which describe objects in terms of attributes — and Vakarelov [16] within the context of property systems — knowledge-based systems which describe objects in terms of properties. It has been furthered by Demri [2], Demri, Orlowska and Vakarelov [3], Orlowska [6,7], Orlowska and Pawlak [9] and Vakarelov [13,14,15, 16] who have considered the questions of axiomatization/completeness and decidability/complexity of modal logics which modalities correspond to informational relations. The aim of these logics is to provide a formal method for reasoning about uncertain knowledge discovered from attribute systems or property systems. All the informational relations in attribute systems and property systems are either indistinguishability relations or distinguishability relations. Indistinguishability relations indicate the way objects share attributes or properties whereas distinguishability relations indicate the way attributes and properties differentiate objects. This paper is devoted to the modal analysis of emptiness relations in property systems. We give in section 2 the formal definition of emptiness relations in property systems. We provide a mathematical definition of the syntax and semantics of our modal logic in section 3 where we mainly address the questions of completeness and decidability. In section 4 we show that the satisfiability problem for any given formula of our modal logic requires polynomial space whereas in section 5 we show that the satisfiability problem for any given formula of our modal logic is decidable in polynomial space.
2
Emptiness Relations
Adapted from Vakarelov [16], a property system will be any algebraic structure (Obj, P ro, f ) where: H. de Swart (Ed.): RelMiCS 2001, LNCS 2561, pp. 15–34, 2002. c Springer-Verlag Berlin Heidelberg 2002
16
P. Balbiani
– Obj is a nonempty set of objects; – P ro is a nonempty set of properties; – f is a function with domain Obj and range the power set of P ro. The property system (Obj, P ro, f ) will be defined to be non-trivial if for all x ∈ Obj, f (x) = ∅ and f (x) = P ro. We should consider, for example, the property system S = (Obj, P ro, f ) defined as follows. Define: - Obj is {Ann, Bob, Cindy, Daniel, Emma}; - P ro is {Arabic, Bulgarian, Castilian, Dutch, English}; - f is the function defined by table 1.
Table 1. Example of a property system. Ann {Arabic, Bulgarian, Castilian}
Bob {Arabic, Bulgarian, Castilian, Dutch}
Cindy {Arabic, Bulgarian}
Daniel {English}
Emma {Castilian, Dutch, English}
Property systems constitute the starting point for the formal examination of sentences of the form “object x is indistinguishable from object y” or sentences of the form “object x is distinguishable from object y”. In this respect, the emptiness relations might play an important role. Given any property system S = (Obj, P ro, f ), the emptiness relations over S are defined by the following universal conditions: x S y iff f (x) ∩ f (y) = ∅; x S y iff f (x) ∩ f (y) = ∅;
x S y iff f (x) ∩ f (y) = ∅; x S y iff f (x) ∩ f (y) = ∅.
Within the context of attribute systems, Orlowska [8] has named “forward inclusion”, “backward inclusion”, “right orthogonality”, “left orthogonality” our emptiness relations S , S , S , S . The property system of table 1 is such that Ann S Bob, Ann S Cindy, Ann S Daniel and Ann S Emma. We leave it to the reader to prove the following lemma. Lemma 1. Let S = (Obj, P ro, f ) be a non-trivial property system. The following universal conditions are satisfied. x S x; x S x; If x S y then y S x; If x S y then y S x; If x S y and y S z then x S z; If x S y and y S z then x S z; If x S y and y S z then x S z; If x S y and y S z then x S z; xS x; xS x; If x S y then y S x; If x S y then y S x; If x S y and y S z then x S z; If x S y and y S z then x S z; If x S y and y S z then x S z; If x S y and y S z then x S z.
Emptiness Relations in Property Systems
17
Lemma 1 motivates the following definition. An emptiness structure will be any relational structure (W, , , , ) where: – W is a nonempty set of possible worlds; – , , , are binary relations on W subject to the universal conditions of lemma 1. The following theorem explains the connection between emptiness structures and non-trivial property systems. Theorem 1. (Characterization theorem for emptiness structures) Let (W, , , , ) be an emptiness structure. There is a non-trivial property system S = (Obj, P ro, f ) such that Obj = W and the following universal conditions are satisfied. x S y iff x y; x S y iff x y;
x S y iff x y; x S y iff x y.
Proof. It follows immediately from the definition that the following universal conditions are satisfied: x x; x x; If x y then y x; If x y then y x; If x y and y z then x z; If x y and y z then x z; If xy then xx; If xy then xx; xx or x y; xx or x y; If xy then yx; If xy then yx; If xy and y z then xz; If xy and y z then xz; xy or yz or x z; xy or yz or x z. By Vakarelov [16], there is a property system S = (Obj, P ro, f ) such that Obj = W and the following universal conditions are satisfied: x S y iff x y; x S y iff x y;
x S y iff x y; x S y iff x y.
In view of the fact that for all x ∈ W , xx and xx, for all x ∈ Obj, f (x) = ∅ and f (x) = P ro. Consequently S is non-trivial.
3
Modal Analysis
Let us be clear that our emptiness relations are nothing but the relations of informational inclusion and the complementary relations of the relations of similarity introduced by Vakarelov [16]. In this section, their modal analysis is presented for the very first time. An abstract structure will be any structure (W, , , , , ≡, R) where: – W is a nonempty set of possible worlds; – , , , are binary relations on W subject to the universal conditions of lemma 1;
18
P. Balbiani
– ≡, R are binary relations on W subject to the following universal conditions: • x ≡ y iff x y and x y; • xRy iff x y and x y. The proof of the following lemma is left as an exercise for the reader. Lemma 2. Let (W, , , , , ≡, R) be an abstract structure. The following universal conditions are satisfied. If x ≡ y then x y; If xRy then x y; If x ≡ y then x y; If xRy then x y; x ≡ x; If x ≡ y then y ≡ x; If xRy then yRx; If x ≡ y and y ≡ z then x ≡ z; If xRy and y ≡ z then xRz; If x ≡ y and yRz then xRz; If xRy and yRz then x ≡ z. Lemma 2 motivates the following definition. A nonstandard abstract structure will be any structure (W, , , , , ≡, R) where: – W is a nonempty set of possible worlds; – , , , are binary relations on W subject to the universal conditions of lemma 1 but the universal conditions “xx”, “xx”; – ≡, R are binary relations on W subject to the universal conditions of lemma 2. The linguistic basis of our modal logic is the propositional calculus enlarged with the modalities [], [], [], [], [≡], [R]. We define the set of all formulas as follows: – A ::= p | ¬A | (A ∨ B) | []A | []A | []A | []A | [≡]A | [R]A; where p ranges over a countably infinite set of propositional variables. The other standard connectives are defined by the usual abbreviations. In particular, A is ¬[]¬A, A is ¬[]¬A, A is ¬[]¬A, A is ¬[]¬A, ≡ A is ¬[≡]¬A, R A is ¬[R]¬A. We follow the standard rules for omission of the parentheses. The number of occurrences of symbols in formula A is denoted length(A). A model (respectively: a nonstandard model) will be any structure (W, , , , , ≡, R, V ) where: – (W, , , , , ≡, R) is an abstract structure (respectively: a nonstandard abstract structure); – V is a function with domain the set of all propositional variables and range the power set of W . Let M = (W, , , , , ≡, R, V ) be either a model or a nonstandard model. We define the relation “formula A is true at possible world x in M ”, denoted M, x |= A, as follows: -
M, x |= p iff x ∈ V (p); M, x |= ¬A iff M, x |= A; M, x |= A ∨ B iff M, x |= A or M, x |= B; M, x |= []A iff for all y ∈ W , if x y then M, y |= A;
Emptiness Relations in Property Systems
-
19
M, x |= []A iff for all y ∈ W , if x y then M, y |= A; M, x |= []A iff for all y ∈ W , if x y then M, y |= A; M, x |= []A iff for all y ∈ W , if x y then M, y |= A; M, x |= [≡]A iff for all y ∈ W , if x ≡ y then M, y |= A; M, x |= [R]A iff for all y ∈ W , if xRy then M, y |= A.
Let LER — logic of emptiness relations — be the smallest normal modal logic that contains the following axioms: []A → A; A → [] A; []A → [][]A; []A → [][]A; A → [] A; []A → [][]A; []A → [][]A; []A → [≡]A; []A → [≡]A; [≡]A → A; A → [≡]≡ A; [≡]A → [≡][≡]A; [R]A → [≡][R]A;
[]A → A; A → [] A; []A → [][]A; []A → [][]A; A → [] A; []A → [][]A; []A → [][]A; []A → [R]A; []A → [R]A; A → [R]R A; [R]A → [R][≡]A; [≡]A → [R][R]A.
A typical result is the following. Theorem 2. (Completeness theorem for LER) LER is complete with respect to the class of all models, the class of all finite nonstandard models and the class of all nonstandard models, i.e. the following conditions are equivalent. 1. 2. 3. 4.
A A A A
is is is is
true at all possible worlds in all models; true at all possible worlds in all finite nonstandard models; true at all possible worlds in all nonstandard models; a theorem of LER.
Proof. (1 implies 2): Let M = (W, , , , , ≡, R, V ) be a finite nonstandard model and M = (W , , , , , ≡ , R , V ) be the structure defined as follows. Define: – W is W × ZZ × ZZ ; – For all x, y ∈ W , for all i1 , i2 ∈ ZZ and for all j1 , j2 ∈ ZZ , (x, i1 , i2 ) (y, j1 , j2 ) iff x y, i1 = j1 , i2 j2 and if i2 j2 then x ≡ y; – For all x, y ∈ W , for all i1 , i2 ∈ ZZ and for all j1 , j2 ∈ ZZ , (x, i1 , i2 ) (y, j1 , j2 ) iff x y, i1 = j1 , i2 j2 and if i2 j2 then x ≡ y; – For all x, y ∈ W , for all i1 , i2 ∈ ZZ and for all j1 , j2 ∈ ZZ , (x, i1 , i2 ) (y, j1 , j2 ) iff x y, i1 = −j1 , i2 −j2 and if i2 −j2 then xRy; – For all x, y ∈ W , for all i1 , i2 ∈ ZZ and for all j1 , j2 ∈ ZZ , (x, i1 , i2 ) (y, j1 , j2 ) iff x y, i1 = −j1 , i2 −j2 and if i2 −j2 then xRy; – For all x, y ∈ W , for all i1 , i2 ∈ ZZ and for all j1 , j2 ∈ ZZ , (x, i1 , i2 ) ≡ (y, j1 , j2 ) iff x ≡ y, i1 = j1 and i2 = j2 ;
20
P. Balbiani
– For all x, y ∈ W , for all i1 , i2 ∈ ZZ and for all j1 , j2 ∈ ZZ , (x, i1 , i2 )R (y, j1 , j2 ) iff xRy, i1 = −j1 and i2 = −j2 ; – For all propositional variables p, V (p) is V (p) × ZZ × ZZ . It is a simple matter to check that M is a model. Moreover, M is a p-morphic image of M , i.e., the following conditions are satisfied: – For all x, y ∈ W , for all i1 , i2 ∈ ZZ (y, j1 , j2 ) then x y; – For all x, y ∈ W , for all i1 , i2 ∈ ZZ (y, j1 , j2 ) then x y; – For all x, y ∈ W , for all i1 , i2 ∈ ZZ (y, j1 , j2 ) then x y; – For all x, y ∈ W , for all i1 , i2 ∈ ZZ (y, j1 , j2 ) then x y; – For all x, y ∈ W , for all i1 , i2 ∈ ZZ (y, j1 , j2 ) then x ≡ y; – For all x, y ∈ W , for all i1 , i2 (x, i1 , i2 )R (y, j1 , j2 ) then xRy;
and for all j1 , j2 ∈ ZZ , if (x, i1 , i2 ) and for all j1 , j2 ∈ ZZ , if (x, i1 , i2 ) and for all j1 , j2 ∈ ZZ , if (x, i1 , i2 ) and for all j1 , j2 ∈ ZZ , if (x, i1 , i2 ) and for all j1 , j2 ∈ ZZ , if (x, i1 , i2 ) ≡ ∈ ZZ and for all j1 , j2 ∈ ZZ , if
and the following conditions are satisfied: – For all x, y ∈ W and for all i1 , i2 ∈ ZZ , there is j1 , j2 ∈ ZZ such that if x y then (x, i1 , i2 ) (y, j1 , j2 ); – For all x, y ∈ W and for all i1 , i2 ∈ ZZ , there is j1 , j2 ∈ ZZ such that if x y then (x, i1 , i2 ) (y, j1 , j2 ); – For all x, y ∈ W and for all i1 , i2 ∈ ZZ , there is j1 , j2 ∈ ZZ such that if x y then (x, i1 , i2 ) (y, j1 , j2 ); – For all x, y ∈ W and for all i1 , i2 ∈ ZZ , there is j1 , j2 ∈ ZZ such that if x y then (x, i1 , i2 ) (y, j1 , j2 ); – For all x, y ∈ W and for all i1 , i2 ∈ ZZ , there is j1 , j2 ∈ ZZ such that if x ≡ y then (x, i1 , i2 ) ≡ (y, j1 , j2 ); – For all x, y ∈ W and for all i1 , i2 ∈ ZZ , there is j1 , j2 ∈ ZZ such that if xRy then (x, i1 , i2 )R (y, j1 , j2 ). As a consequence, see Hughes and Cresswell [4], if A is true at some possible world in M then A is true at some possible world in M . (2 implies 3): Let M = (W, , , , , ≡, R, V ) be a nonstandard model and M = (W , , , , , ≡ , R , V ) be the structure defined as follows. We define the smallest set ΓA of formulas such that: - A ∈ ΓA ; - ΓA is closed under subformulas. It should be remarked that Card(ΓA ) ≤ length(A). Let =A be the equivalence relation on W defined as follows. For all x, y ∈ W , define: – x =A y iff for all formulas B, if B ∈ ΓA then M, x |= B iff M, y |= B. For all x ∈ W , the equivalence class of x modulo =A is denoted x|=A . The quotient set of W modulo =A is denoted W|=A . For all propositional variables p, the quotient set of V (p) modulo =A is denoted by V (p)|=A . Define:
Emptiness Relations in Property Systems
– W is W|=A ; – For all x, y ∈ W , x|=A y|=A iff for all formulas B, if B ∈ ΓA then: • If M, x |= []B then M, y |= []B; • If M, x |= []B then M, y |= []B; • If M, y |= []B then M, x |= []B; • If M, y |= []B then M, x |= []B; – For all x, y ∈ W , x|=A y|=A iff for all formulas B, if B ∈ ΓA then: • If M, x |= []B then M, y |= []B; • If M, x |= []B then M, y |= []B; • If M, y |= []B then M, x |= []B; • If M, y |= []B then M, x |= []B; – For all x, y ∈ W , x|=A y|=A iff for all formulas B, if B ∈ ΓA then: • If M, x |= []B then M, y |= []B; • If M, x |= []B then M, y |= []B; • If M, y |= []B then M, x |= []B; • If M, y |= []B then M, x |= []B; – For all x, y ∈ W , x|=A y|=A iff for all formulas B, if B ∈ ΓA then: • If M, x |= []B then M, y |= []B; • If M, x |= []B then M, y |= []B; • If M, y |= []B then M, x |= []B; • If M, y |= []B then M, x |= []B; – For all x, y ∈ W , x|=A ≡ y|=A iff for all formulas B, if B ∈ ΓA then: • If M, x |= []B then M, y |= []B; • If M, x |= []B then M, y |= []B; • If M, x |= []B then M, y |= []B; • If M, x |= []B then M, y |= []B; • If M, x |= [≡]B then M, y |= [≡]B; • If M, x |= [R]B then M, y |= [R]B; • If M, y |= []B then M, x |= []B; • If M, y |= []B then M, x |= []B; • If M, y |= []B then M, x |= []B; • If M, y |= []B then M, x |= []B; • If M, y |= [≡]B then M, x |= [≡]B; • If M, y |= [R]B then M, x |= [R]B; – For all x, y ∈ W , x|=A R y|=A iff for all formulas B, if B ∈ ΓA then: • If M, x |= []B then M, y |= []B; • If M, x |= []B then M, y |= []B; • If M, x |= []B then M, y |= []B; • If M, x |= []B then M, y |= []B; • If M, x |= [≡]B then M, y |= [R]B; • If M, x |= [R]B then M, y |= [≡]B; • If M, y |= []B then M, x |= []B; • If M, y |= []B then M, x |= []B; • If M, y |= []B then M, x |= []B; • If M, y |= []B then M, x |= []B; • If M, y |= [≡]B then M, x |= [R]B; • If M, y |= [R]B then M, x |= [≡]B; – For all propositional variables p, V (p) is V (p)|=A .
21
22
P. Balbiani
It is a rather remarkable fact that M is a finite nonstandard model. The interesting result is that M is a filtration of M , i.e., the following conditions are satisfied: – – – – – –
For For For For For For
all all all all all all
x, y x, y x, y x, y x, y x, y
∈ W, ∈ W, ∈ W, ∈ W, ∈ W, ∈ W,
if if if if if if
x y then x|=A y|=A ; x y then x|=A y|=A ; x y then x|=A y|=A ; x y then x|=A y|=A ; x ≡ y then x|=A ≡ y|=A ; xRy then x|=A R y|=A ;
and the following conditions are satisfied: – For all x, y ∈ W and for all formulas B, if x|=A y|=A , B ∈ ΓA and M, x |= []B then M, y |= B; – For all x, y ∈ W and for all formulas B, if x|=A y|=A , B ∈ ΓA and M, x |= []B then M, y |= B; – For all x, y ∈ W and for all formulas B, if x|=A y|=A , B ∈ ΓA and M, x |= []B then M, y |= B; – For all x, y ∈ W and for all formulas B, if x|=A y|=A , B ∈ ΓA and M, x |= []B then M, y |= B; – For all x, y ∈ W and for all formulas B, if x|=A ≡ y|=A , B ∈ ΓA and M, x |= [≡]B then M, y |= B; – For all x, y ∈ W and for all formulas B, if x|=A R y|=A , B ∈ ΓA and M, x |= [R]B then M, y |= B. As a consequence, see Hughes and Cresswell [4], if A is true at some possible world in M then A is true at some possible world in M . (3 implies 4): The proof can be obtained by the canonical model construction. (4 implies 1): The proof is trivial because models satisfy the conditions which are needed to verify the axioms of LER. We now turn our attention to the decidability of the satisfiability problem for any given formula of LER. Theorem 3. (Decidability theorem for LER) The satisfiability problem for any given formula of LER is decidable. Proof. By theorem 2, LER is a finitely axiomatizable normal modal logic which has the finite model property. As a consequence, the satisfiability problem for any given formula of LER is decidable. A less obvious result concerns the inherent difficulty of the satisfiability problem for any given formula of LER. What we have in mind is to prove that the satisfiability problem for any given {[≡], [R]}-free formula A of LER is P SP ACEcomplete in length(A). We outline how this result can be proved, but leave the details to the rest of the paper. In section 4, we show how the satisfiability problem for any given formula of S4 can be linearly reduced to the satisfiability problem for any given {[≡], [R]}-free formula of LER. In section 5, we show how the satisfiability problem for any given {[≡], [R]}-free formula of LER can be solved by means of a polynomial-space bounded nondeterministic algorithm.
Emptiness Relations in Property Systems
23
P SP ACE-Hardness
4
We define the set of all formulas of S4 as follows: - A ::= p | ¬A | (A ∨ B) | 2A; where p ranges over a countably infinite set of propositional variables. We follow the standard rules for omission of the parentheses. The number of occurrences of symbols in formula A of S4 is denoted length(A). A model of S4 is a structure (W, |, V ) where: - W is a nonempty set of possible worlds; - | is a binary relation on W such that for all x, y ∈ W : - x | x; - If x | y and y | z then x | z; - V is a function with domain the set of all propositional variables and range the power set of W . Let M = (W, |, V ) be a model of S4. We define the relation “formula A of S4 is true at possible world x in M ”, denoted M, x |= A, as follows: -
M, x |= p iff x ∈ V (p); M, x |= ¬A iff M, x |= A; M, x |= A ∨ B iff M, x |= A or M, x |= B; M, x |= 2A iff for all y ∈ W , if x | y then M, y |= A.
Let t be the function that assigns to each formula A of S4 the {[≡], [R]}-free formula t(A) of LER as follows: -
t(p) = p; t(¬A) = ¬t(A); t(A ∨ B) = t(A) ∨ t(B); t(2A) = []t(A).
The interesting result is the following. Lemma 3. For all formulas A of S4, the following conditions are equivalent. 1. A is true at some possible world in some model of S4; 2. t(A) is true at some possible world in some model of LER. Proof. (1 implies 2): Let M = (W, |, V ) be a model of S4 and M = (W , , , , , V ) be the structure defined as follows. Define: -
W is W ; For all x, y ∈ W , x y iff x | y; For all x, y ∈ W , x y iff y | x; For all x, y ∈ W , x y; For all x, y ∈ W , x y; For all propositional variables p, V (p) is V (p).
24
P. Balbiani
It is a simple matter to check that M is a model of LER. The reader may easily verify by induction on the length of formula A of S4 that if A is true at some possible world in M then t(A) is true at some possible world in M . (2 implies 1): Let M = (W, , , , , V ) be a model of LER and M = (W , | , V ) be the structure defined as follows. Define: - W is W ; - For all x, y ∈ W , x | y iff x y; - For all propositional variables p, V (p) is V (p). It is a rather remarkable fact that M is a model of S4. The reader may easily verify by induction on the length of formula A of S4 that if t(A) is true at some possible world in M then A is true at some possible world in M . Seeing that the satisfiability problem for any given formula A of S4 is P SP ACEhard in length(A), see Ladner [5], we infer immediately the following. Theorem 4. (P SP ACE-hardness theorem for LER) The satisfiability problem for any given {[≡], [R]}-free formula A of LER is P SP ACE-hard in length(A).
P SP ACE-Completeness
5
Following the line of reasoning suggested by Spaan [12], let Cl be the function that assigns to each string ω ∈ {, , , } and to each {[≡], [R]}-free formula A of LER the smallest set Cl(ω, A) of {[≡], [R]}-free formulas of LER as follows: - A ∈ Cl(, A); - Cl(ω, A) is closed under subformulas; - For all {[≡], [R]}-free formulas B of LER: • If []B ∈ Cl(ω, A) then []B ∈ Cl(ω , A) and []B ∈ Cl(ω, A); • If []B ∈ Cl(ω, A) then []B ∈ Cl(ω , A) and []B ∈ Cl(ω, A); • If []B ∈ Cl(ω, A) then []B ∈ Cl(ω , A) and []B ∈ Cl(ω, A); • If []B ∈ Cl(ω, A) then []B ∈ Cl(ω , A) and []B ∈ Cl(ω, A); where denotes the empty string. The following lemma is basic. Lemma 4. Let A be a {[≡], [R]}-free formula of LER. -
Card(Cl(, A)) ≤ length(A); Card(Cl(, A)) ≤ length(A); Card(Cl(, A)) ≤ length(A); Card(Cl(, A)) ≤ length(A); Card(Cl(, A)) ≤ length(A); For all strings ω ∈ {, , , } : • Card(Cl(ω , A)) = Card(Cl(ω , A)); • Card(Cl(ω , A)) < Card(Cl(ω , A)); • Card(Cl(ω , A)) = Card(Cl(ω , A)); • Card(Cl(ω , A)) < Card(Cl(ω , A)); • Card(Cl(ω , A)) < Card(Cl(ω , A)); • Card(Cl(ω , A)) = Card(Cl(ω , A));
Emptiness Relations in Property Systems
• • • • • • • • • •
25
Card(Cl(ω , A)) < Card(Cl(ω , A)); Card(Cl(ω , A)) = Card(Cl(ω , A)); Card(Cl(ω , A)) < Card(Cl(ω, A)); Card(Cl(ω , A)) = Card(Cl(ω, A)); Card(Cl(ω , A)) < Card(Cl(ω, A)); Card(Cl(ω , A)) = Card(Cl(ω, A)); Card(Cl(ω , A)) = Card(Cl(ω A)); Card(Cl(ω , A)) < Card(Cl(ω, A)); Card(Cl(ω , A)) = Card(Cl(ω, A)); Card(Cl(ω , A)) < Card(Cl(ω, A)).
Proof. We first observe that Cl(, A) is the set of all subformulas of A. As a consequence Card(Cl(, A)) ≤ length(A). It is a simple matter to check that Cl(, A) does not contain more formulas than Cl(, A). Hence we have Card(Cl(, A)) ≤ length(A). Similarly Card(Cl(, A)) ≤ length(A), Card(Cl(, A)) ≤ length(A) and Card(Cl(, A)) ≤ length(A). It is a rather remarkable fact that Cl(ω , A) does not contain less formulas than Cl(ω , A). Hence we have Card(Cl(ω , A)) = Card(Cl(ω , A)). Similarly Card(Cl(ω , A)) = Card(Cl(ω , A)), Card(Cl(ω , A)) = Card(Cl(ω , A)), Card(Cl(ω , A)) = Card(Cl(ω , A)), Card(Cl(ω , A)) = Card(Cl(ω, A)), Card(Cl(ω , A)) = Card(Cl(ω, A)), Card(Cl(ω , A)) = Card(Cl(ω A)) and Card(Cl(ω , A)) = Card(Cl(ω, A)). The reader can readily check that Cl(ω , A) contains less formulas than Cl(ω , A). Hence we have Card(Cl(ω , A)) < Card(Cl(ω , A)). Similarly Card(Cl(ω , A)) < Card(Cl(ω , A)), Card(Cl(ω , A)) < Card(Cl(ω , A)), Card(Cl(ω , A)) < Card(Cl(ω , A)), Card(Cl(ω , A)) < Card(Cl(ω, A)), Card(Cl(ω , A)) < Card(Cl(ω, A)), Card(Cl(ω , A)) < Card(Cl(ω, A)) and Card(Cl(ω , A)) < Card(Cl(ω, A)). To continue we introduce the function depth that assigns to each string ω ∈ {, , , } the positive integer depth(ω) as follows: -
depth() = 0; depth() = 0; depth() = 0; depth() = 0; depth() = 0; For all strings ω ∈ {, , , } : • depth(ω ) = depth(ω ); • depth(ω ) = depth(ω ) + 1; • depth(ω ) = depth(ω ); • depth(ω ) = depth(ω ) + 1; • depth(ω ) = depth(ω ) + 1; • depth(ω ) = depth(ω ); • depth(ω ) = depth(ω ) + 1; • depth(ω ) = depth(ω ); • depth(ω ) = depth(ω) + 1; • depth(ω ) = depth(ω);
26
P. Balbiani
• • • • • •
depth(ω ) = depth(ω) + 1; depth(ω ) = depth(ω); depth(ω ) = depth(ω); depth(ω ) = depth(ω) + 1; depth(ω ) = depth(ω); depth(ω ) = depth(ω) + 1.
For our purpose, the crucial property of the function depth is the following. Lemma 5. For all strings ω ∈ {, , , } and for all {[≡], [R]}-free formulas A of LER, if depth(ω) ≥ length(A) then Cl(ω, A) = ∅. Proof. The proof is conducted by induction on the length of {[≡], [R]}-free formula A of LER. Basis: Let p be a propositional variable. For all strings ω ∈ {, , , } , if depth(ω) ≥ length(p) then depth(ω) ≥ 1. Hence Cl(ω, p) = ∅. Hypothesis: Let B, C be {[≡], [R]}-free formulas of LER such that for all strings ω ∈ {, , , } , if depth(ω) ≥ length(B) then Cl(ω, B) = ∅ and if depth(ω) ≥ length(C) then Cl(ω, C) = ∅. Step: For all strings ω ∈ {, , , } , if depth(ω) ≥ length(¬B) then depth(ω) > length(B). It follows that Cl(ω, B) = ∅ and Cl(ω, ¬B) ⊆ Cl(ω, B). Therefore Cl(ω, ¬B) = ∅. For all strings ω ∈ {, , , } , if depth(ω) ≥ length(B ∨C) then depth(ω) > length(B) and depth(ω) > length(C). It follows that Cl(ω, B) = ∅, Cl(ω, C) = ∅ and Cl(ω, B ∨ C) ⊆ Cl(ω, B) ∪ Cl(ω, C). Therefore Cl(ω, B ∨ C) = ∅. For all strings ω ∈ {, , , } , if depth(ω) ≥ length([]B) then depth(ω) > length(B). It follows that there is a string ω ∈ {, , , } such that ω = ω or ω = ω or ω = ω or ω = ω or ω = ω or ω = ω or ω = ω or ω = ω . Suppose that ω = ω . Hence depth(ω ) ≥ length(B). It follows that Cl(ω , B) = ∅ and Cl(ω, []B) ⊆ Cl(ω , B). Therefore Cl(ω, []B) = ∅. The same line of reasoning applies if ω = ω or ω = ω or ω = ω or ω = ω or ω = ω or ω = ω or ω = ω . For all strings ω ∈ {, , , } , the same line of reasoning applies if depth(ω) ≥ length([]B) or depth(ω) ≥ length([]B) or depth(ω) ≥ length([]B). Let S be a set of {[≡], [R]}-free formulas of LER and A be a {[≡], [R]}-free formula of LER. A (, A)-witness of S will be any set T of {[≡], [R]}-free formulas of LER such that A ∈ T and for all {[≡], [R]}-free formulas B of LER: -
If If If If
[]B ∈ S then []B ∈ T ; []B ∈ S then []B ∈ T ; []B ∈ T then []B ∈ S; []B ∈ T then []B ∈ S.
A (, A)-witness of S will be any set T of {[≡], [R]}-free formulas of LER such that A ∈ T and for all {[≡], [R]}-free formulas B of LER:
Emptiness Relations in Property Systems
-
If If If If
27
[]B ∈ S then []B ∈ T ; []B ∈ S then []B ∈ T ; []B ∈ T then []B ∈ S; []B ∈ T then []B ∈ S.
A (, A)-witness of S will be any set T of {[≡], [R]}-free formulas of LER such that A ∈ T and for all {[≡], [R]}-free formulas B of LER: -
If If If If
[]B ∈ S then []B ∈ T ; []B ∈ S then []B ∈ T ; []B ∈ T then []B ∈ S; []B ∈ T then []B ∈ S.
A (, A)-witness of S will be any set T of {[≡], [R]}-free formulas of LER such that A ∈ T and for all {[≡], [R]}-free formulas B of LER: -
If If If If
[]B ∈ S then []B ∈ T ; []B ∈ S then []B ∈ T ; []B ∈ T then []B ∈ S; []B ∈ T then []B ∈ S.
To test the satisfiability problem for any given {[≡], [R]}-free formula A of LER, let SAT (A) be the nondeterministic algorithm defined as follows. Define: function SAT (A) returns Boolean begin for all subsets S of Cl(, A) such that A ∈ S do W ORLD(S, , A) if all these calls return false then return false return true end where W ORLD(L, ω, A) is the nondeterministic algorithm defined as follows. Our nondeterministic algorithm W ORLD is similar to Spaan’s nondeterministic algorithm S4t − W ORLD — a procedure developed by Spaan [12] to investigate the inherent difficulty of the satisfiability problem for any given formula of S4t — in the sense that we keep track of a list L of sets of {[≡], [R]}-free formulas and a string ω ∈ {, , , } . Define: function W ORLD(L, ω, A) returns Boolean begin let S be last(L) for all {[≡], [R]}-free formulas B of LER such that ¬B ∈ Cl(ω, A) do if (¬B ∈ S and B ∈ S) or (¬B ∈ S and B ∈ S) then return false for all {[≡], [R]}-free formulas B, C of LER such that B ∨ C ∈ Cl(ω, A) do if (B ∨ C ∈ S and B ∈ S and C ∈ S) or (B ∨ C ∈ S and B ∈ S) or (B ∨ C ∈ S and C ∈ S) then return false for all {[≡], [R]}-free formulas B of LER such that []B ∈ Cl(ω, A) do begin if []B ∈ S and B ∈ S then return false if []B ∈ S and there is no T in L such that T is a (, B)-witness of S then
28
P. Balbiani
begin for all subsets T of Cl(ω , A) such that T is a (, B)-witness of S do if ω = or ω is ended by or then W ORLD(T, ω , A) else W ORLD(L ◦ T, ω , A) if all these calls return false then return false end end for all {[≡], [R]}-free formulas B of LER such that []B ∈ Cl(ω, A) do begin if []B ∈ S and B ∈ S then return false if []B ∈ S and there is no T in L such that T is a (, B)-witness of S then begin for all subsets T of Cl(ω , A) such that T is a (, B)-witness of S do if ω = or ω is ended by or then W ORLD(T, ω , A) else W ORLD(L ◦ T, ω , A) if all these calls return false then return false end end for all {[≡], [R]}-free formulas B of LER such that []B ∈ Cl(ω, A) do if []B ∈ S and there is no T in L such that T is a (, B)-witness of S then begin for all subsets T of Cl(ω, A) such that T is a (, B)-witness of S do if ω = or ω is ended by or then W ORLD(T, ω, A) else W ORLD(L ◦ T, ω, A) if all these calls return false then return false end for all {[≡], [R]}-free formulas B of LER such that []B ∈ Cl(ω, A) do if []B ∈ S and there is no T in L such that T is a (, B)-witness of S then begin for all subsets T of Cl(ω, A) such that T is a (, B)-witness of S do if ω = or ω is ended by or then W ORLD(T, ω, A) else W ORLD(L ◦ T, ω, A) if all these calls return false then return false end return true end where last(L) denotes the last element of L and L ◦ T denotes the list obtained by extending L with T . The following lemma is basic. Lemma 6. For all {[≡], [R]}-free formulas A of LER and for all subsets S of Cl(, A), the following conditions are equivalent. 1. W ORLD(S, , A) succeeds; 2. There is a model M = (W, , , , , V ) of LER and there is a possible world x ∈ W such that for all {[≡], [R]}-free formulas B of LER, if B ∈ Cl(, A) then M, x |= B iff B ∈ S. Proof. (1 implies 2): Suppose W ORLD(S, , A) succeeds and let M = (W, , , , , V ) be the structure defined as follows. Define:
Emptiness Relations in Property Systems
29
- W is the set of all pairs (T, ω) for which there is a finite sequence (L0 , ω0 ), . . . , (LK , ωK ) such that: - L0 = S and ω0 = ; - last(LK ) = T and ωK = ω; - For all positive integers k, if k ≤ K then W ORLD(Lk , ωk , A) is successfully called in W ORLD(S, , A) at some depth of the recursion; - For all positive integers k, if k < K then W ORLD(Lk+1 , ωk+1 , A) is called by W ORLD(Lk , ωk , A); - For all (T, ω), (T , ω ) ∈ W , (T, ω) (T , ω ) iff there is a finite sequence (L0 , ω0 ), . . . , (LK , ωK ) such that: - L0 = S and ω0 = ; - last(LK ) = T and ωK = ω; - For all positive integers k, if k ≤ K then W ORLD(Lk , ωk , A) is successfully called in W ORLD(S, , A) at some depth of the recursion; - For all positive integers k, if k < K then W ORLD(Lk+1 , ωk+1 , A) is called by W ORLD(Lk , ωk , A); and there is a finite sequence (L0 , ω0 ), . . . , (LK , ωK ) such that: - L0 = S and ω0 = ; - last(LK ) = T and ωK = ω ; - For all positive integers k, if k ≤ K then W ORLD(Lk , ωk , A) is successfully called in W ORLD(S, , A) at some depth of the recursion; , A) is - For all positive integers k, if k < K then W ORLD(Lk+1 , ωk+1 called by W ORLD(Lk , ωk , A); and either (LK , ωK ) = (LK , ωK ) or W ORLD(LK , ωK , A) is called by W ORLD(LK , ωK , A) in the []-segment or the finite sequence (L0 , ω0 ), . . . , (LK , ωK ) is a prefix of the finite sequence (L0 , ω0 ), . . . , (LK , ωK ), the call W ORLD(LK , ωK , A) enters in the []-segment and there is a {[≡], [R]}-free formula B of LER such that []B ∈ Cl(ωK , A), []B ∈ T and T is a (, B)-witness of T ; - For all (T, ω), (T , ω ) ∈ W , (T, ω) (T , ω ) iff there is a finite sequence (L0 , ω0 ), . . . , (LK , ωK ) such that: - L0 = S and ω0 = ; - last(LK ) = T and ωK = ω; - For all positive integers k, if k ≤ K then W ORLD(Lk , ωk , A) is successfully called in W ORLD(S, , A) at some depth of the recursion; - For all positive integers k, if k < K then W ORLD(Lk+1 , ωk+1 , A) is called by W ORLD(Lk , ωk , A); and there is a finite sequence (L0 , ω0 ), . . . , (LK , ωK ) such that: - L0 = S and ω0 = ; - last(LK ) = T and ωK = ω ; - For all positive integers k, if k ≤ K then W ORLD(Lk , ωk , A) is successfully called in W ORLD(S, , A) at some depth of the recursion; , A) is - For all positive integers k, if k < K then W ORLD(Lk+1 , ωk+1 called by W ORLD(Lk , ωk , A); and either (LK , ωK ) = (LK , ωK ) or W ORLD(LK , ωK , A) is called by W ORLD(LK , ωK , A) in the []-segment or the finite sequence (L0 , ω0 ), . . . , (LK , ωK ) is a prefix of the finite sequence (L0 , ω0 ), . . . , (LK , ωK ), the call W ORLD(LK , ωK , A) enters in the []-segment and there is a {[≡], [R]}-free
30
P. Balbiani
formula B of LER such that []B ∈ Cl(ωK , A), []B ∈ T and T is a (, B)-witness of T ; - For all (T, ω), (T , ω ) ∈ W , (T, ω) (T , ω ) iff there is a finite sequence (L0 , ω0 ), . . . , (LK , ωK ) such that: - L0 = S and ω0 = ; - last(LK ) = T and ωK = ω; - For all positive integers k, if k ≤ K then W ORLD(Lk , ωk , A) is successfully called in W ORLD(S, , A) at some depth of the recursion; - For all positive integers k, if k < K then W ORLD(Lk+1 , ωk+1 , A) is called by W ORLD(Lk , ωk , A); and there is a finite sequence (L0 , ω0 ), . . . , (LK , ωK ) such that: - L0 = S and ω0 = ; - last(LK ) = T and ωK = ω ; - For all positive integers k, if k ≤ K then W ORLD(Lk , ωk , A) is successfully called in W ORLD(S, , A) at some depth of the recursion; , A) is - For all positive integers k, if k < K then W ORLD(Lk+1 , ωk+1 called by W ORLD(Lk , ωk , A); and either W ORLD(LK , ωK , A) is called by W ORLD(LK , ωK , A) in the []-segment or the finite sequence (L0 , ω0 ), . . . , (LK , ωK ) is a prefix of the finite sequence (L0 , ω0 ), . . . , (LK , ωK ), the call W ORLD(LK , ωK , A) enters in the []-segment and there is a {[≡], [R]}-free formula B of LER such that []B ∈ Cl(ωK , A), []B ∈ T and T is a (, B)-witness of T ; - For all (T, ω), (T , ω ) ∈ W , (T, ω) (T , ω ) iff there is a finite sequence (L0 , ω0 ), . . . , (LK , ωK ) such that: - L0 = S and ω0 = ; - last(LK ) = T and ωK = ω; - For all positive integers k, if k ≤ K then W ORLD(Lk , ωk , A) is successfully called in W ORLD(S, , A) at some depth of the recursion; - For all positive integers k, if k < K then W ORLD(Lk+1 , ωk+1 , A) is called by W ORLD(Lk , ωk , A); and there is a finite sequence (L0 , ω0 ), . . . , (LK , ωK ) such that: - L0 = S and ω0 = ; - last(LK ) = T and ωK = ω ; - For all positive integers k, if k ≤ K then W ORLD(Lk , ωk , A) is successfully called in W ORLD(S, , A) at some depth of the recursion; , A) is - For all positive integers k, if k < K then W ORLD(Lk+1 , ωk+1 called by W ORLD(Lk , ωk , A); and either W ORLD(LK , ωK , A) is called in W ORLD(LK , ωK , A) in the []-segment or the finite sequence (L0 , ω0 ), . . . , (LK , ωK ) is a prefix of the finite sequence (L0 , ω0 ), . . . , (LK , ωK ), the call W ORLD(LK , ωK , A) enters in the []-segment and there is a {[≡], [R]}-free formula B of LER such that []B ∈ Cl(ωK , A), []B ∈ T and T is a (, B)-witness of T ; - For all propositional variables p, V (p) is the set of all (T, ω) ∈ W such that p ∈ T. Now the definition of M can be completed. Firstly define:
Emptiness Relations in Property Systems
-
31
0 is ∪ −1 ; 0 is ∪ −1 ; 0 is ∪ −1 ; 0 is ∪ −1 .
Secondly define for all positive integers k: -
k+1 is (k ◦ k ) ∪ (k ◦ k ); k+1 is (k ◦ k ) ∪ (k ◦ k ); k+1 is (k ◦k ) ∪ (k ◦ k ); k+1 is (k ◦k ) ∪ (k ◦ k ).
Thirdly define: - is {k : k is a positive integer}; - is {k : k is a positive integer}; - is {k : k is a positive integer}; - is {k : k is a positive integer}. It is a simple matter to check that M is a model of LER. What is more one can establish by induction on the positive integer k the remarkable result that for all {[≡], [R]}-free formulas B of LER and for all (T, ω), (T , ω ) ∈ W : -
If If If If If If If If
[]B ∈ T []B ∈ T []B ∈ T []B ∈ T []B ∈ T []B ∈ T []B ∈ T []B ∈ T
and (T, ω) k (T , ω ) then []B ∈ T ; and (T, ω) k (T , ω ) then []B ∈ T ; and (T, ω) k (T , ω ) then []B ∈ T ; and (T, ω) k (T , ω ) then []B ∈ T ; and (T, ω) k (T , ω ) then []B ∈ T ; and (T, ω) k (T , ω ) then []B ∈ T ; and (T, ω) k (T , ω ) then []B ∈ T ; and (T, ω) k (T , ω ) then []B ∈ T .
It follows immediately by induction on the length of {[≡], [R]}-free formula B of LER that for all (T, ω) ∈ W , if B ∈ Cl(ω, A) then: - M, (T, ω) |= B iff B ∈ T . Hence we have that for all {[≡], [R]}-free formulas B of LER, if B ∈ Cl(, A) then: - M, (S, ) |= B iff B ∈ S. (2 implies 1): The reader may easily prove by induction on the depth of string ω ∈ {, , , } that for all sets T of {[≡], [R]}-free formulas of LER, for all models M = (W, , , , , V ) of LER and for all possible worlds x ∈ W , if T ⊆ Cl(ω, A) and: - For all {[≡], [R]}-free formulas B of LER, if B ∈ Cl(ω, A) then M, x |= B iff B ∈ T ; then W ORLD(T, ω, A) succeeds. These considerations prove that if there is a model M = (W, , , , , V ) of LER and there is a possible world x ∈ W such that:
32
P. Balbiani
- For all {[≡], [R]}-free formulas B of LER, if B ∈ Cl(, A) then M, x |= B iff B ∈ S; then W ORLD(S, , A) succeeds. It follows immediately from lemma 6 that the nondeterministic algorithm SAT (A) works correctly, i.e., SAT (A) succeeds iff A is true at some possible world in some model of LER. Much more difficult than lemma 6 are the following important results. Lemma 7. For all {[≡], [R]}-free formulas A of LER, for all lists L of sets of {[≡], [R]}-free formulas and for all strings ω ∈ {, , , } , if W ORLD(L, ω, A) is called in W ORLD(S, , A) at some depth of the recursion then length(L) = O(length(A)2 ). Proof. Similar to the line of reasoning suggested by Demri [2] at pages 225–226. Lemma 8. For all {[≡], [R]}-free formulas A of LER, for all lists L of sets of {[≡], [R]}-free formulas and for all strings ω ∈ {, , , } , if W ORLD(L, ω, A) is called in W ORLD(S, , A) at some depth of the recursion then depth(ω) ≤ length(A). Proof. Suppose that depth(ω) > length(A). It follows that there is a string ω ∈ {, , , } such that ω = ω or ω = ω or ω = ω or ω = ω or ω = ω or ω = ω or ω = ω or ω = ω . Suppose that ω = ω . Hence depth(ω ) ≥ length(A). It follows that Cl(ω , A) = ∅: a contradiction. Therefore depth(ω) ≤ length(A). The same line of reasoning applies if ω = ω or ω = ω or ω = ω or ω = ω or ω = ω or ω = ω or ω = ω . By applying lemma 7 and lemma 8, we infer immediately that the nondeterministic algorithm SAT (A) is polynomial-space bounded in length(A). We can summarize the results proved above as follows for all {[≡], [R]}-free formulas A of LER: - SAT (A) succeeds iff A is true at some possible world in some model of LER; - The total space to run SAT (A) is polynomial in length(A). We get that the satisfiability problem for any given {[≡], [R]}-free formula A of LER is in N P SP ACE of length(A). Seeing that N P SP ACE = P SP ACE, see Savitch [11], we therefore conclude the following. Theorem 5. (P SP ACE-completeness theorem for LER) The satisfiability problem for any given {[≡], [R]}-free formula A of LER is P SP ACE-complete in length(A).
6
Conclusion
The key feature of this paper is the modal analysis of emptiness relations in property systems. The proof that the satisfiability problem for any given {[≡], [R]}-free formula A of LER is P SP ACE-complete in length(A) goes
Emptiness Relations in Property Systems
33
through a line of reasoning similar to that applied by Spaan [12] to investigate the inherent difficulty of the satisfiability problem for any given formula of S4t . Let us note that our proof does not consider the full language of our modal logic, so it might not be suitable for studying the inherent difficulty of the satisfiability problem for any given formula of LER, a question that remains unsolved. Previous modal analyses of informational relations in property systems have been given by Vakarelov [16] who has mainly considered the relations of informational inclusion — our emptiness relations and — and the relations of similarity — the complementary relations of our emptiness relations and . On that occasion, Vakarelov [16] gave a completeness theorem and a decidability theorem for a modal logic which modalities correspond to the relations of informational inclusion and the relations of similarity in property systems. Modal analyses of our emptiness relations in property systems together with other informational relations like similarity relations are not known.
References 1. Balbiani, P., Vakarelov, D.: First-order characterization and modal analysis of indiscernibility and complementarity in information systems. In Benferhat, S., Besnard, P. (Editors): Symbolic and Quantitative Approaches to Reasoning with Uncertainty. Springer-Verlag, Lecture Notes in Artificial Intelligence 2143 (2001) 772–781. 2. Demri, S.: The nondeterministic information logic NIL is PSPACE-complete. Fundamenta Informaticæ42 (2000) 211–234. 3. Demri, S., Orlowska, E., Vakarelov, D.: Indiscernibility and complementarity relations in information systems. In Gerbrandy, J., Marx, M., de Rijke, M., Venema, Y. (Editors): JFAK: Essays Dedicated to Johan van Benthem on the Occasion of his 50th Birthday. Amsterdam University Press (1999) http://turing.wins.uva.nl/˜j50/cdrom/contribs/demri/index.html. 4. Hughes, G., Cresswell, M.: A Companion to Modal Logic. Methuen (1984). 5. Ladner, R.: The computational complexity of provability in systems of modal propositional logic. SIAM Journal on Computing 6 (1977) 467–480. 6. Orlowska, E.: Logic of nondeterministic information. Studia Logica 44 (1985) 91– 100. 7. Orlowska, E.: Kripke semantics for knowledge representation logics. Studia Logica 49 (1990) 255–272. 8. Orlowska, E. (Editor): Incomplete Information: Rough Set Analysis. PhysicaVerlag, Studies in Fuzziness and Soft Computing 13 (1998). 9. Orlowska, E., Pawlak, Z.: Representation of nondeterministic information. Theoretical Computer Science 29 (1984) 27–39. 10. Pawlak, Z.: Information systems — theoretical foundations. Information Systems 6 (1981) 205–218. 11. Savitch, W.: Relationships between nondeterministic and deterministic tape complexities. Journal of Computer and System Sciences 4 (1970) 177–192. 12. Spaan, E.: The complexity of propositional tense logics. In de Rijke, M. (Editor): Diamonds and Defaults: Studies in Pure and Applied Intensional Logic. Kluwer, Synthese Library 229 (1993) 287–307.
34
P. Balbiani
13. Vakarelov, D.: A modal logic for similarity relations in Pawlak knowledge representation systems. Fundamenta Informaticæ15 (1991) 61–79. 14. Vakarelov, D.: Modal logics for knowledge representation systems. Theoretical Computer Science 90 (1991) 433–456. 15. Vakarelov, D.: A duality between Pawlak’s knowledge representation systems and BI-consequence systems. Studia Logica 55 (1995) 205–228. 16. Vakarelov, D.: Information systems, similarity relations and modal logics. In Orlowska, E. (Editor): Incomplete Information: Rough Set Analysis. PhysicaVerlag, Studies in Fuzziness and Soft Computing 13 (1998) 492–550.
Pregroups: Models and Grammars Wojciech Buszkowski Faculty of Mathematics and Computer Science, Adam Mickiewicz University, Pozna´ n, Poland
[email protected] Abstract. Pregroups, introduced in Lambek [12], are a generalization of partially ordered groups. In [5], we have proven several theorems on pregroups and grammars based on the calculus of free pregroups, in particular, the weak equivalence of these grammars and context-free grammars. In the present paper, we obtain further results of that kind. We consider left and right pregroups, study concrete left and right pregroups consisting of monotone functions on a poset and of monotone relations on a poset, and adjust the equivalence theorem to grammars based on left (right) pregroups.
1
Pregroups
A structure (G, ≤, ·, 1) is called a partially ordered monoid (p.o. monoid), if (G, ·, 1) is a monoid (i.e. a semigroup with unit 1), (G, ≤) is a poset (i.e. a partially ordered set), and the following monotonicity condition holds: (MON) if a ≤ b then ca ≤ cb and ac ≤ bc, for all a, b, c ∈ G. A pregroup is defined as a structure (G, ≤, ·, l, r, 1) such that (G, ≤, ·, 1) is a p.o. monoid, and l, r are unary operations on G, satisfying the following inequalities: al a ≤ 1 ≤ aal and aar ≤ 1 ≤ ar a , (1) for all a ∈ G. The element al (resp. ar ) is called the left (resp. right) adjoint of a. Standard examples of pregroups are partially ordered groups (p.o. groups), i.e. structures (G, ≤, ·,∪ , 1) such that (G, ≤, ·, 1) is a p.o. monoid, and ∪ is a unary operation on G, satisfying the equalities: a∪ a = 1 = aa∪ , for all a ∈ G. An extensive exposition of the theory of p.o. groups is given in Fuchs [8]. Clearly, in a p.o. group one can define al = ar = a∪ . A pregroup G is said to be commutative, if ab = ba, for all a, b ∈ G. If G is commutative, then al = ar , for all a ∈ G, and G is, actually, a p.o. group. Accordingly, the notion of a pregroup is a generalization of that of a p.o. group which can be innovatory for noncommutative structures only. A pregroup is said to be proper, if it is not a p.o. group (that means, al = ar , for some element a). H. de Swart (Ed.): RelMiCS 2001, LNCS 2561, pp. 35–49, 2002. c Springer-Verlag Berlin Heidelberg 2002
36
W. Buszkowski
Pregroups are related to residuated monoids, abstract models for the Lambek calculus with possibly empty antecedents [11,4]. A residuated monoid is a structure (G, ≤, ·, \, /, 1) such that (G, ≤, ·, 1) is a p.o. monoid, and \, / are binary operations on G, fulfilling the equivalences: ab ≤ c iff b ≤ a\c iff a ≤ c/b ,
(2)
for all a, b, c ∈ G. (Actually, (MON) follows from (2).) For a pregroup G, one defines a\b = ar b and a/b = abl , which yields a residuated monoid. One can also define al = 1/a and ar = a\1. The latter definitions of adjoints, applied to an arbitrary residuated monoid, lead to inequalities al a ≤ 1 and aar ≤ 1, but the inequalities 1 ≤ aal and 1 ≤ ar a need not hold. The Lambek calculus with possibly empty antecedents is a formal system whose theorems are the inequalities valid in residuated monoids. Consequently, all theorems of the Lambek calculus are also valid in pregroups under the above interpretation of residuals a\b and a/b. The converse is not true even for productfree types [5]. Grammars based on the Lambek calculus are a kind of categorial grammars, extensively studied nowadays as a framework for deductive parsing (see [14,15,3]). Lambek [12] proposes alternative parsing strategies, based on the calculus of free pregroups (also see [2,7]). We discuss these matters in section 3. Using (1), one easily proves the following equalities hold in every pregroup: 1l = 1r = 1, alr = arl = a, (ab)l = bl al , (ab)r = br ar .
(3)
By (1), we also derive: aal a = a, aar a = a ,
(4)
and (MON) with (1) yield: (MON’) a ≤ b iff bl ≤ al iff br ≤ ar . By Z we denote the set of integers. Following [12], for any pregroup G, a ∈ G, and n ∈Z, we define the element a(n) ∈ G: a(0) = a, a(n+1) = (a(n) )r , for n ≥ 0, and a(n−1) = (a(n) )l , for n ≤ 0. Using (3), one shows the latter equalities hold for all n ∈Z. Also: a(n) a(n+1) ≤ 1 ≤ a(n+1) a(n) , for all n ∈ Z ,
(5)
(ab)(n) = a(n) b(n) , and if a ≤ b then a(n) ≤ b(n) (n is even) ,
(6)
(n)
(ab)
=b
a
(n) (n)
, and if a ≤ b then b
(n)
≤a
(n)
(n is odd) .
(7)
Proposition 1. Let G be a p.o. monoid. Then, for any a ∈ G, there exist at most one b ∈ G such that ba ≤ 1 ≤ ab and at most one c ∈ G such that ac ≤ 1 ≤ ca. Proof. Assume ba ≤ 1 ≤ ab and b a ≤ 1 ≤ ab . Then b ≤ bab ≤ b and b ≤ b ab ≤ b, hence b = b .
Pregroups: Models and Grammars
37
As a consequence, al is the only element b such that ba ≤ 1 ≤ ab, and ar is the only element c such that ac ≤ 1 ≤ ca (this yields an easy proof for (3)). Further, ()l , ()r , ()(n) can be regarded as partial operations on arbitrary p.o. monoids, and they satisfy (1), (MON’), (3), (4), (5), (6), (7), if the relevant elements exist; for instance, if a(n) , b(n) exist, then (ab)(n) exists and the equation from (6) holds. In a pregroup, right and left adjoints are uniquely determined by the underlying p.o. monoid. Accordingly, we can also define pregroups as p.o. monoids in which every element admits the two adjoints (then, the signature of pregroups equals that of p.o. monoids, and we may say, for example: a given pregroup is a substructure of a given p.o. monoid). Proposition 2. For every p.o. monoid G, there exists a (unique) largest pregroup which is a substructure of G. Proof. Let H consist of all a ∈ G such that a(n) exists, for all n ∈Z. By (6), (7), H is a submonoid of G. By the definition, H admits adjoints. Every pregroup H which is a substructure of G must be contained in H. The pregroup H, defined above, will be denoted by Pr(G). Let (X, ≤) be a poset. A function f : X → X is said to be monotone, if x ≤ y entails f (x) ≤ f (y). By F (X, ≤) we denote the p.o. monoid of all monotone functions f : X → X with the identity function I being the unit, the function composition being the operation ·, and the ordering being defined by: f ≤ g iff f (x) ≤ g(x), for all x ∈ X. Proposition 3. Every pregroup (G, ≤, ·, l, r, 1) is embeddable into Pr(F (G, ≤)). Proof. For a ∈ G, the function fa : G → G is defined as follows: fa (x) = ax. By (MON), fa is monotone. We have: fab = fa ◦ fb , f1 = I, and a ≤ b iff fa ≤ fb . Consequently, the mapping h(a) = fa is a monomorphism of p.o. monoids. Since fal ◦ fa ≤ I ≤ fa ◦ fal and fa ◦ far ≤ I ≤ far ◦ fa , then fal = fal and far = far , hence h is a monomorphism of pregroups. Consequently, every pregroup is isomorphic to a pregroup of monotone functions on a poset. If (X, ≤) is a poset and Y ⊆ X, then maxY and minY will denote the greatest and the least, respectively, element of Y (thus, maxY is the l.u.b. of Y belongimg to Y , and minY is the g.l.b. of Y belonging to Y ). Proposition 4. For every function f ∈Pr(F (X, ≤)), there hold the equalities: f l (x) =min{y ∈ X : x ≤ f (y)}, for all x ∈ X, f r (x) =max{y ∈ X : f (y) ≤ x}, for all x ∈ X. Further, if M is a submonoid of F (X, ≤) and, for every f ∈ M , the functions f l , f r defined by these equations exist and belong to M , then M is a pregroup. Proof. The first part is easy (see [5]). We prove the second part. Let M satisfy the condition. We prove (1). Since f (x) ≤ f (x), then f l (f (x)) ≤ x (so, f l ◦ f ≤ I). Also x ≤ f (f l (x)), since f l (x) is a y such that x ≤ f (y).
38
W. Buszkowski
These equalities can equivalently be expressed by the following Galois connections: f l (x) ≤ y iff x ≤ f (y) , (8) y ≤ f r (x) iff f (y) ≤ x ,
(9)
for all x, y ∈ X. As a consequence, all functions in Pr(F (X, ≤)) must be upward and downward unbounded, that means: ∀x∃y(x ≤ f (y)), ∀x∃y(f (y) ≤ x) ,
(10)
for all x, y ∈ X.
2
Left and Right Pregroups
A left (resp. right) pregroup is a p.o. monoid in which every element admits a left (resp. right) adjoint, that means, for every element a there exists an element al (resp. ar ) such that the left (resp. right) conjunct of (1) holds. Right pregroups are dual to left pregroups: (G, ≤, ·, 1) is a left pregroup iff (G, ≥, ·, 1) is a right pregroup. A pregroup is a p.o. monoid which is both a left pregroup and a right pregroup. Everything which has been said above on pregroups can also be said on left (right) pregroups provided one confine herself to left (right) adjoints. In particular, every p.o. monoid G contains a largest left (resp. right) pregroup to be denoted by LPr(G) (resp. RPr(G)); its elements are all a ∈ G such that a(n) exists, for all n ≤ 0 (resp. n ≥ 0). Clearly, Pr(G) equals the meet of LPr(G) and RPr(G). If we prove later on that no left (right) pregroup having a given property exists, then it follows that no pregroup having this property exists. Of course, some arguments sound for pregroups need not be sound for left (right) pregroups. For example, to prove (MON’) for pregroups one can prove: a ≤ b entails bl ≤ al and br ≤ ar , and then, substitute bl for a and al for b, and use the second equalities of (3) to obtain: bl ≤ al entails a ≤ b. For left pregoups, the latter entailment can be shown in another way. Assume bl ≤ al . Then: a ≤ bbl a ≤ bal a ≤ b. By an analogue of proposition 4, all functions in LPr(F (X, ≤)) must be upward unbounded, and all functions in RPr(F (X, ≤)) must be downward unbounded. These claims are dual to each other, since LPr(F (X, ≤)) equals RPr(F (X, ≥)). Proposition 5. Let a be the least (resp. greatest) element of poset (X, ≤). Then, for every function f ∈RPr(F (X, ≤)) (resp. LPr(F (X, ≤))) there holds: f (x) = a iff x = a, for all x ∈ X. Proof. Let f ∈RPr(F (X, ≤)). Then, f (a) = a, since f is downward unbounded. Assume f (x) = a. Then a = f r (a) ≥ x, by proposition 4, hence x = a.
Pregroups: Models and Grammars
39
In [5], it has been shown that, for any totally ordered set (X, ≤) containing more than two elements, Pr(F (X, ≤)) consists of all both upward and downward unbounded, monotone functions f : X → X if and only if (X, ≤) is isomorphic to the set of integers with the natural ordering. We prove a left (right) analogue of this theorem. Theorem 1. Let (X, ≤) be a totally ordered set containing at least two elements. Then, LPr(F (X, ≤)) consists of all upward unbounded, monotone functions f : X → X if and only if (X, ≤) is a well ordered set without the greatest element. Dually, RPr(F (X, ≤)) consists of all downward unbounded, monotone functions f : X → X if and only if (X, ≥) is a well ordered set without the greatest element. Proof. Assume (X, ≤) be a nonempty, well ordered set without the greatest element. Let M denote the set of all upward unbounded, monotone functions f : X → X. M is closed under function composition and I ∈ M , hence M is a submonoid of F (X, ≤). Further, for any f ∈ M , the function f l defined by the equation from proposition 4 exists and belongs to M (Caution: the nonexistence of the greatest element is needed to show that F l is upward unbounded.) So, M is a left pregroup. But all functions in LPr(F (X, ≤)) are upward unbounded, hence M equals the latter. Now, let (X, ≤) be a totally ordered set containing at least two elements. Assume LPr(F (X, ≤)) consist of all upward unbounded, monotone functions f : X → X. Suppose a to be the greatest element of X. Define f (x) = a, for all x ∈ a. Clearly, f is upward unbounded and monotone, hence f ∈LPr(F (X, ≤)), which contradicts proposition 5. Suppose (X, ≤) not to be well ordered. Then, there exists an infinite sequence an ∈ X, n ≥ 0, such that an+1 < an , for all n. Let Y denote the set of all y ∈ X such that an+1 ≤ y ≤ an , for some n. Define f (x) = a0 , for x ∈ Y , and f (x) = x, for x ∈ Y . Clearly, f is upward unbounded and monotone. But f l (a0 ) =min(Y ) does not exist. Contradiction. By ω we denote the set of nonnegative integers. It follows from theorem 1 that LPr(F (ω, ≤)), where ≤ is the natural ordering, contains all upward unbounded, monotone functions from ω into ω. It is easy to see that this left pregroup is proper (that means, it is not a p.o. group). Below we show how to do that. An element a of a left (resp. right) pregroup is said to be injective, if al a = 1 (resp. ar a = 1), and surjective, if aal = 1 (resp. aar = 1); it is said to be bijective, if it is both injective and surjective. In a left pregroup of functions, a function f is injective (resp. surjective) iff it is an injective (resp. a surjective) mapping. We prove this fact. Assume f to be injective. Then, f l ◦ f = I, so f is an injective mapping, since I is so. In a similar way, we show that, if f is surjective, then f is a surjective mapping. Assume f be an injective mapping. By (4), f ◦ f l ◦ f = f = f ◦ I, and consequently, f l ◦ f = I, so f is injective. In a similar way, we show that, if f is a surjective mapping, then f is surjective. The same holds true for right pregroups of functions.
40
W. Buszkowski
Accordingly, in a left (right) pregroup of functions, a function f is bijective iff f is a bijective mapping. A left (right) pregroup is proper iff it contains at least one nonbijective element. Thus, LPr(F (ω, ≤)) is proper, since it contains nonbijective mappings, e.g. f (x) = 2x. Proposition 6. For any left pregroup G and all a, b ∈ G, the following conditions hold true: (i) a is injective (resp. surjective) iff al is surjective (resp. injective), (ii) if ab is injective (resp. surjective), then b is injective (resp. a is surjective). Proof. If al a = 1, then al all = 1, and if aal = 1, then all al = 1, by (3). The converse conditionals also hold, by (3) and (MON’). This proves (i). If (ab)l ab = 1, then 1 = bl al ab ≤ bl b ≤ 1, hence bl b = 1, and if ab(ab)l = 1, then 1 = abbl al ≥ aal ≥ 1, hence aal = 1, which proves (ii).
An analogous fact is true for right pregroups. Proposition 7. If all elements of a left (right) pregroup are injective or surjective, then all elements are bijective. Proof. Assume that all elements of a left pregroup G be injective or surjective. Let a ∈ G. We prove that a is injective. Suppose not. Then, al a is not injective, hence al a is surjective, and consequently, al is surjective. So, a is injective. Contradiction. In a similar way, one proves that a is surjective. A left (right) pregroup is said to be strictly proper, if 1 is its only element which is injective or surjective. Every left (right) pregroup G contains a (unique) largest strictly proper left (right) pregroup G which consists of all elements a of the former such that a = 1 or a is neither injective, nor surjective (by proposition 6, this set is closed under · and adjoints). If G is a pregroup, then G is also a pregroup. By proposition 7, G is proper iff G = {1}. Theorem 2. No totally ordered left (right) pregroup is proper. Proof. We show: al a = 1 iff a ≥ all . First, al a = 1 iff al a ≥ 1, by (1). If al a ≥ 1, then all ≤ all al a ≤ a.If a ≥ all , then al a ≥ al all ≥ 1. In a similar way, one shows: aal = 1 iff a ≤ all . In a totally ordered left pregroup, a ≤ all or all ≤ a, for every element a. Consequently, all elements are injective or surjective, and we use proposition 7. By duality, the same holds for right pregroups. It is known [8] that there are no finite p.o. groups in which a < b holds, for some elements a, b. For a < b entails 1 < ba∪ , and 1 < c entails 1 < c < cc < ccc < . . .. This argument cannot be applied for pregroups. We nonetheless prove: Theorem 3. If G is a finite left (right) pregroup, then the ordering in G is the identity relation (hence, G is a group).
Pregroups: Models and Grammars
41
Proof. Let (G, ≤, ·, l, 1) be a finite left pregroup. The inequality xa ≤ b has a solution (x = bal ), for all a, b ∈ G. For a ∈ G, define a mapping fa (x) = xa. Let M denote the set of minimal elements of G. For any b ∈ M , there is x ∈ G such that fa (x) = b, hence there is x ∈ M such that fa (x) = b. Thus, fa maps a subset of M onto M . Since M is finite, this subset equals M , and fa restricted to M is a bijection of M onto itself. Consequently, xa ∈ M , for every x ∈ M . The inequality b ≤ ax has a solution (x = al b), for all a, b ∈ G. For a ∈ G, define a mapping ga (x) = ax. Let N denote the set of maximal elements of G. For any b ∈ N , there is x ∈ G such that ga (x) = b, hence there is x ∈ N such that ga (x) = b. As above, one shows that ga restricted to N is a bijection of N onto itself. Consequently, ax ∈ N , for every x ∈ N . Let a ∈ N . By the first paragraph, all minimal elements are of the form xa, for x ∈ M . By the second paragraph, these elements are maximal. We have shown that all minimal elements are maximal. Thus, the ordering is the identity relation. As a consequence, no finite left (right) pregroup can be proper. Below we obtain some results on the nonexistence of proper left or right pregroups of monotone functions on a poset. Lemma 1. Let (X, ≤) be a poset, and Y, Z ⊆ X be such that X = Y ∪ Z, Y ∩ Z = ∅. Let G be a left (right) pregroup contained in F (X, ≤) such that, for all f ∈ G, f [Y ] ⊆ Y and f [Z] ⊆ Z. Let GY consist of all functions f ∈ G restricted to Y , and GZ consist of all f ∈ G restricted to Z. Then, both GY and GZ are left (right) pregroups contained in F (Y, ≤) and F (X, ≤), respectively (here, ≤ is the restricted ordering). Proof. Clearly, both GY and GZ are closed under composition. Since I ∈ G, then IY ∈ GY and IZ ∈ GZ (here, IY is the identity function restricted to Y , and similarly for IZ ). If f ∈ G, then f l ∈ G, and consequently, f l restricted to Y belongs to GY . The condition f l (f (x)) ≤ x ≤ f (f l (x), for all x ∈ X, holds true in G, hence it also holds true in GY and GZ for f and f l restricted to Y and Z, respectively. So, GY and GZ admit adjoints. The direct product G × H of left (right) pregroups G, H is defined in the usual way. If G, H are left (right) pregroups, then G × H is a left (right) pregroup. In lemma 1, G is embeddable into GY × GZ . G is isomorphic to GY × GZ , if either y ≤ z, for all y ∈ Y , z ∈ Z, or neither y ≤ z, nor z ≤ y. for all y ∈ Y , z ∈ Z. A poset (X, ≤) is called a tree, if {y ∈ X : y ≤ x} is well ordered, for all x ∈ X. It is said to be well founded, if every nonempty subset of X has a minimal element. Every tree is well founded. Theorem 4. If (X, ≤) is well founded, then no right pregroup contained in F (X, ≤) is proper. Proof. We fix a well founded poset (X, ≤) and a right pregroup G contained in F (X, ≤). It is sufficient to show that all functions in G are bijective. We define a transfinite sequence of sets Mα . M0 is the set of all minimal elements in (X, ≤). For α > 0, if β 0, by xi we denote the relation inductively defined as follows: x1 = x, and xi+1 = x;xi . Definition 6. We define the calculus ω-CCFA as the extension of the CCFA obtained by adding the following inference rule: , 71 ≤y
xi ≤ y 7 xi+1 ≤ y 7 x∗ ≤ y
(i ∈ IN)
Definition 7. A model of the identities provable in ω-CCFA will be called an omega closure fork algebras. The class of omega closure fork algebras is denoted by ω-CFA. The intended (standard) models of the ω-CCFA are the Proper Closure Fork Algebras (PCFA for short). In order to define the class PCFA, we will first define the class of Pre Proper Closure Fork Algebras, denoted by •PCFA. Definition 8. Let E be a binary relation on a set U , and let R be a set of binary relations satisfying: / 1. R ⊆ E, 2. Id (the identity relation on the set A), ∅ (the empty binary relation) and E belong to R, 3. R is closed under set union (∪), intersection (∩) and complement relative to E (– ), 4. R is closed under relational composition (denoted by ◦), converse (denoted by & ), and reflexive-transitive closure (denoted by ∗ ). 5. / : U × U → U is one-to-one. 6. R is closed under the fork operator (∇), defined by the condition S ∇ T = { !x, y / z" : !x, y" ∈ S and !x, z" ∈ T } . 7. R is closed under
%
, the set choice operator defined by the condition:
x% ⊆ x and |x% | = 1
⇐⇒
x 3= ∅ .
Then, the structure ! R, U, ∪, ∩, – , ∅, E, ◦, Id,& , ∇, % , ∗ , / " is a •PCFA. Notice that x% denotes an arbitrary pair in x. This is why x% is called a choice operator. We will call the set U in Def. 8 the field of the algebra, and will denote the field of an algebra ÿ by Uÿ . Definition 9. We define the class PCFA as Rd • PCFA where Rd takes reducts to structures of the form ! R, ∪, ∩, – , ∅, E, ◦, Id,˘, ∇, % , ∗ ".
Interpretability of First-Order Dynamic Logic in a Relational Calculus
71
, , Notice that given ÿ ∈ PCFA, the terms (1 ∇1)˘ and (1∇1 )˘ denote respectively the binary relations { !a / b, a" : a, b ∈ Uÿ } and { !a / b, b" : a, b ∈ Uÿ }. Thus, they behave as projections with respect to the injection /. We will denote these terms by π and ρ, respectively. If we call splitting an object a ∈ Uÿ for which , there exist b, c ∈ Uÿ such that a = b / c, then 1 U is a part of the identity relation whose domain contains all the non–splitting objects. Ax. 9 states that the set on non–splitting objects (that we will call urelements) is nonempty. We denote the set of urelements of a PCFA ÿ by Urel ÿ . From the fork operator we define the binary operator ⊗ (cross) by the condition x⊗y = (π ;x) ∇ (ρ;y). When interpreted in an algebra ÿ ∈ PCFA, ⊗ behaves as a parallel product: x⊗y = { !a / b, c / d" : !a, c" ∈ x ∧ !b, d" ∈ y } . Definition 10. Given an algebra þ and a class of algebras K, þ is representable in K if there exists ÿ ∈ K such that þ is isomorphic to ÿ. If h is such an isomorphism, then the pair !ÿ, h" is a representation of þ in K. This notion generalizes as follows: a class of algebras K1 is representable in a class of algebras K2 if every member of K1 is representable in K2 . We finally present the representation theorem for omega closure fork algebras, which is reproduced and proved as Thm 3 in the Appendix. Theorem 1. The class ω-CFA is representable in PCFA.
4
Interpretability of DL in ω-CCFA
For the following definitions, σ and τ will be sequences of numbers increasingly sorted. Intuitively, these sequences will contain the indices of those variables that occur free in the formulas (or terms) being translated. Notation 1. – P os(n, σ) denotes the position of the index n in the sequence σ. – σ ⊕ n denotes the extension of the sequence σ with the index n. – σ(k) projects the element in the k-th position of σ. – If τ is a subsequence of σ, σ−τ denotes the sequence containing those indices in σ but not in τ . – Let τ and σ be disjoint and with Length(τ ) = l1 and Length(σ) = l2 . Let a = a1 / · · ·/ al1 and b = b1 / · · ·/ bl2 encode the values for the variables whose indices occur in τ and σ, respectively. M ergeτ,σ , from the input a / b, builds the encoding for the values of those variables whose indices occur either in τ or σ. Example: Let τ = [3, 5] and σ = [2, 4]. Let a = a1 / a2 and b = b1 / b2 . Then, M ergeτ,σ sends the object a / b to the object b1 / a1 / b2 / a2 . The relational expression for M ergeτ,σ (which of course depends on the specific τ and σ being considered) is (ρ;δσ (v2 )) ∇ (π ;δτ (v3 )) ∇ (ρ;δσ (v4 )) ∇ (π ;δτ (v5 )). – Finally, given a formula or term e, by σe we denote the sequence of all indices of variables with free occurrences in e.
72
M.F. Frias, G.A. Baum, and T.S.E. Maibaum
A relation x is right ideal if x = x;1. Intuitively, each element in the domain is related to all the elements in the universe. Definition 11. By ω-CCFA+ (Σ) we denote the extension of ω-CCFA obtained by adding the following equations as axioms. , , 1. 1 s ≤ 1 U (elements from s do not split). 2. In the first paragraph of Section 2 we assumed that actions modify finitely many state variables. If a ∈ A reads or modifies k state variables, we add the equation , , , , (1 s ⊗ · · · ⊗1 s );a;(1 s ⊗ · · · ⊗1 s ) = a . "! $ # "! $ # k–times k–times 3. For each f : sk → s ∈ F , we add the equation , , , f˘;(1 s ⊗ · · · ⊗1 s );f ≤ 1 s , "! $ # k–times stating that f is a functional relation of the right arity. 4. For each p ∈ P of arity k, we add the equation , , (1 s ⊗ · · · ⊗1 s );p;1 = p, "! $ # k–times stating that p is a binary right-ideal relation relating the right amount of inputs. Notice that given finite A, F and P , only finitely many equations are introduced in items 1–4 above. As before, we will assume a fixed but arbitrary signature Σ, and use the notation ω-CCFA+ as a shorthand for ω-CCFA+ (Σ). Definition 12. A model for ω-CCFA+ is a structure A = ! ÿ, mA " where ÿ ∈ ω-CFA and mA is the meaning function. When A is clear from context, we will note mA simply by m. The meaning function operates on the language of ω-CCFA+ as follows. 1. m(s) is a partial identity belonging to ÿ. 2. For each a ∈ A reading or modifying k variables, m(a) satisfies the typing condition in item 2 of Def. 11, and extends homomorphically to complex action terms. 3. For each f ∈ F , m(f ) ∈ ÿ is a functional relation. If f : sk → s, then m(f ) satisfies the typing condition in item 3 of Def. 11. 4. For each p ∈ P , m(p) is a right-ideal relation. If p has arity k, then m(p) satisfies the typing condition in item 4 of Def. 11. Definition 13. An ω-CCFA+ model ! ÿ, m " is full if ÿ has domain P (S × S) for some nonempty set S, and the operations have their intended set–theoretical meaning.
Interpretability of First-Order Dynamic Logic in a Relational Calculus
73
In the following paragraphs we will define a function mapping F or(Σ) to terms in ω-CCFA+ . The definition proceeds in two steps, because terms (as well as formulas) must be translated. Notation 2. We will denote by IT (S) the set of terms over individuals from DL that can be built from constant and function symbols belonging to S. Definition 14. The function δσ : IT (F ) → T (F ), mapping individual terms whose variables have indices in σ to relational terms, is defined inductively by the conditions: & ρP os(i,σ)−1 ;π if i is not the last index in σ , δσ (vi ) = ρLength(σ)−1 if i is the last index in σ . δσ (f (t1 , . . . , tm )) = (δσ (t1 )∇ · · · ∇δσ (tm ));f for each f ∈ F . Before defining the mapping Tσ translating F or(Σ), we need to define some auxiliary terms. For σ of length l, n ∈ IN and k = P os(n, σ ⊕ n), we define the term ∆σ,n by the condition1
ÿ
∆σ,n =
δσ (vσ(1) )∇ · · · ∇δσ (vσ(k−1) )∇1s ∇δσ (vσ(k) )∇ · · · ∇δσ (vσ(l) ) δσ (vσ(1) )∇ · · · ∇δσ (vσ(l) )∇1s
if k ≤ l, if k = l + 1.
The term ∆σ,n can be understood as a cylindrification [7, 8] in the k–th coordinate of an l-dimensional space. Given a subsequence, τ , of σ, of length k, by Πσ,τ we denote the term δσ (vτ (1) )∇ · · · ∇δσ (vτ (k) ) . This term, given an object storing values for the variables whose indices occur in σ, builds an object storing values for the variables occurring in τ . In the next definition we present functions Mσ and Tσ mapping action terms and formulas, respectively, to relational terms. The general assumption in the definition, is that whenever we apply Mσ to a program p, σ includes all the indices of state variables occurring in p. Similarly, if we apply Tσ to a formula α, then σ includes all the indices of variables with free occurrences in α. Definition 15. The functions Mσ and Tσ are mutually defined by 1. 2. 3. 4. 5. 6. 7. 8. 9. 1
Mσ (a) = (Πσ,σa ;a ∇ Πσ,σ−σa ) ;M ergeσa ,σ−σa , for each a ∈ A, ∗ Mσ (R∗ ) = Mσ (R) , Mσ (R ∪ S) = Mσ (R)+Mσ (S), Mσ (R;S) = Mσ (R);Mσ (S), , Mσ (α?) = Tσ (α)·1 , Tσ (p(t1 , . . . , tn )) = (δσ (t1 )∇ · · · ∇δσ (tn )) ;p, Tσ (¬α) = Tσ (α), Tσ (α ∨ β) = Tσ (α)+Tσ (β), Tσ ((∃vi ) α) = ∆σ,i ;Tσ⊕i (α),
, By 1s we denote the relation 1; 1 s .
74
M.F. Frias, G.A. Baum, and T.S.E. Maibaum
10. Tσ (!p" α) = Mσ (p);Tσ (α). Notation 3. We will denote by 7ω-CCFA the entailment relation in the calculus ω-CCFA. As usual, a formula α will be called a sentence if it does not contain any free occurrences of variables. The next theorem states the interpretability of presentations of theories from DL as equational theories in ω-CCFA. A detailed sketch of the proof is given in the Appendix, Thm. 8. Theorem 2. Let Γ ∪ { ϕ } be a set of sentences. Then, ' ) ⇐⇒ T!" (γ) = 1 : γ ∈ Γ 7ω-CCFA+ T!" (ϕ) = 1 . Γ |=DL ϕ
5
Conclusions
Since the interpretability of first-order dynamic logic allows one to reason about UML–like diagrams with dynamic content in an equational calculus, the results in this paper are at the heart of relational methods in computer science. The contributions of this paper can be summarized as follows: 1. 2. 3. 4. 5.
Definition of the class of omega closure fork algebras. Definition of the class of proper closure fork algebras. Proof of a representation theorem of ω-CFA in PCFA. Definition of a complete calculus for PCFA. Proof of an interpretability theorem of first-order dynamic logic in the calculus ω-CCFA.
Reasoning about properties of UML–like diagrams with dynamic content can be done in two ways. One way is by using formulas from dynamic logic to represent diagrams, and then translate the formulas to relational equations. The other way is by skipping the passage through dynamic logic and using relational equations directly.
References 1. Burris, S. and Sankappanavar, H.P., A Course in Universal Algebra, Graduate Texts in Mathematics 78, Springer–Verlag, 1981. 2. Frias M. F., Baum G. A. and Haeberer A. M., Fork Algebras in Algebra, Logic and Computer Science, Fundamenta Informaticae Vol. 32 (1997), pp. 1–25. 3. Frias, M. F., Haeberer, A. M. and Veloso, P. A. S., A Finite Axiomatization for Fork Algebras, Logic Journal of the IGPL, Vol. 5, No. 3, 311–319, 1997. 4. Frias, M. F. and Orlowska E., Equational Reasoning in Non–Classical Logics, Journal of Applied Non Classical Logic, Vol. 8, No. 1–2, 1998. 5. Gyuris, V., A Short Proof for Representability of Fork Algebras, Theoretical Computer Science, vol. 188, 1–2, pp. 211–220, 1997. 6. Harel, D., Kozen, D. and Tiuryn, J., Dynamic Logic, MIT Press, 2000.
Interpretability of First-Order Dynamic Logic in a Relational Calculus
75
7. Henkin, L., Monk, D. and Tarski, A., Cylindric Algebras Part I, Studies in Logic and the Foundations of Mathematics, vol.64, North–Holland, 1971. 8. Henkin, L., Monk, D. and Tarski, A., Cylindric Algebras Part II, Studies in Logic and the Foundations of Mathematics, vol.115, North–Holland, 1985. 9. Lyndon, R., The Representation of Relational Algebras, Annals of Mathematics (Series 2) vol.51 (1950), 707–729. 10. Maddux, R.D., Topics in Relation Algebras, Doctoral Dissertation, Univ. California, Berkeley, 1978, pp. iii+241. 11. Maddux, R.D., Finitary Algebraic Logic, Zeitschr. f. math. Logik und Grundlagen d. Math. vol. 35, pp. 321–332, 1989. 12. Maddux, R.D., Pair-Dense Relation Algebras, Transactions of the A.M.S., vol. 328, No. 1, pp. 83–131, 1991. 13. Maddux, R.D., Relation Algebras, Chapter 2 of Relational Methods in Computer Science, Springer Wien–New York, 1997. 14. Orlowska, E., Relational proof systems for modal logics. In: Wansing,H. (ed) Proof Theory of Modal Logics. Kluwer, pp. 55–77, 1996. 15. Rasiowa, H. and Sikorski, R. The Mathematics of Metamathematics. Polish Science Publishers, Warsaw, 1963. 16. Tarski, A., On the Calculus of Relations, Journal of Symbolic Logic, Vol. 6, 73–89, 1941. 17. Tarski, A. and Givant, S.,A Formalization of Set Theory without Variables, A.M.S. Coll. Pub., vol. 41, 1987.
76
M.F. Frias, G.A. Baum, and T.S.E. Maibaum
A
On ω-CCFA
Theorem 3. The class ω-CFA is representable in PCFA. Proof. Let ÿ ∈ ω-CFA. Let us consider the fork algebra reduct ÿ) of the algebra ÿ. By the representation theorem for fork algebras [3, 5] there exists a representation ! ) , h" of ) . Let us consider the structure = ! ) , % , ∗ ", where % and ∗ are defined by the conditions:
ÿ
þ
x% = h
ÿ
ÿ
*+ *+ -% , -∗ , h−1 (x) and x∗ = h h−1 (x)
þ
ÿ
for each x ∈
ÿ. )
→ is an ω-CFA isomorphism. It is also It is easy to check that h : ' ) easy to check using the infinitary rule that for x ∈ , x∗ = lub xi : i ∈ ω . Thus, using the isomorphism h, we can prove that for each x ∈ ,
þ
ÿ
' ) x∗ = lub xi : i ∈ ω .
(1)
ÿ
Notice that it is not necessarily the case that for x ∈ , x∗ equals the reflexivetransitive closure of x, since infinite sums may not correspond to the infinite union of the binary relations2 . Let be ’s relation algebra reduct. Since is point–dense (see [11, 12] for details on point-density), by [11, Thm. 8] has a complete representation3 ! ) , g". Since the representation is complete, g satisfies
ý ÿ
ý
ý
g (lub { xi : i ∈ I }) =
.
ý
g (xi ) .
i∈I
Let
ü = ! ý , ∇, )
% ∗
, ", where ∇,
%
and
∗
are defined as follows:
+ x∇y = g g −1 (x)∇g−1 (y) , *+ -% , x% = g g −1 (x) , *+ , ∗ x∗ = g g −1 (x) .
ÿ ü ü
Then, g : → is also an isomorphism. That ∇ as defined is a fork operation on follows from the fact relations g(π) and g(ρ) are a pair of quasi-projections [11, 17] in . Since % satisfies Ax. 4–Ax. 6, it is a choice operator. Finally, let us check that x∗ is the reflexive-transitive closure of x, for each x ∈ ÿ.
ü
2 3
ÿ
Examples of proper relation algebras in which lubA does not agree with A (for some subset A of the universe) are given in [9, 10] and elsewhere. A representation !ÿ, h" is complete if h maps infinite suprema into the corresponding infinite set unions.
Interpretability of First-Order Dynamic Logic in a Relational Calculus .
*+ -∗ , g −1 (x) (, * %+ -i = g lub g −1 (x) : i ≥ 0 . *+ -i , = g g −1 (x)
x∗ = g
77
(by Def. x∗ ) (by (1)) (g complete representation)
i≥0
=
.+ + --i g g −1 (x)
(g homomorphism)
i≥0
=
.
xi .
(g one-to-one)
i≥0
The last union computes the reflexive-transitive closure of the relation x, as was to be proved, and therefore ÿ ∈ PCFA. Finally, !ÿ, g ◦ h" is a representation of the ω-CFA þ in PCFA. 9
B
On the Interpretability of DL in the ω-CCFA+
Notation 4. Given a structure A = ! s, m ", a valuation of the individual variables ν, and þ ∈ PCFA such that Urelÿ ⊇ s, by sν,σ we denote the element a1 / · · · / ai / · · · / an ∈ Uÿ such that: 1. n = Length(σ), 2. ai = ν(vσ(i) ) for all i, 1 ≤ i ≤ n. In case σ = !", sν,σ denotes an arbitrary element from Uÿ . Definition 16. By the FullPCFA with set of urelements U (U a nonempty set), we refer to the algebra þ constructed as follows: 1. Let ! U ' , / " be the totally free groupoid with set of free generators U . 2. Let ÿ be the full algebra of binary relations with field U' . 3. Let þ = ! ÿ, ∇, % , ∗ " where ∇, % and ∗ are defined by: R∇S = { !x, y / z" : xRy ∧ xS z } . R% = { !a, b" } , with !a, b" ∈ R, arbitrary . R∗ = the reflexive-transitive closure of R . Definition 17. A ω-CCFA+ model ! þ, m " is full if
þ is a FullPCFA.
Theorem 4. Let α ∈ F or(S) and let A be a structure for Σ. Then there exists a full ω-CCFA+ model B = ! ÿ, m) " such that4 A, ν |=DL α 4
⇐⇒
sν,σα ∈ dom (m) (Tσα (α))) .
Given a binary relation R, by dom (R) we denote the domain set of R.
78
M.F. Frias, G.A. Baum, and T.S.E. Maibaum
Sketch of the Proof. Assume A = ! s, m ". Let us define B as follows. 1. 2. 3. 4.
Let ÿ be the FullPCFA with set of urelements s, m) (a) = { !sν,σa , sν ! ,σa " : !ν, ν ) " ∈ m(a) }, for each a ∈ A, m) (f ) = { !a1 / · · · / ai , b" : m(f )(a1 , . . . , ai ) = b }, for each f ∈ F , m) (p) = { !a1 / · · · / ai , b" : m(p)(a1 , . . . , ai ) and b ∈ Uÿ }, for each p ∈ P .
The proof continues by proving by simultaneous induction on the structure of programs and formulas the following properties. Let σ be an arbitrary sequence of indices. Then: a) for each R ∈ RT (Σ) with σ extending σR , m) (Mσ (R)) = { !sν,σ , sν ! ,σ " : !ν, ν ) " ∈ m(R) } . b) for each β ∈ F or(S) and σ extending σβ , dom (m) (Tσ (β))) = { sν,σ : A, ν |=DL β } . Finally, the theorem follows from b) choosing σ to be σα .
9
Corollary 1. Let α ∈ F or(Σ) be a sentence, and let A be a structure for Σ. Then there exists a full ω-CCFA+ model B = ! ÿ, m " such that + A |=DL α ⇐⇒ m T!" (α) = 1 . Proof. A |=DL α ⇐⇒ A, ν |=DL α for each ν + + -⇐⇒ sν,!" ∈ dom m T!" (α) for each ν .
(by Def. |=DL ) (by Thm. 4)
Since sν,!" denotes an arbitrary element from Uÿ , A |=DL α ⇐⇒ e ∈ dom (m (T0 (α))) for each e ∈ Uÿ . + Finally, since m T!" (α) is a right-ideal relation, + A |=DL α ⇐⇒ m T!" (α) = 1 . 9 Notation 5. Let þ ∈ PCFA, s = a1 / · · · / ak (aj ∈ Urel þ for all j, 1 ≤ j ≤ k), and i such that 0 ≤ i ≤ k. By νs,i we denote the set of valuations of individual variables ν satisfying ν(vj ) = aj for all 1 ≤ j ≤ i. In case i = 0, for each s ∈ Uþ νs,i denotes the set of all valuations. Theorem 5. Let α ∈ F or(Σ) and A = ! þ, m " be a full ω-CCFA+ model. Then there exists a structure B such that s ∈ dom (m (Tσα (α)))
⇐⇒
B, ν |=DL α for all ν ∈ νs,σα .
Interpretability of First-Order Dynamic Logic in a Relational Calculus
79
Sketch of the Proof. Let B = ! s, m) " with 1. s = Urel ÿ , 2. (a) For each a ∈ A, m) (a) = {!ν, ν ) " : !sν,σα , sν ! ,σα " ∈ m(a)} . (b) For each f : sk → s ∈ F , m) (f )(a1 , . . . , ak ) = b
iff
!a1 / · · · / ak , b" ∈ m(f ) .
(c) For each p ⊆ sk ∈ P , !a1 , . . . , ak " ∈ m) (p)
iff
a1 / · · · / ak ∈ dom (m(p)) .
The remaining part of the proof follows by induction on the structure of terms and formulas and is analogous to the proof of Thm. 4. 9
Corollary 2. Let α ∈ F or(S) be a sentence, and let A = ! ÿ, m " be a full ω-CCFA+ model. Then there exists a structure B such that + ⇐⇒ m T!" (α) = 1 . B |=DL α Proof. Let B as in Thm. 5. Then B |=DL α ⇐⇒ B, ν |=DL α for each ν (by Def. |=DL ) ⇐⇒ for each s ∈ Uÿ , B, ν |=DL α for each ν ∈ νs,!" (by Def. νs,!" ) + + -⇐⇒ s ∈ dom m T!" (α) for each s ∈ Uÿ (by Thm. 5) + + -⇐⇒ dom m T!" (α) = Uÿ (by Def. dom) + ⇐⇒ m T!" (α) = 1 . (by T!" (α) right-ideal) 9 Notation 6. We will denote by |=Full the semantic consequence relation on the class of full ω-CCFA models. Theorem 6. Let Γ ∪ { ϕ } be a set of sentences. Then, ' ) Γ |=DL ϕ ⇐⇒ T!" (γ) = 1 : γ ∈ Γ |=Full T!" (ϕ) = 1 . Proof. In order to prove the theorem, it suffices to show that ' ) Γ 3|=DL ϕ ⇐⇒ T!" (γ) = 1 : γ ∈ Γ 3|=Full T!" (ϕ) = 1 .
(2)
Formula (2) follows from Cors. 1 and 2.
9
Theorem 7. Let V be the variety generated by FullPCFA. Then, V = ω-CFA. Proof. From Thm. 1, every ω-CFA is isomorphic to some PCFA. A proof similar to the one showing that proper relation algebras are isomorphic to subalgebras of products of full proper relation algebras, shows that5 PCFA ⊆ ISPFullPCFA. Thus, ω-CFA ⊆ ISPFullPCFA ⊆ HSPFullPCFA. Since FullPCFA satisfies the equations defining ω-CFA and equations are preserved by H, S and P, HSPFullPCFA ⊆ ω-CFA. 9 5
Given a class of algebras K, by IK we denote the closure of K under isomoprhisms. By HK, the closure of K under homomorphic images. By SK, the closure of K under subalgebras. By PK, the closure of K under direct products.
80
M.F. Frias, G.A. Baum, and T.S.E. Maibaum
Notation 7. Given a class of algebras K, by EqT h(K) we denote the equational theory of K, i.e., the set of all valid equations in K. Theorem 8. Let Γ ∪ { ϕ } be a set of sentences. Then, ' ) ⇐⇒ T!" (γ) = 1 : γ ∈ Γ 7ω-CCFA T!" (ϕ) = 1 . Γ |=DL ϕ Proof. By Thm. 6, Γ |=DL ϕ
⇐⇒
'
T!" (γ) = 1 : γ ∈ Γ
)
|=Full T!" (ϕ) = 1 .
It then suffices to show that given a set of equations E ∪ { e }, E |=Full e
⇐⇒
E 7ω-CCFA e .
⇒) E |=Full e ⇒ E |=ω-CFA e ⇒ E ∪ EqTh (ω-CFA) |= e ⇒ E ∪ EqTh (ω-CFA) 7 e .
(by Thm. 7) (by Def. ω-CFA) (by completeness of eq. logic)
Notice now that an equational proof of e from E ∪ EqTh (ω-CFA) will be a finite height tree whose leaves are equations in EqTh (ω-CFA). In order to show that E 7ω-CCFA e, it suffices to replace each leaf in EqTh (ω-CFA) by its corresponding proof in ω-CCFA. ⇐) Let us proceed by induction on the height of proof trees. • If the tree has height 1, then e ∈ E or e is an axiom of ω-CCFA. Clearly, E |=Full e. • Assume the implication holds if there is a proof of an equation e) from E of height n, and let Π be a proof of e from E of height n + 1. The only nontrivial case is when Π is of the form Π ,0 71 ≤y
···
Πi+1 Πi 7 xi ≤ y 7 xi+1 ≤ y 7 x∗ ≤ y
···
By inductive hypothesis, , E |=Full 1 ≤ y, . . . , E |=Full xi ≤ y, E |=Full xi+1 ≤ y, . . . , which implies E |=Full x∗ ≤ y.
9
Relations in GUHA Style Data Mining Petr H´ajek Institute of Computer Science AS CR, Pod Vod´ arenskou vˇeˇz´ı 2, Prague 8, Czech Republic
Abstract. The formalism of GUHA style data mining is confronted with the approach of relational structures of Orlowska and others. A computational complexity result on tautologies with implicational quantifiers is presented.
1
Introduction
The first aim of the present paper is to compare the logic of observational calculi, presented is depth in the monograph [5] and serving as foundation for the GUHA method of automated generation of hypotheses, (see e.g. [6,7]) with the formalism of relational structures as presented in [11] (see e.g. Chapter 16 by D¨ untsch and Orlowska). Our main aim is to contribute to “building bridges” between these two approaches by showing some possiblities of mutual influence and application. For this purpose we survey, in Section 2,various kinds of relations emerging in this context and present, as an illustration, the notion of an i-algebra (Boolean algebra with an implicational relation) and formulate a corresponding representability problem. The second aim is to present a result on the computational complexity of an important class of logical formulas called implicational tautologies, i.e. formulas built using a binary quantifier ⇒∗ and logically true for each interpretation of ⇒∗ as an implicational quantifier (in the sense of GUHA): the set of all such tautologies is shown to be co-NP-complete. The two parts of the paper can be read independently of each other; the reader not familiar with GUHA finds in the first part all definitions needed for the second part. Interestingly, the second part makes non-trivial use of mathematical fuzzy logic. Support of COST Action 274 TARSKI (Theory and Applications of Relational Structures as Knowledge Instruments) is acknowledged. The second part of the paper is also relevant for the grant project A13004/00 of the Grant Agency of the Academy of Sciences of the Czech Republic. Thanks are due to an anonymous referee, whose remarks helped to improve the paper.
2
Relations – Where Are They from
One basic notion common to both approaches is that called information system U in the terminology of [9,10,12,13] and data matrix or observational model in GUHA. It is given by a finite non-empty set U of objects, a finite non-empty set A of attributes, each a ∈ A having a finite domain Va and an evaluation H. de Swart (Ed.): RelMiCS 2001, LNCS 2561, pp. 81–87, 2002. c Springer-Verlag Berlin Heidelberg 2002
82
P. H´ ajek
function fa assigning to each u ∈ U an element fa (u) ∈ Va . More generally, one may assume fa (u) ⊆ Va ; less generally, one may assume fa (u) ∈ {0, 1}, i.e. Va = {0, 1} for all a. Obviously, the most general notion is reducible to particular case of {0, 1}-valued attributes by an appropriate coding; for simplicity we shall restrict ourselves to this particular case.1 Without loss of generality assume U = {1, . . . , m} and A = {1, . . . , n}. Then the system U is obviously coded by a matrix {ui,j }j=1,...,n i=1,...,m of zeros and ones where ui,j = fj (i). Name the j-th column of the matrix by a predicate Pj . GUHA uses monadic predicate calculus with generalized quantifiers. Open formulas are built from atomic formulas Pj (x) (where Pj is a unary predicate and x is a variable) using logical connectives; for open formulas ϕ, ψ containing just one free variable x, the formula ϕ ∼ ψ is built using the quantifier ∼ (read ϕ ∼ ψ “ϕ is associated with ψ”). Given such a matrix U with columns named by predicates P1 , . . . , Pn , the ith object satisfies the formula Pj (x) iff ui,j = 1, i.e. the value in the i-th row and j-th column is 1. For each data matrix U let a, b, c, d be the number of objects satisfying ϕ&ψ, ϕ&¬ψ, ¬ϕ&ψ, ¬ϕ&¬ψ respectively. The quadruple (a, b, c, d) is commonly called the four-fold table of ϕ, ψ in U) and can be presented as follows: ψ ¬ψ ϕ a b r ¬ϕ c d s k l m (Here r = a + b, s = c + d, k = a + c, l = b + d are the marginal sums; r + s = k + l = a + b + c + d = m.) The semantics of ∼ is given by a truth function tr∼ assigning to each (a, b, c, d) the truth value tr∼ (a, b, c, d) ∈ {0, 1}. A quantifier ∼ is associational if a1 ≥ a2 , b1 ≤ b2 , c1 ≤ c2 , d1 ≥ d2 and tr∼ (a1 , b1 , c1 , d1 ) = 1 implies tr∼ (a2 , b2 , c2 , d2 ) = 1. It is implicational if a1 ≥ a2 , b1 ≤ b2 and tr∼ (a1 , b1 , c1 , d1 ) = 1 implies tr∼ (a2 , b2 , c2 , d2 ) = 1. (Thus each implicational quantifier is associational.) Some examples are presented in Sect. 3. The GUHA software works with particular associational/implicational quantifiers making use of their theory. It generates formulas ϕ ∼ ψ true in the data. Orlowska’s approach uses information systems to define binary relations on U, particularly information relations of similarity and diversity, e.g. 1
But note that this is equivalent to the assumption of having just one attribute a ˆ and faˆ (u) ⊆ Vaˆ is the item set in the terminology of [1] - the set of attributes for which u has the value 1.
Relations in GUHA Style Data Mining
83
uRv iff u, v have the same attributes, uR v iff u, v have no attribute in common. For deep modal and algebraic aspects of this approach see [11]. Our question now reads: does the GUHA approach lead to some interesting relations? Let us offer three possibilities. First, note that each information system, i.e. data matrix ({0, 1}-valued) can be seen as a binary relation, subset of U × A, namely (u, a) ∈ R iff fa (u) = 1. Evidently, this relation uniquely extends to a relation on U × F orm, where F orm is the set of all open formulas built from atoms P1 (x), P2 (x), . . .. Each ntuple {ϕ1 , . . . , ϕn } of such open formulas determines the corresponding relation on U × {ϕ1 , . . . , ϕn } – the truth table of ϕ1 , . . . , ϕn given by U. Note that this representation is very near to that used in Agrawal’s kind of mining associational rules [1] mentioned above. See also [14,2]. Second, assume a data matrix U and a quantifier ∼ to be fixed; this defines a binary relation HYP on F orm, namely of all pairs ϕ, ψ such that ϕ ∼ ψ is true in U. If ∼ is associational then HYP represents all pairs (ϕ, ψ) of open formulas (composed properties) such that ϕ is associated with ψ. GUHA generates subrelations of this relation given by syntactical conditions on ϕ, ψ (think e.g. on ϕ as a combination of symptoms and ψ a combination of diseases or so). Let us stress the importance of logical rules (deduction rules) for an optimal representation of true hypotheses; for example if ∼ is symmetric (ϕ ∼ ψ logically implies ψ ∼ ϕ) one can make use of it. Third, let U and ∼ be as before; instead of pairs of formulas let us think of pairs of subsets of U definable by open formulas. We get a binary relation on the Boolean algebra BU of definable subsets of U; properties of the quantifier ∼ (for example ∼ being an implicational quantifier) determine properties of the relation. In more details, for u, v ∈ BU let auv = card(u ∩ v), buv = card(u ∩ −v), cuv = card(−u ∩ v), duv = card(−u ∩ −v). Let tr∼ be the truth function of an implicational quantifier ∼ and put RU = {(u, v)|tr∼ (auv , buv , cuv , duv ) = 1}. This leads us to the following Definition 1. A Boolean algebra with an implicational relation (briefly, an iBoolean algebra) is a structure B = (B, ∧, ∨, −, R) where (B, ∧, ∨, −) is a Boolean algebra and R ⊆ B ×B is a relation such that for each u1 , u2 , v1 , v2 ∈ B, if u1 ∧ v1 ≤ u2 ∧ v2 , u1 ∧ −v1 ≤ u2 ∧ −v2 and (u1 , v1 ) ∈ R then (u2 , v2 ) ∈ R. Clearly, for each U and each implicational quantifier ∼, the algebra of definable subsets of U together with the relation HYP as above is an i-Boolean algebra. This leads us to the following problem: Is each (finite) i-Boolean algebra isomorphic to the algebra BU given by a data matrix U and an implicational quantifier ∼? This may be easy to decide; but
84
P. H´ ajek
it seems at least to show that our approach may offer some possibly interesting purely algebraic problems. The methods of the next section seem to be useful for solution of the present problem.
3
Complexity of GUHA-Implicational Quantifiers
Implicational quantifiers2 were defined in the preceding section; for full treatment see [5]. Note that implicational quantifiers are usually denoted by the symbol ⇒∗ and by similar symbols. Just recall the following examples: – classical implicational quantifier – ϕ ⇒ ψ is (∀x)(ϕ(x) → ψ(x)), i.e. ϕ ⇒ ψ is true iff b = 0; – founded implication ϕ ⇒p,s ψ is true iff a ≥ s and a ≥ p(a + b) where s is a positive natural number and 0 < p ≤ 1, ψ is true iff – lower critical implication: ϕ ⇒L a+b a+b ip,α LIM P Lp (a, b) = i=a i p (1 − p)a+b−i ≤ α For example, LIM P L0.9 (240, 3) = 2.76.10−8 . Consider the language with unary predicates P1 , P2 , . . . and a binary quantifier ⇒∗ (as well as object variables and logical connectives). Let x be an object variable called the designated variable. A pure prenex formula has the form ϕ ⇒∗ ψ, where ϕ and ψ are quantifier-free open formulas containing no variable except x (i.e. Boolean combinations of formulas P1 (x), P2 (x), . . .). It is understood that ⇒∗ binds the variable x; the pedantical writing would be (⇒∗ x)(ϕ(x), ψ(x)). The normal form theorem (see [5] 3.1.30) says that each closed formula of our language is logically equivalent to a Boolean combination of pure prenex formulas. Note the the proof of this fact is constructive and uniform: given a closed formula Φ one finds a normal form N F (Φ) which is a (possibly empty) disjunction of elementary conjunctions of pure prenex formulas and is logically equivalent to Φ for any semantics of the quantifier ⇒∗ . It follows that if you have only finitely many predicates (atributes) P1 , . . . , Pn then there is a finite set NF of normal form formulas such that for each Φ, N F (Φ) can be taken from NF. Call Φ an implicational tautology (or a tautology with implicational quantifiers) if Φ is true in each data matrix U (of appropriate arity) for each implicational quantifier ⇒∗ . Φ is implicationally satisfiable (briefly, i-satisfiable) if there is a data matrix U (interpreting predicates occuring in Φ) and an implicational quantifier ⇒∗ such that Φ is true in U w.r.t ⇒∗ . Examples of implicational tautologies (easy to verify): ((ϕ&ψ) ⇒∗ χ) → (ϕ ⇒∗ (χ ∨ ¬ψ)), (ϕ ⇒∗ (ψ&χ)) → ((ϕ&ψ) ⇒∗ χ). (Compute the corresponding four-fold tables.) 2
An alternative name is multitudinal quantifiers, cf. [4].
Relations in GUHA Style Data Mining
85
The normal form theorem gives a cheap decidability result: For each fixed n, the set of all implicational tautologies containing no predicate except P1 , . . . , Pn , is decidable (since the set of the corresponding normal forms is finite, cf. [5] Chapter III, Problem 9). We are going to show decidability and determine the computational complexity of the set of implicational tautologies having arbitrary many predicates. Since a formula is an implicational tautology iff its negation is not i-satisfiable, everything is solved by the following Theorem 1. The set all of implicationally satisfiable formulas in normal form is NP-complete. Proof. The reader is assumed to know basics notions and facts of polynomial complexity theory as: a set X of formulas (or words in an alphabet) is in NP (accepted by a non-deterministic Turing machine working in polynomial time); X is NP-complete (is in NP and each Y in NP is reducible to X in polynomial time); the set of satisfiable formulas of propositional logic is NP-complete and, in more details, the function assigning to each propositional formula A and each evaluation of its atoms by zeros and ones the corresponding truth value of A is computable in polynomial time. See e.g. [3] for details. Formulas in question result from propositional formulas A(p1 , . . . , pn ) and pure prenex formulas Φ1 , . . . Φn (Φi being ϕi ⇒∗ ψi ) as A(Φ1 , . . . , Φn ) (substituting Φi for pi in A(p1 , . . . pn )). We describe a non-deterministic algorithm accepting exactly all implicationally satisfiable formulas; checking that it works in polynomial time is routine. This will show that the set of all implicationally satisfiable formulas are in NP; then we show NP-completeness. Given A(p1 , . . . , pn ) and Φ1 , . . . Φn , first guess an evaluation e assigning to each pi a truth value e(pi ) ∈ {0, 1} and check if e makes A(p1 , . . . pn ) true. (If not, fail.) If it does, we want to find a data matrix (information structure) U and an implicational quantifier ⇒∗ such that U respects e for ⇒∗ , i.e. for i = 1, . . . , n, Φi is true in U iff e(pi ) = 1. If all this succeeds then A(Φ1 , . . . , Φn ) is implicationally satisfiable and vice versa. Call Φi positive if e(pi ) = 1, and negative if e(pi ) = 0. Recall that Φi is ϕi ⇒∗ ψi . If U and ⇒∗ are given then only the frequency of ϕi &ψi and of ϕi &¬ψi (number of objects satisfying the respective formula) decide of Φi is true in U or not. Call all the formulas ϕi &ψi , ϕi &¬ψi (i = 1, . . . , n) critical formulas. Their frequencies (in U) define a linear preorder ≤U on them: for any critical γ1 , γ2 , we put γ1 ≤U γ2 if f reqU (γ1 ) ≤ f reqU (γ2 ). For an arbitrary linear preorder ≤ of the critical formulas, call ≤ acceptable for e of the following holds for any 1 ≤ i, j ≤ n : whenever ϕi ⇒∗ ψi is positive (with respect to e), ϕj &ψj ≥ ϕi &ψi and ϕj &¬ψj ≤ ϕi &¬ψi then ϕj ⇒∗ ψj is also positive. From the definition of an implicational quantifier it follows immediately that the above preorder ≤U is acceptable (for any U, ⇒∗ ). Thus we continue as follows: Having e, guess an acceptable linear preorder ≤ of critical formulas and ask whether there is an U such that ≤ is ≤U (say then that ≤ is realizable). We show that this can be done in non-deterministic polynomial time. To do this we use the result of [8] on the complexity of fuzzy probabilistic logic over L ukasiewicz propositional calculus. (This calculus is also described in [4].) For each critical formula γ, let P(γ) be the formula saying “γ is probable”. Let
86
P. H´ ajek
γ1 , . . . , γn be a sequence of all critical formulas non-decreasing with respect to ≤. Let ε be a new atomic formula. For each 1 ≤ i < k, if γi < γk let Λi be the ukasiewicz strong disjunction, formula (P(γi ) ⊕ P(ε) → P(γi+1 ) (where ⊕ is L also denoted by ∨); if γi ≤ γi+1 and γi+1 ≤ γi then Λi is P(γi ) ≡ P(γi+1 ). Let T be the finite theory over FPL whose axioms are all the Λi (i = 1, . . . k−1) and also P (γi0 ) → P (¬ε) where i0 is the largest index such that γio < γk . (The singular case that all γi are ≤-equivalent is left to the reader as an exercise.) Observe that there is a probability on open formulas built from predicates occuring the critical formulas and coherent with ≤ iff the theory T has a model over FPL in which P(ε) has a non-extremal value, i.e. P(ε) ∨ ¬P(ε) has not value 1. Using the method of [8] and [4] one easily reduces this problem to a Mixed Integer Programming problem, showing that the last problem is in NP. Three things remain: First, to show that if such probability exists then we may assume it has rational values on all open formulas from our predicates – this is done as in [4] 8.4.16. Thus multiplying by the common denominator we may get a finite model alias data matrix alias information system U such that frequencies of objects satisfying critical formulas order them in accordance with ≤ . Second, let (ai , bi , ci , di ) be four-fold tables of pairs (ϕi , ψi ) of formulas occuring in the sentences ϕi ⇒∗ ψi ; define a quantifier ⇒∗ by letting tr⇒∗ (a, b, c, d) = 1 iff for some i, a ≥ ai , b ≤ bi , c ≤ ci and d ≥ di . This is an implicational quantifier and U respects e for this ⇒∗ . This shows that our problem of satisfiability of Φ) is in NP. The last (third) thing is to show NP-hardness. To this end we reduce the satisfiability problem of propositional logic (equivalently, satisfiability of open formulas built from atom P1 (x), . . . , Pn (x), . . .) to our problem. Let ε be as before; for each open formula ϕ as above let ϕ∗ be (ε ⇒∗ ε) → (ε ⇒∗ ϕ). Let us show that ϕ is a Boolean tautology logic iff ϕ∗ is an implicational tautology; thus ϕ is satisfiable in Boolean logic iff ¬(¬ϕ)∗ is satisfiable in our logic with an implicational quantifier. Indeed, if ϕ is a Boolean tautology and (a, b, c, d) is a four-fold table of (ε, ϕ) given by an U; then b = d = 0; if in the same U the formula ε ⇒∗ ε is true for an implicational quantifier ⇒∗ then tr⇒∗ (a, 0, 0, c) = 1 (note that (a, 0, 0, c) is the four-fold table of , in U) and hence also tr⇒∗ (a, 0, c, 0) = 1. Then ⇒∗ ϕ is true and hence ϕ∗ is true in our U. Conversely, if ϕ is not a Boolean tautology then take an U in which all objects satisfy ¬ϕ and all satisfy ε, thus ε ⇒∗ ε is true and ε ⇒∗ ϕ is false for the classical implicational quantifier ⇒∗ (γ ⇒∗ δ being (∀x)(γ → δ)). This completes the proof. Corollary 1. The set of all implicational tautologies is co-NP-complete.
Remark 1. The same idea can be used to show that the problem of associational satisfiability (more precisely, the problem of showing that a formula Φ of the language with one binary quantifier ∼ in normal form, is satisfiable in a U by an associational quantifier) is in NP. The reader may try to show NP-completeness (possibly easy).
Relations in GUHA Style Data Mining
87
Conclusion. We hope to have shown that the logic of observational calculi as calculi speaking on data (alias information systems) is interesting and relevant for the study of relational structures and have contributed to the study of its computational complexity.
References 1. Agrawal R., Mannila H., Srikant R., Toivonen H., Verkamo A. I.: Fast discovery of association rules. In: (Fayyad V. et. al, ed.) Advances in knowledge discovery and data mining, AAA Press/MIT Press 1996, pp. 307–328. 2. D¨ untsch I., Gediga G., Orlowska E.: Relational attribute systems, International Journal of Human Computer Studies, 55 (2001) pp. 293-309. 3. Garrey M. R., Johnson D. S: Computers and intractability, New York, W. J. Freeman and Co., 1979. 4. H´ ajek P.: Metamathematics of fuzzy logic, Kluwer 1998. 5. H´ ajek P., Havr´ anek T.: Mechanizing hypothesis formation – Mathematical foundations for a general theory, Springer Verlag 1978. Internet version (free): www.cs.cas.cz/∼ hajek/guhabook 6. H´ ajek P., Holeˇ na M.: Formal logics of discovery and hypothesis formation by machine. In: (Arikawa et al, ed.) Discovery Science, Springer Verlag 1998, 291-302 7. H´ ajek P., Sochorov´ a A., Zv´ arov´ a J.: GUHA for personal computers. Comp. Statistics and Data Analysis 19(1995), 149–153. 8. H´ ajek P., Tulipani S.: Complexity of fuzzy probability logic. Fundamenta Informaticae 45 (2001), 207–213. 9. Lipski W.: Informational systems with incomplete information. In Michaelson S., Milner R., editors. Third International Colloquium on Automata, Languages and Programming, University of Edinburgh, Edinburgh University Press, 1976, pp. 120-130. 10. Orlowska E., Pawlak Z: Logical foundationas of knowledge representation, ICS Research Report 573, Polish Academy of Sciences, 1984. 11. Orlowska E., Szalas A. (ed.): Relational methods in computer science applications, Springer-Physica Verlag 2001, 263–285. 12. Pawlak Z.: Mathematical foundations of information retrieval, ICS Research Report 101, Polish Academy of Sciences, 1973. 13. Pawlak Z.: Rough sets. Internat. Journal Comput. Inform. Sci., 11 (1982) pp. 341356. 14. Vakarelov D.: Information systems, similarity relations and modal logics. In: Orlowska E., editor, Incomplete Information – Rough Set Analysis, Heidelberg, Physica-Verlag 1997, pp. 492-550.
Groups in Allegories Yasuo Kawahara Department of Informatics, Kyushu University 33, Fukuoka 812-8581, Japan
[email protected] Abstract. Groups are one of the most fundamental notions in mathematics. This paper provides a foundation of group theory in allegories. Almost all results in the paper can be applied to theory of fuzzy groups.
1
Introduction
The motivation of the paper arose from the following three fundamental exercises in group theory: (a) Let G = (G, ·) be a semigroup, that is, a binary operation · : G × G → G is associative. Show that if ∀x ∈ G : xe = x and ∀x ∈ G∃y ∈ G : xy = e for some element e ∈ G, then ∀x ∈ G : ex = x and ∀x ∈ G∃y ∈ G : yx = e. (b) Let H be a subgroup of a group G = (G, ·, e, ·−1 ). Prove that the (binary) relation x ≡ y on G, defined by x−1 y ∈ H, is an equivalence relation. (c) Show that the set of all normal subgroups of a group G forms a modular lattice. In other words, the following modular law holds for three normal subgroups S, T and U of G: S ⊆ U =⇒ ST ∩ U ⊆ S(T ∩ U ). To define group objects in categories [9] is not new, but it is difficult to directly treat algebraic relations, such as residual relations (b) induced by subgroups, in the ordinary category theory. Allegories [4], as a kind of relation categories, give a natural and suitable setting for manipulating algebraic binary relations in group theory, lattice theory [8] and so on. This paper provides a foundation of group theory in allegories, and solves the above three questions (a), (b) and (c). The paper is organised as follows: In section 2 we recall the definition of allegories [4] and remark some fundamentals on relational products in allegories. In section 3 we review a simple sharpness property [5,10,6] on relational products of relations, which was initiated by Schmidt and will play an important rˆ ole in the proof of the main results. In section 4 we explore some fundamental properties of relational binary operations and show a suitable formalisation (Theorem 2) of the associative law, as well as the commutative law, using with relational global elements. The inverse law and the absorption laws (in lattice theory) are of course not the case. In section 5 we mention unitary and inverse operations for binary operations in allegories. We also give a relational version (Theorem 4) for H. de Swart (Ed.): RelMiCS 2001, LNCS 2561, pp. 88–103, 2002. c Springer-Verlag Berlin Heidelberg 2002
Groups in Allegories
89
the first question (a). In section 6 we define notions of (functional) groups and subgroups, and prove an elementary fact (b) that the residual relation induced by a subgroup is an equivalence relation. In section 7 we describe a notion of normal subgroups in allegories, and answer the final question (c) that the set of all normal subgroups of a group forms a modular lattice.
2
Allegories
In this section we recall the fundamentals on relation categories, called allegories [4]. Throughout this paper, a morphism α from an object X into an object Y in an allegory (which will be defined below) will be denoted by a half arrow α : X Y , and the composite of a morphism α : X Y followed by a morphism β : Y Z will be written as αβ : X Z. Also we will denote the identity morphism on X as idX . Definition 1. An allegory A is a category satisfying the following: D1. [Meet Semi-Lattice] For all pairs of objects X and Y the hom-set A(X, Y ) consisting of all morphisms of X into Y is a meet semi-lattice with the greatest morphism ∇XY . Its semi-lattice structure will be denoted by A(X, Y ) = (A(X, Y ), , , ∇XY ). D2. [Converse] There is given a converse operation : A(X, Y ) → A(Y, X). That is, for all morphisms α, α : X Y , β : Y Z, the following converse laws hold: (a) (αβ) = β α , (b) (α ) = α, (c) If α α , then α α . D3. [Dedekind Formula] For all morphisms α : X Y , β : Y Z and γ : X Z the Dedekind formula αβ γ α(β α γ) holds. D4. [Sub-Distributivity] The composition preserves order: If α α and β β , then αβ α β . The fundamental properties of relational categories is referred to [1,4,11,7]. The following is a basic property of allegories. Proposition 1. Let α : X Y and γ : Y X be morphisms in an allegory A. If αγ = idX and γα = idY , then α = γ . A unit I in an allegory A is an object such that idI = ∇II . A morphism α : X Y is total if idX αα (or equivalently, α∇Y X = ∇XX ). A morphism f : X Y such that f f idY (univalent) is called a function and may be introduced as f : X → Y . In what follows the word relation is used as synonym for morphisms in allegories. Definition 2. A pair (A, B) of objects in an allegory A has a relational product if there exists a pair (p : T → A, q : T → B) of total functions such that p q = ∇AB and pp qq = idT . The pair (p : T → A, q : T → B) is called a pair of projections for (A, B). An allegory A has a relational product if every pair of objects in A has a relational product.
90
Y. Kawahara
Let (p : T → A, q : T → B) be a pair of projections for a pair (A, B) of objects. For each pair of relations α : X A and β : X B, we define a relation α β : X T by α β = αp βq . It is trivial that p q = idT .
The following proposition is a list of elementary properties of relational products in allegories. The proof is trivial and so omitted. Proposition 2. Let (p : T → A, q : T → B) be a pair of projections for (A, B) and let α, α : X A, β, β : X B, γ : Y A, δ : Y B and ξ : Z X be relations.
Then the following statements hold: If α α and β β , then α β α β , ξ(α β) ξα ξβ, ξ(α ∇XY δ) = ξα ∇ZY δ, ξ(∇XY γ β) = ∇ZY γ ξβ, If ξ : Z X is a function, then ξ(α β) = ξα ξβ. If α and β are total functions, then α β is a unique total function such that (α β)p = α and (α β)q = β, (g) (α α ) β = (α β) (α β) and α (β β ) = (α β) (α β ).
(a) (b) (c) (d) (e) (f)
As in ordinary category theory, the common domain T of projections p : T → A and q : T → B is uniquely determined up to isomorphism by the virtue of the last Proposition 2(f). This enables us to write the object T as A × B and the pair of projections for (A, B) as (p, q), if it exists. Let (p, q) be a pair of projections for (A, A). The twisting function t : A × A → A × A is defined by t = q p (= qp pq ). That is, t is a unique total function such that tp = q and tq = p. Then tt = t(q p) = tq tp = p q = idA×A .
Groups in Allegories
91
In addition we assume that (p1 , q1 ) and (p2 , q2 ) are pairs of projections for (A × A, A) and (A, A × A), respectively. Ao
q
p
A ×O A
/A O
p1
p2
(A × A) × A
a
/ A × (A × A)
q
A×A
q1
q2
Ao
p
/A
The associative function a : (A × A) × A → A × (A × A) is defined by a = p1 p (p1 q q1 ). That is, a is a unique total function such that ap2 = p1 p, aq2 p = p1 q and aq2 q = q1 . Another associative function b : A × (A × A) → (A × A) × A is defined by b = (p2 q2 p) q2 q. That is, b is a unique total function such that bp1 p = p2 , bp1 q = q2 p and bq1 = q2 q. It is trivial that a and b are mutually inverses, that is, ab = id(A×A)×A and ba = idA×(A×A) , and consequently b = a by Proposition 1. It is easy to see that a pair of total functions p2 = p1 p and q2 = p1 q q1 is a pair of projections for (A, A × A). This fact indicates that if a relational product for (A × A, A) exists iff a relational product for (A, A × A) exists. Obviously a = b = id(A×A)×A when p2 = p2 and q2 = q2 . The following proposition can be readily seen by a simple computation: Proposition 3. Let α, β, γ : X A be relations. Then the following identities hold: (a) {(α β) γ}a = α (β γ), (b) {α (β γ)}b = (α β) γ.
Let (p, q) and (p0 , q0 ) be pairs of relational products for (A, B) and (X, Y ), respectively. For each pair of relations ξ : A X and η : B Y we define a relation ξ × η : A × B X × Y by ξ × η = pξ qη = pξp0 qηq0 . Ao
p
ξ
Xo
A×B
q
ξ×η
p0
X ×Y
q0
/B /Y
η
If f : A → X and g : B → Y are total functions, then f × g is a unique function such that (f × g)p0 = pf and (f × g)q0 = qg. Remark that an equality
92
Y. Kawahara
(f × g)(f × g ) = f f × gg always holds for all total functions f, f , g and g . Schmidt (Cf. [10,6]) initially suggested that the so-called sharpness property (ξ × η)(ξ × η ) = ξξ × ηη does not hold for relations ξ, ξ , η and η in general. In the rest of the section we review a simple sharpness property [5] needed in the later disscusion. Theorem 1 (Sharpness). (a) If two conditions (1) α β γ δ ρ τ and (2) αα β β hold, then αγ βδ (αρ βτ )(ργ τ δ ). (b) If p : T → A and q : T → B is a pair of total functions with p q = ∇AB , and if there exists a relation η : T X such that η η idX and η p = ∇XA , then αγ βδ = (αp βq )(pγ qδ ).
Proof. (a) It is direct from the following computation. { αγ βδ α(γ α βδ ) = α{γ (γ δ α β)δ } {
α{γ (ρ τ α β)δ } {
α{γ (ρ α βτ )τ δ } {
α(ρ α βτ ){(ρ τ β α)γ τ δ } {
(αρ αα βτ )(ργ τ δ ) {
(αρ βτ )(ργ τ δ ). {
Dedekind Formula } DF } (1) } DF } DF } Sub-distributive } (2) }
(b) An inclusion (αp βq )(pγ qδ ) αγ βδ is trivial from the univalency p p idT and q q idT . To see the converse inclusion, we set ρ = pp , τ = q, α ˆ = αp η and γˆ = γp . We verify that ρ, τ , α ˆ , β, γˆ and δ satisfy two conditions (1) and (2) of (a): (1) α ˆ β γˆ δ ∇T B p∇AB = pp q = ρ τ by the totality of p and p q = ∇AB . (2) α ˆα ˆ β η ηβ β by the univalency η η idY . Therefore we have αγ βδ (α βδ γ)γ βδ { DF }
(α η p)γ βδ { ∇XA = η p }
(αp η )pγ βδ { DF } =α ˆ γˆ βδ
(ˆ αρ βτ )(ρˆ γ τ δ ) {(a)}
(αp βq )(pγ qδ ). { p p idT }
Groups in Allegories
93
Corollary 1. Let (p, q) and (p0 , q0 ) be pairs of projections for (A, B) and (Y, Z), respectively. (a) If X or Y is identical with one of three objects A, B and I, then αγ βδ = (α β)(γ δ) . (b) If X or Y × Z is identical with one of three objects A, B and I, then αξ βη = (α β)(ξ × η).
Proof. It is a direct corollary of Theorem 1(b).
3
Binary Operations
In this section we will study fundamental properties of binary operations in allegories. Throughout of the rest of the paper we assume that G is an object in an allegory A and A has at least two pairs (p, q) and (p1 , q1 ) of projections for (G, G) and (G × G, G), respectively. Under this condition a pair (p2 , q2 ) of projections for (G, G × G) exists, as stated in the section 2. Also we assume that all relations are in A unless othewise stated. Definition 3. A binary operation on G is a relation µ : G × G G.
Let µ : G × G G be a binary operation. Then we define α β = (α β)µ for a pair of relations α, β : X G. Remark that p q = µ. Lemma 1. Assume µ : G × G G is a binary operation on G. Let α, α , β, β : X G, γ, δ : Y G and ξ : Z X be relations in A.
94
Y. Kawahara
Then the following hold: (a) (b) (c) (d) (e) (f)
If α α and β β , then α β α β , ξ(α β) ξα ξβ, ξ(α ∇XY δ) = ξα ∇ZY δ, ξ(∇XY γ β) = ∇ZY γ ξβ, If ξ : Z X is a function, then ξ(α β) = ξα ξβ, (α α ) β (α β) (α β) and α (β β ) (α β) (α β ),
Proof. It is just a corollary of Proposition 2.
The next proposition is a simple result from the sharpness Corollary 1, but it will play an important rˆ ole in the proof of Theorems 5 and 6. Proposition 4. Assume that a binary operation µ : G × G G is total. If X = G or X = I, then αγ βδ (α β)(γ δ) holds. Proof. It is immediate from Corollary 1 (sharpness): { Corollary 1(a) } αγ βδ = (α β)(γ δ)
(α β)µµ (γ δ) { idG×G µµ (µ : total) } = (α β)(γ δ) . Definition 4. A (relational) semigroup G = (G, µ) in A is a pair of an object G in A and a binary operation µ : G × G G, satisfying the associative law (µ×idG )µ = a(idG ×µ)µ, where a : (G×G)×G → G×(G×G) is the associative function. The associative law is an indispensable property on binary operations to simplify iterations of operations. There are often difficulties when one manipulates the associative law in terms of morphisms. So it is convenient to use the traditional form of the associative law (x · y) · z = x · (y · z). The following theorem guarantees that the associative law in terms of morphisms and the traditional form of the associative law using global elements are equivalent. Theorem 2. A binary operation µ : G × G G satisfies the associative law (µ × idG )µ = a(idG × µ)µ if and only if (α β) γ = α (β γ) for all relations α, β, γ : X G.
Groups in Allegories
95
Proof. (⇒) (α β) γ = {(α β)µ γ}µ = {(α β) γ}(µ × idG )µ { = {(α β) γ}a(idG × µ)µ { = {α (β γ)}(idG × µ)µ { = {α (β γ)µ}µ { = α (β γ)
Corollary 1(b) (sharpness) } associative law } Proposition 3(a) } Corollary 1(b) (sharpness) }
(⇐) (µ × idG )µ = (p1 µ q1 )µ { = p1 (p q) q1 { = (p1 p p1 q) q1 { = p1 p (p1 q q1 ) { = ap2 (aq2 p aq2 q) { { = a(p2 q2 µ)µ = a(idG × µ)µ {
µ × idG = p1 µ q1 } µ=p q } Lemma 1(e) } (associative law) } ap2 = p1 p, aq2 p = p1 q, aq2 q = q1 } Lemma 1(e), µ = p q } idG × µ = p2 q2 µ }
Remark. It is trivial that a binary operation µ : G × G G satisfies the commutative law tµ = µ (where t : G × G → G × G is the twisting function defined by t = q p) if and only if α β = β α for all relations α, β : X G.
4
Unitary and Inverse Operations
A total function x : I → X is called an I-point of an object X. A relation ρ : I X is nonempty if there is some I-point x : I → X such that x ρ. In this section we assume G = (G, µ) is a semigroup in A. Proposition 5. Let ε, ε : I G be relations. Then (a) (b) (c) (d)
idG ∇GI ε = idG iff α ∇XI ε = α for all relations α : X G. ∇GI ε idG = idG iff ∇XI ε α = α for all relations α : X G, If idG ∇GI ε = idG and ∇GI ε idG = idG , then ε = ε . If idG ∇GI ε = idG or ∇GI ε idG = idG , then µ is total.
Proof. (a) (⇒) α = α idG = α(idG ∇GI ε) { idG ∇GI ε = idG } = α idG ∇XI ε { Lemma 1(c) } = α ∇XI ε
96
Y. Kawahara
(⇐) Set α = idG . Then idG ∇GI ε = idG simply follows from α = α ∇XI ε. (b) It is similar to (a). (c) ε = ε ∇II ε { (a) : idG ∇GI ε = idG } = ε ε { ∇II = idI } = ∇II ε ε { idI = ∇II } =ε { (b) : ∇GI ε idG = idG } (d) Assume idG ∇GI ε = idG . Then ∇GG = ∇GG (idG ∇GI ε)µ ∇GG×G µ and so ∇GG×G µ = ∇GG , which is equivalent to idG µ µ. A relation ε : I G is called a unitary operation for µ if it satisfies two conditions idG ∇GI ε = idG and ∇GI ε idG = idG . As in the ordinary group theory a unitary operation for a binary operation is unique by Proposition 5(c). Corollary 2. Let ε : I G a unitary operation for µ, and let α, β, γ : X G be relations. If α β = β γ = ∇XI ε, then α = γ. Proof.
α = α ∇XI ε { = α (β γ) { = (α β) γ { = ∇XI ε γ { = γ. {
Proposition 5(a) : idG = idG ∇GI ε } β γ = ∇XI ε } (associative) } α β = ∇XI ε } Proposition 5(b) : idG = ∇GI ε idG }
The following states that a nonempty unitary operation for a total binary operation is an I-point. Theorem 3. Let ε : I G be a unitary operation for µ. If µ is total and ε is nonempty, then ε is an I-point. Proof. As ε is nonempty there is an I-point e : I → G such that e ε. Then idG ∇GI e = (idG ∇GI e)µ is total (since µ is total by the assumtion) and idG ∇GI e idG ∇GI ε = idG . Hence idG ∇GI e = idG (remarking that idG ∇GI e is total and idG is a function), and so ε = e by Proposition 5(c). Remark. Every total relation ε : I G is nonempty under the relational axiom of choice: For all relations α : A B there exists a function f : A → B such that f α and f ∇BA = α∇BA . The notion of inverse operations of course depends on unitary operations. Here we only mention a few general properties on inverse-like relations. Corollary 3. Let ε : I G be a unitary operation for µ. If two relations ι, ι : G G satisfy idG ι = ι idG = ∇GI ε, then ι = ι . Proof. It directly follows from Corollary 2.
Recall a fundamental excercise in the ordinary group theory: If ∀x ∈ G : xe = x and ∀x ∈ G∃y ∈ G : xy = e for some element e ∈ G,
Groups in Allegories
97
then ∀x ∈ G : ex = x and ∀x ∈ G∃y ∈ G : yx = e. (Answer. Assume xy = e and yz = e. Then yx = (yx)e = (yx)(yz) = (y(xy))z = (ye)z = yz = e and ex = (xy)x = x(yx) = xe = x.) The above fact reflects the following theorem on our framework: Theorem 4. Let ε : I G and ι : G G be relations satisfying idG ∇GI ε = idG and idG ι = ∇GI ε. Then the following four conditions are equivalent: (a) (b) (c) (d)
ι is a total function, ι ι2 = ∇GI ε, ∇GI ε ι2 = idG and ∇GI ε idG = idG , ι2 = idG .
Proof. (a)⇒(b) ι ι2 = ι(idG ι) { Lemma 1(e) ι : function } { idG ι = ∇GI ε } = ι∇GI ε = ∇GI ε. { ι : total } (b)⇒(c) (i) ∇GI ε ι2 = (idG ι) ι2 { = idG (ι ι2 ) { = idG ∇GI ε { = idG , {
idG ι = ∇GI ε } (associative) } (b) } idG ∇GI ε = idG }
ι idG = ι (∇GI ε ι2 ) { = (ι ∇GI ε) ι2 { = ι ι2 { = ∇GI ε, {
(i) } (associative) } idG ∇GI ε = idG } (b }
(ii)
(iii) ∇GI ε idG = (idG ι) idG { = idG (ι idG ) { { = idG ∇GI ε = idG . {
idG ι = ∇GI ε } (associative) } (ii) } idG ∇GI ε = idG }
(c)⇒(d) ι2 = ι2 (∇GI ε idG ) { (c) } = ∇GI ε ι2 { Lemma 1(d) } = idG . { (c) } (d)⇒(a) It is trivial by Proposition 1.
Remark. Supposed the relational axiom of choice. If idG ∇GI ε = idG and idG ι = ∇GI ε, and if µ : G × G G and ι : G G are total and ε : I G is a function, then ι is a (unique) total function. (By the relational axiom of choice there exists a total function i : G → G such that i ι. Then idG i
idG ι = ∇GI ε and so idG i = ∇GI ε, since idG i is total and ∇GI ε is a function. Hence, by the virtue of the last Theorem 4, i is a right and left inverse, and finally we have ι = i by Corollary 3.)
98
5
Y. Kawahara
Groups
In this section we define the notion of subgroups and residual relations induced by subgroups, and show a fundamental fact that the residual relations are equivalence relations in allegories. However we limit all the operations of group structures to be total functions for the sake of simplicity. Definition 5. A group G = (G, m, e, i) in A is a quartet of an object G and three total functions m : G × G → G, e : I → G and i : G → G satisfying the following conditions: (Associative Law) (m × idG )m = a(idG × m)m, (Right Unitary) (idG ∇GI e)m = idG , (Right Inverse) (idG i)m = ∇GI e. The following proposition is trivial from the arguments in the previous sections: Proposition 6. Let G = (G, m, e, i) be a group in A and α, β, γ : X G relations. Then the following identities hold: (a) (b) (c) (d) (e) (f)
(α β) γ = α (β γ) (associative), α ∇XI e = ∇XI e α = α (unitary), idG i = i idG = ∇GI e, (inverse), i2 = idG , ei = e and e e = e, mi = qi pi.
Proof. The statement (a) has already been seen in Theorem 2. The statements (b), (c) and (d) follow from Proposition 5 and Theorem 4. (e) An identity e e = e is trivial by (b) and it is easily seen that ei = ei e = e(i idG ) = e∇GI e = e. (f) It follows from Corollary 2, since we have mi m = (qi pi) m = ∇G×GI e as follows: mi m = m(i idG ) { Lemma 1(e) } = m∇GI e { i idG = ∇GI } = ∇G×GI e { m : total } and
m (qi pi) = (p q) (qi pi) { = {p (q qi)} pi { = {p q(idG i)} pi { = (p q∇GI e) pi = (p ∇G×GI e) pi { = p pi { = p(idG i) = p∇GI e = ∇G×GI e. {
m=p q } (associative) } Lemma 1(e) } q : total } Lemma 1(e) } p : total }
Note that qi pi = t(i × i)m, where t = q p : G × G → G × G is the twisting function.
Groups in Allegories
99
In what follows we assume G = (G, m, e, i) is a group in A. In relational calculus there are a few different ways how to specify subobjects. For example, Schmidt and Str¨ ohlein [11] made use of “vectors”, to represent subobjects in relation algebras. We are going to use relations from a unit I into some object G; these do in fact satisfy the vector equation ∇II ρ = ρ. Definition 6. A nonempty relation ρ : I G is a subgroup of G if ρi ρ and ρ ρ ρ. Note that every subgroup ρ : I G contains the unitary operation e : I → G, that is, e ρ. Assume x ρ for some I-point x : I → G. Then e = x∇GI e = x(i idG ) = xi x ρi ρ ρ ρ ρ. Thus a relation ρ : I G is a subgroup of G iff e ρ, ρi = ρ and ρ ρ = ρ. The unitary operation e and the universal relation ∇IG are trivial subgroups of G. Proposition 7. If ρ, ρ : I G are two subgroups of G, then so is ρ ρ .
Proof. It is trivial.
Now let us go back to an exercise (that is, the second question (b) in the introduction) in classical group theory: Let S be a subgroup of a group G. Prove that the right residual relation x ≡ y on G, defined by x−1 y ∈ S, is an equivalence relation. (Answer. The reflexive law: x−1 x = e ∈ S, the symmetric law: If x−1 y ∈ S, then y −1 x = (x−1 y)−1 ∈ S, and the transitive law: If x−1 y ∈ S and y −1 z ∈ S, then x−1 z = (x−1 y)(y −1 z) ∈ S.) The following theorem is a generalization of this fundamental fact. Theorem 5. If ρ : I G is a subgroup, then θ = idG ∇GI ρ is an equivalence relation on G. Proof. Reflexivity idG θ is direct from idG = idG ∇GI e idG ∇GI ρ = θ. Next we can show transitivity θθ θ by the following computation: θθ = θ(idG ∇GI ρ) = θ ∇GI ρ { = (idG ∇GI ρ) ∇GI ρ = idG (∇GI ρ ∇GI ρ) { { = idG ∇GI (ρ ρ)
idG ∇GI ρ { = θ.
Lemma 1(c) } (associative) } ∇GI : total function, Lemma 1(e) } ρ ρ ρ }
100
Y. Kawahara
Finally we will see symmetry θ θ. The proof is somewhat complicated as follows: θ = (p ∇GI ρq )m = {p ∇GI ρ(q ρ ∇IG×G )}m { = {p ∇GI ρ(q iρ ∇IG×G )}m {
{p ∇GI ρ(idG i)(q ∇G×GI ρ) }m { = {p ∇GI ρ∇GI e(q ∇G×GI ρ) }m { = {p ∇GI e(q ∇G×GI ρ) }m {
(idG ∇GI e){p (q ∇G×GI ρ)} m { = {(p q) ∇G×GI ρ} m { = (m ∇G×GI ρ) m { = {m(idG ∇GI ρ)} m { = θ m m
θ . {
ρq = ρ(q ρ ∇IG×G ) : DF } ρi = ρ, i = i } Proposition 4 } idG i = ∇GI e } ρ∇GI = idI } Proposition 4 } idG ∇GI e = idG } p q =m } Lemma 1(c) } m : function }
6
Normal Subgroups
In the final section we define normal subgroups of groups in allegories and show the main result that normal subgroups of a group form a modular lattice. In classical group theory a subgroup of a group is normal iff the right and the left residual relations coincide. We adopt this property to define normal subgroups in our framework: Definition 7. A subgroup ρ : I G of G is called normal if idG ∇GI ρ = ∇GI ρ idG . It is trivial that a subgroup ρ : I G is normal iff α ∇XI ρ = ∇XI ρ α for all relations α : X G. The following is also an analogy from the classical case: Proposition 8. If ρ, σ : I G are normal subgroups of G, then so is ρ σ. Proof. It is easy to see that ρ σ is a subgroup of G. We only show the normality: idG ∇GI (ρ σ) = idG (∇GI ρ ∇GI σ) { = (idG ∇GI ρ) ∇GI σ = (∇GI ρ idG ) ∇GI σ { = ∇GI ρ (idG ∇GI σ) = ∇GI ρ (∇GI σ idG ) { = (∇GI ρ ∇GI σ) idG = ∇GI (ρ σ) idG . {
Lemma 1(e) } ρ : normal } σ : normal } Lemma 1(e) }
Let ρ, σ : I G be normal subgroups of G. Then it is obvious that the normal subgroup ρ σ is the supremum (or, join) of ρ and σ with respect to the relational inclusion of normal subgroups.
Groups in Allegories
101
Definition 8. Let m : X × X → X be a binary operation on X. A relation θ : X X is a congruence with respect to m if it is an equivalence relation such that (θ × θ)m mθ. Proposition 9. If ρ : I G is a normal subgroup of G, then θ = idG ∇GI ρ is a congruence with respect to m. Proof. (θ × θ)m = (pθ qθ)m = pθ qθ = p(idG ∇GI ρ) q(idG ∇GI ρ) = (p ∇G×GI ρ) (q ∇G×GI ρ) { = p (∇G×GI ρ q) ∇G×GI ρ { = p (q ∇G×GI ρ) ∇G×GI ρ { = (p q) (∇G×GI ρ ∇G×GI ρ) { = m ∇G×GI (ρ ρ) { = m ∇G×GI ρ { { = m(idG ∇GI ρ) = mθ
Lemma 1(e) } (associative) } ρ : normal } (associative) } m = p q, Lemma 1(e) } ρ : subgroup } Lemma 1(c) }
In the ordinary group theory it is well-known that the set of all normal subgroups of a group forms a modular lattice. In other words, for three normal subgroups S, T and U of a group G the following modular law holds: S ⊆ U =⇒ ST ∩ U ⊆ S(T ∩ U ). The proof of the fact is fundamental: Assume S ⊆ U and x ∈ ST ∩ U . Then x ∈ U and x ∈ ST , so x = st for some s ∈ S and t ∈ T . Hence t = s−1 x ∈ U holds since x ∈ U and s ∈ S ⊆ U and U is a subgroup. That is, x = st with s ∈ S and t ∈ T ∩ U . Hence x ∈ S(T ∩ U ). To see the above fact in allegories we need the next lemma: Lemma 2. If τ : I G is a subgroup of G, then τ p τ m τ q . Proof. First recall that τ = τ i. Applying Proposition 4
we have
τ p τ m = τ i p τ m { τ = τ i }
(τ τ )(pi m) { Proposition 4 }
τ (pi m) . { τ τ τ }
102
Y. Kawahara
On the other hand we can see the identity pi m = q from pi m = pi (p q) = (pi p) q = p(i idG ) q = p∇GI e q = ∇G×GI e q = q.
{ { { { { {
m=p q } (associative) } Lemma 1(e) } i idG = ∇GI e } p : total } Proposition 6(b) }
Finally we prove a main result that the set of all normal subgroups in allegories forms a modular lattice, as in the classical case: Theorem 6. Let ρ, σ, τ : I G be normal subgroups of a group G. If ρ τ , then (ρ σ) τ ρ (σ τ ). Proof. It follows from the simple computation: (ρ σ) τ = (ρp σq )m τ
(ρp σq τ m )m { DF } = {ρp σq (τ p τ m )}m { ρ τ }
(ρp σq τ q )m { Lemma 2 : τ p τ m τ q } = {ρp (σ τ )q }m = ρ (σ τ ).
7
Conclusion
Group theory is an important field of mathematics with applications in many other science. In this paper the author has tried to demonstrate some fundamental exercises of group theory entirely at the point-free relation-algebraic level. The paper identifies allegories with a unit and relational products as the appropriate relational setting for this aim and presents purely calculational proofs for group theoretic theorems that cannot be easily be shown in the ordinary category-theoretical approach to group theory. Of course, since allegories are sufficiently abstract idea for the relational methods, all the results in the paper can be applicable to theory of fuzzy or L-fuzzy groups, for example. It is interesting to contrast these proofs with graphical approach to relational reasoning due to Curtis and Lowe [2] and Dougherty and Gutierrez [3]. We just remark that by using the graphical rewriting laws 1, 3, 5, 6, 7 and 9 in [2] one can easily prove that αγ βδ ρ τ implies αγ βδ (αρ βτ )(ργ τ δ ). But, as stated in the section 2, Maddux [10] gave a counterexample for the sharpness problem in finitely generated relation algebras. Also the law 2 and 10 seems to show the uniformity condition ∇XY ∇Y Z = ∇XZ for all objects
Groups in Allegories
103
X, Y, Z. These examples indicate that the graphical calculus is based on theory of (homogeneous) binary relations on sets, and the laws of graphical calculus form a stronger setting than those of relation algebras, allegories and Dedekind categories in some sense. Finally the author would like to thank the anonymous referee for his valuable comments and suggestions to the paper.
References 1. C. Brink, W. Kahl and G. Schmidt (eds.), Relational methods in computer science. Advances in Computing Science, (Springer, Wien, New York, 1997). 2. S. Curtis and G. Lowe, A graphical calculus, Lecture Notes in Computer Science 947 (1995), 214-231. 3. D. Dougherty and C. Guti´errez, Normal forms and reduction for theories of binary relations, Lecture Notes in Computer Science, 1833(2000), 95–109. 4. P. Freyd and A. Scedrov, Categories, allegories (North-Holland, Amsterdam, 1990). 5. J. Desharnais, Monomorphic characterization of n-ary direct products, Information Sciences 119(1999), 275–288. 6. W. Kahl and G. Schmidt, Exploring (finite) relation algebras using tools written in Haskell, Technical Report 2000-02, Fakult¨ at f¨ ur Informatik, Universit¨ at der Bundeswehr M¨ unchen, October 2000. 7. Y. Kawahara, Relational set theory, Lecture Notes in Computer Science, 953(1995), 44–58. 8. Y. Kawahara, Lattices in Dedekind categories, In: Orlowska, E. and Szalas, A. (Eds), Relational Methods for Computer Science Applications, Physica-Verlag, 2001, 247–260. 9. S. Mac Lane, Categories for the working mathematician, (Springer-Verlag, 1972). 10. R. Maddux, On the derivation of identities invloving projection functions, Logic Colloquium ’92, ed. Csirmaz, Gabbay, de Rijke, Center for the Study of Language and Information Publications, Stanford, 1995, 145–163. 11. G. Schmidt and T. Str¨ ohlein, Relations and graphs – Discrete Mathematics for Computer Scientists – (Springer-Verlag, 1993). 12. A. Rosenfeld, Fuzzy groups, J. Math. Anal. Appl. 35(1971), 512–517.
Distributed Conceptual Structures Robert E. Kent The Ontology Consortium
[email protected] Abstract. The theory of distributed conceptual structures, as outlined in this paper, is concerned with the distribution and conception of knowledge. It rests upon two related theories, Information Flow and Formal Concept Analysis, which it seeks to unify. Information Flow (IF) [2] is concerned with the distribution of knowledge. The foundations of Information Flow are explicitly based upon the Chu Construction in *autonomous categories [1] and implicitly based upon the mathematics of closed categories [6]. Formal Concept Analysis (FCA) [3] is concerned with the conception and analysis of knowledge. In this paper, we connect these two studies by categorizing the basic theorem of Formal Concept Analysis, thus extending it to the distributed realm of Information Flow. The main result is the representation of the basic theorem as a categorical equivalence at three different levels of functional and relational constructs. This representation accomplishes a rapprochement between Information Flow and Formal Concept Analysis.
1
Introduction
Figure 1 illustrates distributed conceptual knowledge as a two-dimensional structure. The first dimension is along the distribution/conception distinction. Information Flow exists on the distributional side, whereas Formal Concept Analysis extends this toward the conceptual direction. The second dimension is along a functional/relational distinction. All of the development of Information Flow has taken place on the functional level, but some of Formal Concept Analysis 6relational '$ distributional
-
conceptual
&% functional
?
Fig. 1. Two Dimensions of Distributed Conceptual Structures H. de Swart (Ed.): RelMiCS 2001, LNCS 2561, pp. 104–123, 2002. c Springer-Verlag Berlin Heidelberg 2002
Distributed Conceptual Structures
105
extends this into the relational direction, which still might be considered terra incognita. To a large extent the foundation of distributed conceptual structures is based upon binary relations (or matrices) and centered upon the axiom of adjointness between composition and residuation. This composition/residuation adjointness axiom is similar to the axiom of adjointness between conjunction and implication. Since composition and residuation are binary, the axiom has two statements: (1) Left composition is (left) adjoint to left residuation: r◦s ⊆ t iff s ⊆ r\t, for any compatible binary relations r, s and t. (2) Right composition is (left) adjoint to right residuation: r◦s ⊆ t iff r ⊆ t/s, for any compatible binary relations r, s and t. Some derived properties are that residuation preserves composition: (r1 ◦r2 ) \ t = r2 \(r1 \t) and t/(s1 ◦s2 ) = (t/s2 )/s1 and that residuation preserves identity: IdA \t = t and t/IdB = t. The involutions of transpose and negation are of secondary importance. The axiom for transpose states that transpose dualizes residuation: (r\t)∝ = t∝ /r∝ and (t/s)∝ = s∝ \t∝ . There are two important associative laws – one unconstrained the other constrained. There is an unconstrained associative law: (r\t)/s = r\(t/s), for all t ⊆ A×B, r ⊆ A×C and s ⊆ D×B. There is also an associative law constrained by closure: if t is an endorelation and r and s are closed with respect to t, r = t/(r\t) and s = (t/s)\t, then (t/s)\r = s/(r\t), for all t ⊆ A×A, r ⊆ A×B and s ⊆ C×A. Functions have a special behavior with respect to derivation. If function f and relation r are composable, then f ∝ \r = f ◦r. If relation s and the opposite of function g are composable, then s/g = s◦g ∝ . The paper consists of five sections: Introduction, Basic Notions, Architecture, Limit/Colimit Constructions, and Summary and Future Work. The section on Basic Notions is a review of some of the basic ideas of Information Flow and Formal Concept Analysis (classifications, concept lattices, and functional infomorphisms), and an introduction of some new ideas (relational infomorphisms, bonds, and bonding pairs). The section on Architecture, the central section of the paper, is principally concerned with the three categorical equivalences between the distributional pole and the conceptual pole that represent the fundamental theorem of Formal Concept Analysis. The section on Limit/Colimit Constructions, the closest to applications, gives an enhanced fibrational description of the architecture, thereby situating various basic constructions. The final section gives a summary and points out future plans to apply the logic of distributed conceptual structures to actual distributed representational frameworks on the Internet.
2
Basic Notions
There are six basic notions in distributed conceptual structures: classifications, concept lattices, functional infomorphisms, relational infomorphisms, bonds, and bonding pairs. The first two are object-like structures, with classifications being the central object-like notion of Information Flow, and concept lattices being the central object-like notion of Formal Concept Analysis. The last four are
106
R.E. Kent
morphism-like, with functional infomorphisms being the central morphism notion of Information Flow, relational infomorphisms being newly defined in this paper, bonds being used as an analytic tool in Formal Concept Analysis, and bonding pairs being equivalent (in a categorical sense) to complete homomorphisms, the central morphism notion in FCA. 2.1
Objects
Classifications. According to the theory of Information Flow [2], information presupposes a system of classification. Classifications have been important in library science for the last 2,000 years. Major classification systems in library science include the Dewey Decimal System (DDS) and the Library of Congress (LC). However, the library science classification system most in accord with the philosophy and techniques of IF is the Colon classification system invented by the library scientist Ranganathan. A domain-neutral notion of classification is given by the following abstract mathematical definition. A classification A = inst(A), typ(A), |=A consists of (1) a set, inst(A), of things to be classified, called the instances of A, (2) a set, typ(A), of things used to classify the instances, called the types of A, (3) and a binary classification relation, |=A , between inst(A) and typ(A). The notation a |=A α is read “instance a is of type α in inst(A)”. A classification has an alternate expression in either a suitable extension of the theory of closed categories [6] or *-autonomous categories [1]: a classification is a triple A = inst(A), typ(A), |=A , where inst(A) and typ(A) are sets and |=A : typ(A)×inst(A) → 2 is a function, representing a 2-valuated matrix. Classifications are known as formal contexts in Formal Concept Analysis [3]. In FCA, types are called formal attributes and instances are called formal objects. Define the following pair of derivation operators: AA = {α ∈ typ(A) | a |=A α for all a ∈ A} for any instance subset B ⊆ inst(A), and Γ A = {a ∈ inst(A) | a |=A α for all α ∈ Γ } for any type subset Γ ⊆ typ(A). When A : inst(A) → 1 and Γ : typ(A) ← 1 are regarded as relations, derivation is seen to be residuation, AA = A\A and Γ A = A/Γ , where the classification relation is represented as A. An extent of A is a subset of instances of the form Γ A , and an intent of A is a subset of types of the form AA . For any instance a ∈ inst(A), the intent or type set of a is the set typ(a) = aA = {α ∈ typ(A) | a |=A α}. Intent induces a preorder on the instances inst(A) defined by: a ≤A a when A aA ⊇ a . For any type α ∈ typ(A), the extent or instance set of α is the set inst(α) = αA = {a ∈ inst(A) | a |=A α}. Extent induces a preorder on types A typ(A) defined by: α ≤A α when αA ⊆ α . Note that a ∈ αA iff α ∈ aA for any instance a ∈ inst(A) and any type α ∈ typ(A). As befitting such an important and generic notion, classifications abound. Organisms (instances) are classified by scientists into categories (types), such as Plant, Animal, Fungus, Bacterium, Alga, Eukaryote, Prokaryote, etc. Words (instances) are classified in a dictionary by parts of speech (types), such as Noun, Verb, Adjective, Adverb, etc. The following is a motivating example in Barwise and Seligman [2]: Given a first-order language L, the truth classification
Distributed Conceptual Structures
107
of L has L-structures as instances, sentences of L as types, and satisfaction as the classification relation: M |= φ when sentence φ is true in structure M . An ontology forms a classification with either explicit subtyping or subsumption being the classification relation. Any preorder P = P, ≤P is a classification P = P, P, ≤P , where the preorder elements function as both types and instances, and the ordering relation is the classification. Systemic examples of classifications also abound. Given any set A (of instances), the instance powerset classification ℘A = A, ℘A, ∈ associated with A, has elements of A as instances, subsets of A as types, with the membership relation serving as the classification relation. Given any classification A = ∝ inst(A), typ(A), |=A , the dual classification A∝ = typ(A), inst(A), |=A is the involution of A. This is the classification, whose instances are types of A, whose types are instances of A, and whose classification is the transpose of the A classification. The involution operator applies also to morphisms of classifications and limiting constructions on classifications. Concept Lattices. The basic notion of FCA is the notion of a formal concept. A formal concept is a pair a = (A, Γ ) of subsets, A ⊆ inst(A), called the extent of the concept a and denoted extA (a), and Γ ∈ typ(A), called the intent of the concept a and denoted intA (a), that satisfies the closure properties A = Γ A and Γ = AA . There is a naturally defined concept order: a1 ≤ a2 when a1 is more specific than a2 or dually when a2 is more generic than a1 : extA (a1 ) ⊆ extA (a2 ), or equivalently, intA (a1 ) ⊇int A (a2 ). This partial order is part of a complete lattice L(A) = L(A), ≤A , A , A underlying the concept lattice of A, where AA the meet and join are defined by: A (C) = ∩c∈C extA (c), (∪c∈C intA (c)) AA and A (C) = (∪c∈C extA (c)) , ∩c∈C intA (c) for any collection of concepts C ⊆ L(A). Define the instance embedding relation ιA : inst(A) → L(A), as follows: for every instance a ∈ inst(A) and every formal concept a ∈ L(A) the relationship aιA a holds when a ∈ ext(a), a is in the extent of a. This relation is closed on the right with respect to lattice order. Instances are mapped into the lattice by the instance embedding function ιA : inst(A) → L(A), where ιA (a) = aAA , aA for each instance a ∈ inst(A). This function is expressed in terms of the relation as the meet ιA (a) = A aιA . Concepts in ιA [inst(A)] are calledinstance concepts. Any concept a ∈ L(A) can be expressed as the join a = a∈ext(a)ιA (a) of a subset of instance concepts – ιA [inst(A)] is join-dense in L(A). Dually, define the type embedding relation τA : L(A) → typ(A), as follows: for every type α ∈ typ(A) and every formal concept a ∈ L(A) the relationship aτA α holds when α ∈ int(a), α is in the intent of a. This relation is closed on the left respect to lattice order. Types are mapped into the the type embedding func lattice by tion τA : typ(A) → L(A), where τA (α) = αA , αAA for each type α ∈typ(A). This function is expressed in terms of the relation as the join τA (α) = A τA α. Concepts in τA [typ(A)] are called type concepts. Any concept a ∈ L(A) can be expressed as the meet a = α∈int(a) τA (α) of a subset of type concepts – τA [typ(A)] is meet-dense in L(A). Any classification a |=A α can be expressed in
108
R.E. Kent
terms of the instance/type mappings as: ιA (a) ≤A τA (α) – the classification relation decomposes as the relational composition: |=A = ιA ◦ ≤A ◦τA ∝ . The quintuple L(A) = L(A), inst(A), typ(A), ιA , τA , is called the concept lattice of the classification A. More abstractly, a concept lattice L = L, inst(L), typ(L), ιL , τL consists of a complete lattice L, two sets inst(L) and typ(L) called the instance set and type set of L, respectively; along with two functions mapping to the lattice, the instance embedding ιL : inst(L) → L and the type embedding τL : typ(L) → L, such that ιL [inst(L)] is join-dense in L and τL [typ(L)] is meetdense in L. For each classification A the associated quintuple L(A) is a concept lattice. Let A be any classification and let A be any (indexing) set. A collective A-instance indexed by A is any relation a : inst(A) → A, and a collective Atype indexed by A is any relation a : A → typ(A). Let [inst(A), A] denote the preorder of A-indexed A-instances, and let [A, typ(A)] denote the preorder of A-indexed A-types. The equivalence α ≤ a\A iff a◦α ≤ A iff a ≤ A/α states that derivation with respect to A forms an adjoint pair (Galois connection) ( )\A A/( ) : [inst(A), A] [A, typ(A)]op . This observation can be extended to the statement that derivation forms a natural transformation §A : typA ⇒ instA : Relationop → Adjoint between two functors between Relationop the opposite of the category of sets and binary relations and Adjoint the category of preorders and adjoint pairs of monotonic functions. Define the functor typA ∝ to map sets A to the preorder of collective A-types [A, typ(A)] and relations r : A ← B to the adjoint pair r\( ) r◦( ), and define the functor instA to map sets A to the preorder of collective A-instances [inst(A), A] and relations r : A ← B to the adjoint pair ( )◦r ( )/r. For each set X let §A,X denote the derivation adjoint pair ( )\A A/( ). Then §A is a natural transformation with X th component §A,X . An A-indexed collective A-concept is a pair (a, α), where a : inst(A) → A is a collective A-instance, α : A → typ(A) is a collective A-type, which satisfy the closure conditions a = A/α and α = a\A. Let A A denote the collection (lattice) of all A-indexed collective A-concepts. The basic example of a collective A-concept is the pair (ιA , τA ) consisting of the instance relation ιA : inst(A) → L(A) and the type relation τA : L(A) → typ(A), since the closure conditions ιA \A = τA and A/τA = ιA hold. This collective concept is indexed by the concept lattice L(A). Any A-indexed collective A-concept (a, α) induces a unique mediating function f : A → L(A) satisfying the constraints . α = f ◦ τA = f ∝ \τA and a = ιA ◦ f ∝ = ιA /f . The definition f (x) = (ax, xα) is well-defined, since the closure conditions are equivalent to the (pointwise) fact that (ax, xα) ∈ L(A) is a formal concept for each indexing element x ∈ A. Conversely, for any function f : A → L(A), the pair (ιA ◦ f ∝ , f ◦ τA ) is an A-indexed collective A-concept, since the conditions ιA ◦ f ∝ = A/(f ◦ τA ) = (A/τA )/f = ιA /f and f ◦ τA = (ιA ◦ f ∝ )\A = f ∝ \(ιA \A) = f ∝ \τA hold. So, we A have the isomorphism A A ∼ = L(A) , representing the fact that any A-indexed collective A-concept can equivalently be define as a function f : A → L(A).
Distributed Conceptual Structures
109
The left adjoint r\( ) preserves intents: if (b, β) is a B-indexed collective Aconcept, then (A/((b◦r)\A), r\β) is an A-indexed collective A-concept, since r\β = r\(b\A) = (b◦r)\A, an A-indexed A-intent; in particular, if β is a B-indexed collective A-intent, then r\β is an A-indexed collective A-intent. The right adjoint ( )/r preserves extents: if (a, α) is an A-indexed collective A-concept, then (a/r, (A/(r◦α))\A) is a B-indexed collective A-concept, since a/r = (A/α)/r = A/(r◦α), a B-indexed A-extent; in particular, if a is an Aindexed collective A-extent, then a/r is a B-indexed collective A-extent. These two observations are concentrated in an adjoint pair of monotonic functions, φr , ψr : A B A A, where the left adjoint φr : A B → A A is defined by φr ((b, β)) = (A/((b◦r)\A), r\β) and the right adjoint ψr : A B ← A A is defined by ψr ((a, α)) = (a/r, (A/(r◦α))\A). Adjointness follows from the expressions α ≤ r\β iff r◦α ≤ β implies b ≤ a/r and the expressions b ≤ a/r iff b◦r ≤ a implies α ≤ r\β. 2.2
Morphisms
Functional Infomorphisms. Classifications are connected through infomorphisms. Infomorphisms connect classifications and provide a way to move information back and forth between classifications. There exist two generality levels of infomorphism: functional infomorphisms and relational infomorphisms. Functional infomorphisms are the infomorphisms that are used in the literature on Information Flow [2] and *-autonomous categories and the Chu construction [1]. Relational infomorphisms are newly defined in this paper. This subsection gives the definition of functional infomorphism. The next subsection will define the relational version. A (functional ) infomorphism f = inst(f ), typ(f ) : A B from classification A to classification B is a contravariant pair of functions, a function fˆ = typ(f ) : typ(A) → typ(B) in the forward direction between types and a function fˇ = inst(f ) : inst(A) ← inst(B) in the reverse direction between instances, satisfying the following fundamental property fˇ(b) |=A α iff b |=B fˆ(α) for each instance b ∈ inst(B) and each type α ∈ typ(A). In the theory of the Chu construction [1] an infomorphism (a morphism in the category Chu(Set, 2)) is known as a Chu transformation. Systemic examples of functional infomorphisms abound. For any two sets (of instances) A and B, any function f : A ← B and its inverse image function f −1 : ℘A → ℘B form a instance powerset infomorphism ℘f = f, f −1 : ℘A ℘B between the instance powerset classifications. Given any classification A = inst(A), typ(A), |=A , the intent infomorphism ηA : A ℘(inst(A)) from A to the powerset classification of the instance set of A, is the identity function on instances and the intent function ηA (α) = intA (α) on types. Given any two infomorphisms f = inst(f ), typ(f ) : A B and g = inst(g ), typ(g ) : B C , there is a composite infomorphism f ◦ g = inst(g ) · inst(f ), typ(f ) · typ(g ) : A C defined by composing the type and instance functions. Given any classification A = inst(A), typ(A), |=A , the pair of identity functions on types and instances forms an identity infomorphism
110
R.E. Kent
idA : idinst(A) , idtyp(A) : A A (with respect to composition). Given any infomorphism f = inst(f ), typ(f ) : A B, the dual infomorphism is the infomorphism f ∝ : typ(f ), inst(f ) : B ∝ A∝ , whose source classification is the dual of the target of f , whose target classification is the dual of the source of f , whose instance function is the type function of f , and whose type function is the instance function of f . Classifications and functional infomorphisms form a category Classification with involution ( )∝ . For any infomorphism f = inst(f ), typ(f ) : A B, the instance function is a monotonic function between instance preorders inst(f ) : inst(A) ← inst(B), and the type function is a monotonic function between type preorders typ(f ) : typ(A) → typ(B). These facts are represented as two (projection) functors: the functor inst (contravariant) from Classification to Preorderop , the opposite of the category of preorders and monontonic functions, maps each classification to its instance preorder and each infomorphism to its instance monotonic function; and the functor typ from Classification to Preorder maps each classification to its type preorder and each infomorphism to its type monotonic function. An adjoint pair of monotonic functions f = lef t(f ), right(f ) : P Q from preorder P = P, ≤P to preorder Q = Q, ≤Q is a contravariant pair of functions, a monotonic function lef t(f ) : P → Q in the forward direction and a monotonic function right(f ) : P ← Q in the reverse direction, satisfying the following fundamental adjointness property lef t(f )(p) ≤Q q iff p ≤P right(f )(q) for each element p ∈ P and each element q ∈ Q. The fundamental property of functional infomorphisms is clearly related to the notion of adjointness. Preorders and adjoint pairs of monotonic functions form the category Adjoint. Projecting out to the left and right adjoint monotonic functions give rise to two (projection) functors: the functor left from Adjoint to Preorder, and the functor right from Adjoint to Preorderop . Any adjoint pair of monotonic functions is a functional infomorphism between the associated classifications. This fact is expressed as the functor incl from Adjoint to Classificationop that commutes with projection functors: incl ◦ instop = left and incl ◦ typop = right. Relational Infomorphisms. A (relational ) infomorphism r = inst(r ), typ(r ) : A B from classification A to classification B is a pair of binary relations, a type relation rˆ = typ(r ) : typ(A) → typ(B) between types and an instance relation rˇ = inst(r ) : inst(A) → inst(B) between instances, satisfying the following fundamental property: rˇ\A = B/ˆ r. This property is equivalent to either of the properties in Table 1. The common relation rˇ\A = B/ˆ r : inst(B) → typ(A) in the fundamental property is called the bond of r . Given any two relational infomorphisms r = inst(r ), typ(r ) : A B and s = inst(s), typ(s) : B C , there is a composite infomorphism r ◦ s = inst(r )◦inst(s), typ(r )◦typ(s) : A C defined by composing the instance and type relations of the component infomorphisms, and whose fundamental property follows from the preceding composition and associative laws. Given any classification A = inst(A), typ(A), |=A , the pair of identity rela-
Distributed Conceptual Structures
111
Table 1. Relational Infomorphism α∈(ˇ r b)A iff b∈(αˆ r )B r ˇb⊆αA iff αˆ r ⊆bB r ˇB⊆Γ A iff ∀b∈B,α∈Γ rˇb⊆αA iff ∀b∈B,α∈Γ αˆ r ⊆bB iff Γ r ˆ⊆B B r ˇ◦B⊆A/Γ iff B◦Γ ⊆ˇ r \A iff B◦Γ ⊆B/ˆ r iff Γ ◦ˆ r ⊆B\B
typ(A) inst(A)
typ(B)
|=A
inst(B ) r ˇ\A=B/ˆ r
|=B
∀b∈inst(B) and ∀a∈inst(A) ∀B⊆inst(B) and ∀Γ ⊆typ(A) ∀B:inst(B)→1 and ∀Γ :1→typ(A)
typ(A)
r ˆ
- typ(B)
Q A/ˆr 6 k 6 |=A Q r ˇ\A Q - inst(B) inst(A) |=A
r ˇ
tions on instances and types, with the bond being A, forms an identity infomorphism idA = idinst(A) , idtyp(A) : A A (with respect to composition). For any given infomorphism r = inst(r ), typ(r ) : A B the dual infomorphism r ∝ : typ(r ), inst(r ) : B ∝ A∝ is the infomorphism, whose source classification is the dual of the target of r , whose target classification is the dual of the source of r , whose instance relation is the transpose of the type relation of r , and whose type relation is the transpose of the instance relation of r . The fundamental property of relational infomorphisms for composition, identity and involution follow from basic properties of residuation. Classifications and relational infomorphisms form a category Classificationrel with involution ( )∝ . Any functional infomorphism f = fˇ, fˆ = inst(f ), typ(f ) : A B has an • ˇ ˆ associated relational infomorphism fn2rel(f ) = f• , f : A B, whose bond is the relation in the fundamental property [fˇ(b) |=A α iff b |=B fˆ(α)] : inst(B) → typ(A). The definition of the instance and type relations uses the induced orders on instances and types: the type relation fˆ• : typ(A) → typ(B) is defined by fˆ• (α, β) = fˆ(α) ≤B β, and the instance relation fˇ• : inst(A) → inst(B) is defined by fˇ• (a, b) = a ≤A fˇ(b). The operator fn2rel, which maps classifications to themselves, is a functor from Classification to Classificationrel . The theory of classifications with relational infomorphisms can profitably be regarded as a theory of Boolean matrices, with classifications being matrices, infomorphisms being matrix pairs, composition involving matrix multiplication in the two dual senses of relational composition and residuation, and involution using matrix transpose. Bonds. A bond [3] F : A → B between two classifications A and B is a classification F = inst(B), typ(A), |=F , sharing types with A and instances with B, that is compatible with A and B in the sense of closure: type sets {bF | b ∈ inst(B)} are intents of A and instance sets {F α | α ∈ typ(A)} are extents of B. Closure can be expressed categorically as (A/F )\A = F and B/(F \B) = F . The first expression says that (A/F , F ) is an inst(B)-indexed collective A-concept, whereas the second says that (F , F \B) is a typ(A)-indexed
112
R.E. Kent
collective B-concept. A bond is order-closed on left and right: b ≤B b, bF α imply b F α, and bF α, α ≤A α imply bF α . Or, b ≤B b implies b F ⊇ bF , and α ≤A α implies F α ⊆ F α . For any two bonds F : A → B and G : B → C , the composition is the . bond F G = (B/G)\F : A → C defined using left and right residuation. Since both F and G being bonds are closed with respect to B, an equiva. lent expression for the composition is F G = G/(F \B) : A → C . Pointwise, the composition is F G = {(c, α) | F α ⊇ (cG)B }. To check closure, (A/(F G))\A = (A/((B/G)\F ))\A = (B/G)\F , since F being a collective A-intent means that (B/G)\F is also a collective A-intent. With respect to bond composition, the identity bond at any classification A is the classification relation |=A : A → A; its closed relational infomorphism is the instance-type order pair ≤A , ≤A : A → A. With bond composition and bond identities, classifications and bonds form the category Bond. For any relational infomorphism r = ˇ r, rˆ = inst(r ), typ(r ) : A B, the common relation bond bond(r ) = rˇ\A = B/ˆ r in the fundamental property is a bond: for any instance b ∈ inst(B) the bth row of the bond rˇ\A is an intent of A, since b(ˇ r\A) = {α ∈ typ(A) | rˇb ∈ αA } = (ˇ rb)A , and for any th type α ∈ typ(A), the α column of the bond B/ˆ r is an extent of B, since r)B . Any bond F : A → B is the (B/ˆ r)α = {b ∈ inst(B) | αˆ r ∈ bB } = (αˆ bond of some (closed) relational infomorphism r : A B – just make the . . . . definitions: rˇ = A/F and rˆ = F \B, or pointwise rˇb = (bF )A and αˆ r = (F α)B . There is a naturally defined equivalence relation on the collection of relational infomorphisms between any two classifications A and B: two infomorphisms are equivalent when they have the same bond. Since the bond of a composition is the composition of the bonds and the bond of the identity is the identity bond, this defines a bond quotient functor from relational infomorphisms to bonds, that makes the category Bond a quotient category of Classificationrel . The notion of a bond of a relational infomorphism is (ignoring orientation) the same notion of a bond as defined in [3]. The notion of a relational infomorphism, defined for the first time in this paper, supports a categorical rendering of the notion of bond as defined in [3]. Bonding Pairs. A complete (lattice) homomorphism ψ : L → K between complete lattices L and K is a (monotonic) function that preserves both joins and meets. Being meet-preserving, ψ has a left adjoint φ : K → L, and being join-preserving ψ has a right adjoint θ : K → L. So, a complete homomorphism is the middle monotonic function in two adjunctions φ ψ θ. Let Complete Lattice denote the category of complete lattices and complete homomorphisms. The bond equivalent to a complete homomorphism would seem to be given by two bonds F : A → B and G : B → A, where the right adjoint ψF : L(A) → L(B) of the complete adjoint A(F ) = φF , ψF : L(B) L(A) of one bond (say F , without loss of generality) is equal to the left adjoint φG : L(A) → L(B) of the complete adjoint A(G) = φG , ψG : L(A) L(B) of the other bond G with the resultant adjunctions, φF ψF = φG ψG , where the middle adjoint
Distributed Conceptual Structures
113
is the complete homomorphism. This is indeed the case, but the question is what constraint to place on F and G in order for this to hold. The simple answer is to identify the actions of the two monotonic functions ψG and φF , and this is exactly the solution given in [3]. Let (A, Γ ) ∈ L(A) be any formal concept in L(A). The action of the left adjoint φG on this concept is (A, Γ ) → (AGB , AG ), whereas the action of the right adjoint ψF on this concept is (A, Γ ) → (Γ F , Γ FB ). So the appropriate pointwise constraints are: AGB = Γ F and Γ FB = AG , for every concept (A, Γ ) ∈ L(A). We now give these pointwise constraints a categorical rendition. A bonding pair F , G : A B between two classifications A and B is a contravariant pair of bonds, a bond F : A → B in the forward direction and a bond G : B → A in the reverse direction, satisfying the following pairing constraints: (1) F /τA = B/(ιA \G) and (2) ιA \G = (F /τA )\B, which state that (F /τA , ιA \G) is an L(A)-indexed collective B-concept. The definitions of the relations F /τA and ιA \G are given as follows: F /τA = {(b, a) | int(a) ⊆ bF } = {(b, a) | ((bF )A , bF ) ≤B a} and ιA \G = {(a, β) | ext(a) ⊆ Gβ} = {(a, β) | a ≤B (Gβ, (Gβ)A )}. Any concept a = (A, Γ ) ∈ L(A) is mapped by the relations as: (F /τA )((A, Γ )) = {b | Γ ⊆ bF } = Γ F and (ιA \G)((A, Γ )) = {β | A ⊆ Gβ} = AG . Hence, pointwise the constraints are Γ F = B GB and AG = Γ FB . These are the original pointwise constraints of [3] discussed above. The pointwise constraints can be lifted to a collective setting – any bonding pair F , G : A B preserves collective concepts: for any A-indexed collective A-concept (a, α), a\A = α and A/α = a, the conceptual image (F /α, a\G) is an A-indexed collective B-concept, B/(a\G) = F /α and (F /α)\B = a\G. An important special case is the L(A)-indexed collective A-concept (ιA , τA ). To state that the F , G-image (F /τA , ιA \F ) is an L(A)-indexed collective B-concept, is to assert the pairing constraints F /τA = B/(ιA \G) and ιA \G = (F /τB )\B. So, the concise definition in terms of pairing constraints, the original pointwise definition of [3], and the assertion that F , G preserves all collective concepts, are equivalent versions of the notion of a bonding pair. Two other special cases are the collective B-concepts (F , A\G) and (F /A, G) with pairing constraints B/(A\G) = F and (F /A)\B = G, which are F , G-images of the collective A-concepts (|=A , ≤A ) and (≤A , |=A ). Since any collective concept uniquely factors in terms of its mediating function, the image of any collective concept can be computed in two steps: (1) Factor an A-indexed collective A-concept (a, α) in terms of the mediating function f : A → L(A) and the basic collective A-concept (ιA , τA ); and (2) Compose the F , G-image (F /τA , ιB \G) with the mediating function f , resulting in the F , G-image A-indexed collective A-concept (F /α, a\G). Let F , G : A B and M , N : B C be two bonding pairs. Define . the bonding pair composition F , G M , N = F M , N G : A C in terms of bond composition. We can check, either categorically or pointwise, that bonding pair composition is well defined. Let Bonding Pair denote the category, whose objects are classifications and whose morphisms are bonding pairs.
114
3
R.E. Kent
Architecture
The architecture of the distribution/conception distinction is a categorical equivalence at both the functional and the relational poles of the other (scope) distinction. This architecture is in one sense a categorical expression of the basic theorem of Formal Concept Analysis. The central architecture is revealed at the functional level to be the equivalence between Classification and Concept Lattice, at the relational level to be the equivalence between Bond and Complete Adjointop ,and at the complete relational level to be the equivalence between Bonding Pair and Complete Lattice. Not to be forgotten is the fact from Information Flow, that Classification is also equivalent to Regular Theory, the category of regular theories and theory morphisms. 3.1
Functional Equivalence
Information Flow (IF) and Formal Concept Analysis (FCA) are intimately connected. Every classification supports, and is equivalent to, an associated complete lattice called its concept lattice. Every infomorphism defines an adjoint pair of monotonic functions between the concept lattices of its source and target classifications. This section formalizes these observations in a theorem on categorical equivalence. B be any The Concept Lattice Functor. Let f = inst(f ), typ(f ) : A functional infomorphism between two classifications with instance function fˇ = inst(f ) : inst(B) → inst(A) and type function fˆ = typ(f ) : typ(A) → typ(B). How are the two concept lattices L(A) and L(B) related? Since for any concept (A, Γ ) ∈ L(A) the equality fˇ−1 [A] = fˆ[Γ ]B holds between direct and inverse images, the mapping (A, Γ ) → (fˇ−1 [A], (fˇ−1 [A])B ) is a well-defined monotonic function L(fˇ) : L(B) ← L(A) that preserves type concepts in the sense that: τA · L(fˇ) = fˆ · τB . Since it is always true that meet-irreducible concepts are type concepts, if L(B) is type reduced, then L(fˇ) preserves meet-irreducibility – it maps meet-irreducible concepts to meetirreducible concepts. Dually, since for any concept (B, ∆) ∈ L(B) the equality fˇ[B]A = fˆ−1 [∆] holds, the mapping (B, ∆) → (fˆ−1 [∆]A , fˆ−1 [∆]) is a welldefined monotonic function L(fˆ) : L(B) → L(A) that preserves instance concepts in the sense that: ιB · L(fˆ) = fˇ · ιA . Since it is always true that joinirreducible concepts are instance concepts, if L(A) is instance reduced, then L(fˆ) preserves join-irreducibility – it maps join-irreducible concepts to join-irreducible concepts. Moreover, L(fˆ), L(fˇ) : L(B) L(A) is a pair of adjoint monotonic functions between concept lattices, with L(fˆ) left adjoint and L(fˇ) right adjoint. The quadruple L(f ) = L(fˆ), L(fˇ), fˇ, fˆ is called the concept lattice morphism of the infomorphism f = fˇ, fˆ = inst(f ), typ(f ) : A B. More abstractly, a ´ h, ` h, ˇ h ˆ = lef t(h), right(h), inst(h), typ(h) : concept lattice morphism h = h, L K between two concept lattices L and K consists of a pair of ordinary ˇ = inst(h) : inst(L) ← inst(K ) and h ˆ = typ(h) : typ(L) → typ(K ) functions h
Distributed Conceptual Structures
115
between instance sets and type sets, respectively; and an adjoint pair adj(h) = ´ h ` = lef t(h), right(h) : K h, L of monotonic functions, where the right ` = right(h) : L → K is a monotonic function in the forward direction adjoint h ´ = lef t(h) : L ← K is a `=h ˆ · τK , and the left adjoint h that preserves types τL · h ´ = h·ι ˇ L. monotonic function in the reverse direction that preserves instances ιK · h Let Concept Lattice denote the category of concept lattices and concept lattice morphisms. For each infomorphism f = A B the quadruple L(f ) : L(A) L(B) is a concept lattice morphism from L(A), the concept lattice of classification A, to L(B), the concept lattice of classification B. The operator L is a functor called concept lattice from Classification, the category of classifications and functional infomorphisms, to Concept Lattice, the category of concept lattices and concept lattice morphisms. Classification Functor. Any concept lattice L = L, inst(L), typ(L), ιL , τL has an associated classification C(L) = inst(L), typ(L), |=L , which has Linstances as its instance set, L-types as its type set, and the relational composition |=L = ιL ◦ ≤L ◦τL op as its classification relation. Associated with any concept lattice morphism h = lef t(h), right(h), inst(h), typ(h) : L K, from concept lattice L to concept lattice K , is the infomorphism C(h) = inst(h), typ(h) : C(L) C(K ). The fundamental property of infomorphisms is an easy translation of the adjointness condition for adj(h) : K L and the commutativity of the instance/type functions with the adjoint pair of monotonic functions (left/right functions). The operator C is a functor called classification from Concept Lattice, the category of concept lattices and their morphisms, to Classification, the category of classifications and infomorphisms. Equivalence. The functor composition C ◦ L is naturally isomorphic to the identity functor IdConcept Lattice . To see this, let L = L, inst(L), typ(L), ιL , τL be a concept lattice with associated classification C(L) = inst(L), typ(L), |=L . Part of the fundamental theorem of concept lattices [3] asserts the isomorphism L∼ In particular, define the map L(C(L)) → L by (A, Γ ) → = L(C(L)). ι (A) = τ (Γ ) for every formal concept (A, Γ ) in L(C(L)), and the map L L L L L → L(C(L)) by x → ({a ∈ inst(L) | ιL (a) ≤L x}, {α ∈ typ(L) | x ≤L τL (α)}) for every element x ∈ L. These are inverse monotonic functions. Let h = lef t(h), right(h), inst(h), typ(h) : L K be a concept lattice morphism from concept lattice L to concept lattice K , with associated infomorphism C(h) = inst(h), typ(h) : C(L) C(K ). Then, up to isomorphism, L(C(h)) = h. This defines the natural isomorphism: C ◦ L ∼ = IdConcept Lattice . The functor composition L ◦ C is equal to the identity functor IdClassification . To see this, consider whether A = C(L(A)) for any classification A. Obviously, the type and instance sets are the same. What about the classification relations? The classification relation in C(L(A)) is defined in terms of the lattice order and instance/type embeddings by |=L = ιL ◦ ≤L ◦τL op ; which is easily seen to be equal |=L . Hence, A = C(L(A)). What about infomorphisms? The functional infomorphisms f and C(L(f )) are equal.
116
3.2
R.E. Kent
Relational Equivalence
Relational infomorphisms are more general than functional infomorphisms. This increased flexibility and expressiveness must be balanced with a decreased number of properties. However, the property of (categorical) equivalence between the distributional side and the conceptual side still holds. This section formalizes these observations in a theorem on categorical equivalence. The Complete Adjoint Functor. Let F : A → B be any bond and let r = ˇ r, rˆ = inst(r ), typ(r ) : A B be any relational infomorphism having F as its bond. We again ask how the two concept lattices L(A) and L(B) are related, but now in terms of relational infomorphisms. More particularly, how are L(A) and L(B) related to L(F ), the concept lattice of the bond itself? Since for any concept (B, Γ ) ∈ L(F ) the intent Γ = B F = b∈B bF is an intent of L(A), the mapping (B, Γ ) → (Γ A , Γ ) is a well-defined monotonic function ∂0 : L(F ) → L(A) that has a right-adjoint-right-inverse ∂˜0 : L(A) → L(F ) defined as the mapping (A, Γ ) → (Γ F , Γ FF ). Let intF = ∂0 , ∂˜0 : L(F ) L(A) denote this adjoint pair (coreflection). Dually, since for any concept (B, Γ) ∈ L(F ) the extent B = Γ F = α∈Γ F α is an extent of L(B), the mapping (B, Γ ) → (B, B B ) is a well-defined monotonic function ∂1 : L(F ) → L(B) that has a left-adjoint-right-inverse ∂˜1 : L(B) → L(F ) defined as the mapping (B, ∆) → (B FF , B F ). Let extF = ∂˜1 , ∂1 : L(B) L(F ) denote this adjoint pair (reflection). Since for any concept (A, Γ ) ∈ L(A) the equality Γ F = F /Γ = (B/ˆ r)/Γ = B/(Γ ◦ˆ r) = (Γ rˆ)B holds, the mapping (A, Γ ) → (Γ F , Γ FB ) is a well-defined monotonic function L(ˆ r) : L(A) → L(B), which is the composition L(ˆ r) = ∂˜0 · ∂1 . Dually, since for any concept (B, ∆) ∈ L(B) the equality B F = B\F = B\(ˇ r\A) = (ˇ r◦B)\A = (ˇ rB)A holds, the mapping (B, ∆) → (B FA , B F ) is a well-defined monotonic function L(ˇ r) : L(A) ← L(B), which is the composition L(ˇ r) = ∂˜1 · ∂0 . Basic properties of residuation show that this is an adjoint pair of monotonic functions L(ˇ r), L(ˆ r) = intF ◦ extF : L(B) L(A), since B F ⊇ Γ iff . F r), L(ˆ r) B\F ⊇ Γ iff B◦Γ ⊆ F iff B ⊆ F /Γ iff B ⊆ Γ . The pair A(F ) = L(ˇ is called the complete adjoint of the bond F : A → B. More abstractly, a complete adjoint h = lef t(h), right(h) : L K between complete lattices L and K is an adjoint pair of monotonic functions. Let Complete Adjoint denote the category of complete lattices and adjoint pairs of monotonic functions. For each bond F : A → B the pair A(F ) : L(B) L(A) is a complete adjoint between the concept lattices of the target and source classifications B and A. The operator A is a functor called complete adjoint from Bond the category of classifications and bonds to Complete Adjointop the opposite of the category of complete lattices and adjoint pairs. The Bond Functor. Associated with any complete lattice L = L, ≤L , L , L is the classification B(L) = L, L, ≤L , which has L-elements as its instances and types, and the lattice order as its classification relation. Associated with any complete adjoint h = φ, ψ = lef t(h), right(h) : K L is the bond
Distributed Conceptual Structures
117
B(h) : B(L) → B(K ) defined by its adjointness property: yB(h)x iff φ(y) ≤L x iff y ≤K ψ(x) for all elements x ∈ L and y ∈ K. The closure property of bonds is obvious, since yB(h) =↑L φ(y) for all elements y ∈ K and B(h)x =↓K ψ(x) for all elements x ∈ L. The operator B is a functor called bond from Complete Adjointop the category of complete lattices and adjoint pairs to Bond the category of classifications and bonds. Equivalence. The functor composition B ◦ A is naturally isomorphic to the identity functor IdComplete Adjointop . To see this, let L = L, ≤L , L , L be a complete lattice with associated classification B(L) = L, L, ≤L . Part of the fundamental theorem of concept lattices [3] asserts the isomorphism L ∼ = A(B(L)). In particular, formal concepts of A(B(L)) are of the form (↓L x, ↑L x) for elements x ∈ L. So, define the obvious maps A(B(L)) → L by (↓L x, ↑L x) → x and L → A(B(L)) by x → (↓L x, ↑L x), for every element x ∈ L. Let h = φ, ψ = lef t(h), right(h) : K L be a complete adjoint with associated bond B(h) : B(L) → B(K ). Then, the right adjoint of A(B(h)) maps (↓L x, ↑L x) → (↓L ψ(x), ↑L ψ(x)) and the left adjoint of A(B(h)) maps (↓K y, ↑K y) → (↓K φ(y), ↑K φ(y)). So, up to isomorphism, A(B(h)) is the same as h. This defines the natural isomorphism: B ◦ A ∼ = IdComplete Adjointop . The basic theorem of Formal Concept Analysis [3] can be framed in terms of two fundamental bonds (relational infomorphisms) between any classification and its associated concept lattice. For any classification A the instance embedding relation is a bond ιA : L(A) → A from the concept lattice to A itself, and the type embedding relation is a bond τA : A → L(A) in the reverse direction. The pair L(A)/ιA , τA : L(A) A is a relational infomorphism whose bond is the instance embedding relation. The pair ιA , τA \L(A) : A L(A) is a relational infomorphism whose bond is the type embedding relation. The instance and type embedding bonds are inverse to each other: ιA τA = IdL(A) and τA ιA = IdA . The functor composition A ◦ B is naturally isomorphic to the identity functor IdBond . To see this, let A be a classification with associated concept lattice A(A). The comments above demonstrate the isomorphism A ∼ = B(A(A)). Let F : A → B be a bond between classifications A and B with associated complete adjoint A(F ) : A(B) A(A). The bond B(A(F )) : B(A(A)) → B(A(B)) contains a conceptual pair (b, a) of the form b = (B, ∆) and a = (A, Γ ) iff B ◦ Γ ⊆ F , where B ◦ Γ = B×Γ a Cartesian product or rectangle, iff B ⊆ Γ F iff Γ ⊆ B F . So, (ιA F )τB = (F /(ιA \A))τB = (F /τA )τB = (B/τB )\(F /τA ) = ιB \(F /τA ) = B(A(F )), by bond composition and properties of the instance and type relations. Hence, B(A(F ))ιB = ιA F . This proves the required naturality condition. 3.3
Complete Relational Equivalence
In the section on relational equivalence, we have seen how bonds are categorically equivalent to the opposite of complete adjoints, adjoint pairs between complete lattices. Unfortunately, these are not the best morphisms for making structural comparisons between complete lattices. Complete homomorphisms are best for
118
R.E. Kent
this [3]. Since complete homomorphisms are special cases of complete adjoints on the conceptual side, we are interested in what constraints to place on bonds on the distributional side. The Complete Lattice Functor. Let F , G : A B be any bonding pair. Then F : A → B is a bond in the forward direction from classification A to classification B, and G : A → B is a bond in the reverse direction to classification A from classification B. Applying the complete adjoint functor, we get two adjoint pairs in opposite directions: φF , ψF : L(B) L(A) in the forward direction and φG , ψG : L(A) L(B) in the reverse direction. It was shown above that for bonding pairs the meet-preserving monotonic function ψF : L(A) → L(B) is equal to the join-preserving monotonic function φG : L(A) → L(B), giving a complete homomorphism. This function is the unique mediating function for the L(A)-indexed collective B-concept (F /τA , ιA \G), the F , G-image of the L(A)-indexed collective A-concept (ιA , τA ), whose closure expressions define the pairing constraints. The complete lattice functor A2 : Bonding Pair → Complete Lattice is the . operator that maps a classification A to its concept lattice A2 (A) = L(A) regarded as a complete lattice only, and maps a bonding pair F , G : A B to . its complete homomorphism A2 (F , G) = ψF = φG : L(A) → L(B). The Bonding Pair Functor. Let ψ : L → K be a complete homomorphism between complete lattices L and K with associated adjunctions φ ψ θ. Then φ, ψ : K L and ψ, θ : L K are morphisms in Complete Adjoint. Application of the bond functor B produces the bonds F = B(φ, ψ) : B(L) → B(K ) and G = B(ψ, θ) : B(K ) → B(L) between the classifications B(L) = L, L, ≤L and B(K ) = K , K , ≤K . Note that for any complete lattice L the instance and type relations for the classification B(L) are both equal to the order relation ιL = ≤L = τL . Since F /τL = F /≤L = F , B(K )/(ιL \G) = ≤K /(≤L \G) = ≤K /G = F , ιL \G = ≤L \G = G and (F /τL )\B(K ) = (F /≤L )\≤K = F \≤K = G, the pair of bonds (F , G) is a bonding pair (F , G) = (B(φ, ψ), B(ψ, θ)) : B(L) B(K ). Let B2 (ψ) denote this pair. The bonding pair functor B2 : Complete Lattice → Bonding Pair is the operator that maps a complete lattice L to its classification L, L, ≤L and maps a complete homomorphism to its bonding pair as above. Since the bond functor B is functorial, so is B2 . Equivalence. The functor composition A2 ◦ B2 is naturally isomorphic to the identity functor IdBonding Pair . Consider any classification A. The type and instance embedding relations form bonding pairs in two different ways, τA , ιA : A L(A) = A2 ◦B2 (A) and ιA , τA : A2 ◦B2 (A) = L(A) A, and these are inverse to each other: τA , ιA ιA , τA = IdA and ιA , τA τA , ιA = IdL(A) . So that each classification is isomorphic in Bonding Pair to its concept lattice: A ∼ B be a bonding pair. = B2 (L(A)) = A2 ◦B2 (A). Let F , G : A As shown above, the naturality conditions for bonds F and G are expressed as
Distributed Conceptual Structures
119
ιA F τB = A◦B(F ) and ιB G τA = A◦B(G). So ιA , τA F , GτB , ιB = ιA F τB , ιB G τA = A◦B(F ), A◦B(G) = A2 ◦B2 (F , G). The functor composition B2 ◦ A2 is naturally isomorphic to the identity functor IdComplete Lattice . Consider any complete lattice L. As we have seen in studying relational equivalence, any complete lattice L is isomorphic to the complete lattice (B2 ◦A2 )(L) = L(L, L, ≤L ) via the bijection x → (↓L x, ↑L x). Now consider any complete homomorphism ψ : L → K between complete lattices L and K with associated adjunctions φ ψ θ. The bonding pair functor maps this to the bonding pair B2 (ψ) = (B(φ, ψ), B(ψ, θ)), and the complete lattice functor maps this to the complete homomorphism A2 (B2 (ψ)) = ˜ ψ˜ : L(L, L, ≤L ) → L(K , K , ≤K ) with associated adjunctions φ˜ ψ˜ θ, ˜ ˜ where φ(↓K y, ↑K y) = (↓L φ(y), ↑L φ(y)), ψ(↓L x, ↑L x) = (↓K ψ(x), ↑K ψ(x)) ˜ K y, ↑K y) = (↓L θ(y), ↑L θ(y)). Clearly, the naturality condition holds and θ(↓ between ψ and (B2 ◦A2 )(ψ). 3.4
Theorem and Architectural Diagram
Figure 2 contains a commuting diagram of functors, categories and equivalences that represents the architecture of distributed conceptual structures. This is the central contribution of this paper. Many of the details that support this diagram are known from the literature ([2], [3]). This paper seeks to bring these facts together into a coherent view. Figure 2 is two-dimensional, having the same orientation as Fig. 1. The vertical dimension contains the functionalrelational distinction. The morphisms of the equivalent categories Classification and Concept Lattice are function-based, whereas the equivalent categories Bond and Complete Adjointop are relation-based. The horizontal dimension contains the distributional-conceptual distinction. This is represented as the three categorical equivalences expressed in Theorem 1. The equivalences between the categories Classification and Concept Lattice at the functional level and the categories Bond and Complete Adjointop at the relational level are expressions of the basic theorem of Formal Concept Analysis [3]. The equivalence between the categories Classification and Regular Theory is one of the main contributions of the theory of Information Flow [2]. Theorem 1. The distributional/conceptual distinction in the architecture of distributed conceptual structures is represented by three categorical equivalences. – The category of classifications and functional infomorphisms is categorically equivalent to the category of concept lattices and concept lattice morphisms Classification ≡ Concept Lattice via the lattice functor and classification functor, which are generalized inverses: L ◦ C = IdClassification and C ◦ L ∼ = IdConceptLattice . – The category of classifications and bonds is categorically equivalent to the category of complete lattices and complete adjoints Bond ≡ Complete Adjointop via the complete adjoint functor and bond functor, which are generalized inverses: A ◦ B ∼ = IdBond and B ◦ A ∼ = IdComplete Adjointop .
120
R.E. Kent Aop
Bonding Pair ∂0
Bond
∂1
Bondop
≡
∂0
-
A
Complete Lattice
Bop
R @
pairing constraints
-
≡
-
≡
ρ
R @
Complete Adjoint
pullback
Complete Adjointop
B
∂1
R @
λop
Preorder
-
A
≡
Bond bond
6
info
?
6
Classificationrel fn2rel
Complete Adjointop
B
forget
6 -
L
Classification
≡ C
Concept Lattice
Fig. 2. Architectural Diagram of Distributed Conceptual Structures
– The category of classifications and bonding pairs is categorically equivalent to the category of complete lattices and complete homomorphisms Bonding Pair ≡ Complete Lattice via the complete lattice functor and bonding pair functor, which are generalized inverses: A2 ◦ B2 ∼ = IdBonding Pair and B2 ◦ A2 ∼ = IdComplete Lattice . Proof. The bulk of this paper provides the proof for this theorem.
4
Limit/Colimit Constructions
Limit/colimit constructions are very important in applications, often harboring the central semantics. Categorical equivalence is more general and more useful than categorical isomorphism. Fact 1 in the appendix expresses and proves the well-known observation that equivalent categories have the same limit/colimit structures. Since the distributional-conceptual distinction, the horizontal dimension illustrated in Fig. 2 and expressed in Theorem 1, has three categorical equivalences, limit/colimit structures will give added meaning to distributive conceptual structures. In brief comments below, although we describe limit/colimit constructions on one side of the equivalence, we are assured of their preservation when mapped to the other side of the equivalence. A true conception of the limit/colimit architecture requires an understanding of the two-dimensional fibrational nature of distributive conceptual structures. Types and instances form fibered spans. The fibrational instance dimension is defined in terms of the instance forgetful functor inst : Classification → Setop , here simplified by ignoring instance order. For any set (of instances) A, the Ath fiber
Distributed Conceptual Structures
121
category of inst, denoted inst−1 (A), is the subcategory of classifications A with instance set inst(A) = A and infomorphisms f : A B with instance function inst(f ) = IdA . The fibrational type dimension is defined dually in terms of the type forgetful functor typ : Classification → Set. The initial object in the A-th fiber category inst−1 (A) is the empty classfication 0A = A, ∅, ∅, the terminal object in inst−1 (A) is the instance powerset classification ℘A = A, ℘A, ∈ – for any classification A, the extent infomorphism ηA : A ℘(inst(A)) is the unique infomorphism in inst−1 (A) from A to ℘(inst(A)). Given two classifications A0 and A1 in the A-th fiber category inst−1 (A), the apposition A0 | A1 is the coproduct of A0 and A1 in inst−1 (A). Dual comments hold for the type fibrational dimension, where for example subposition is the product construction. The full category Classification contains all limits and colimits. Given two classifications A and B, the sum (semiproduct) A+B is the coproduct of A and B in Classification. The dual notion provides the product. All limits/colimits exist when not only products/coproducts but also quotient constructions exist. Given a classification A, a dual invariant is a pair J = (A, R) consisting of a set A ⊆ inst(A) of instances of A and a binary relation R on types of A satisfying the constraint: if αRβ, then for each a ∈ A, a |=A α if and only if a |=A β. The dual quotient of A by J, written A/J, is the classification with instances A, whose types are the R-equivalence classes of types of A, and whose classification is a |=A/J [α] if and only if a |=A α. Dual quotients include kernel image factorizations, coequalizors and pushouts. Suitable colimit constructions in Classification have been used by the author to define the semantics of ontology sharing [4,5].
5
Summary and Future Work
This paper has had as its goal the formulation of a framework for conceptual knowledge representation. For this it uses the language of category theory in order to represent some of the essence of Information Flow and Formal Concept Analysis, thereby unifying these two studies. This has culminated in the recognition of the two-dimensional nature of distributed conceptual structures (Fig. 1), whose particulars, described as a commuting diagram (Fig. 2), represent the fundamental theorem of Formal Concept Analysis in terms of the three categorical equivalences expressed in Theorem 1. Information Flow has initiated development of the distributed nature of the logic of information, principally represented by the morphisms on the distributional side of the diagram. Formal Concept Analysis has developed the conceptual nature of knowledge, principally represented by the objects on the conceptual side of the diagram. The author is currently engaged in the development of the Information Flow Framework (IFF), whose mission is to further the development of the theory of Information Flow and Formal Concept Analysis and to apply these to distributed logic, ontologies, and knowledge representation. The Information Flow Framework is manifest as the IFF Foundation Ontology1 , which is being formulated in a new version of the Knowledge Interchange Format (KIF). The IFF 1
Located at: URL: http://suo.ieee.org/IFF/
122
R.E. Kent
Foundation Ontology represents metalogic. It provides a principled foundation for the metalevel (structural level) of the Standard Upper Ontology (SUO). The SUO metalevel can be used as a logical framework for manipulating collections of object level ontologies. The IFF Foundation Ontology is rather large, but highly structured. It is partitioned into three metalevels: top, upper and lower. These metalevels correspond to the set-theoretic distinction in foundations between the generic, the large and the small. Each metalevel is partitioned into numerous namespaces and services the level below it. The top metalevel provides an interface between the KIF logical language and the upper metalevel of the IFF Foundation Ontology. The upper metalevel is partitioned into three sub-ontologies: the IFF Upper Core Ontology, the IFF Upper Classification Ontology, and the IFF Category Theory Ontology. The lower metalevel contains, amongst other modules, the IFF Model Theory Ontology, the IFF Algebraic Theory Ontology, and the IFF Ontology Ontology. The lower metalevel represents object-level ontologies by providing terminology and axiomatization for (1) the formal or axiomatic semantics of IFF theories, (2) the interpretative semantics of IFF model-theoretic structures, and (3) the combined semantics of IFF logics. Guided by the Upper Classification Ontology, the lower metalevel also offers a complete framework for a conceptual “lattice of theories” as concentrated in the fundamental truth meta-classification and truth concept meta-lattice2 . The IFF Upper Classification Ontology, which enables metalevel reasoning about truth, formalizes the terminological and axiomatic content of this paper.
References 1. Barr, M.: *-Autonomous categories and linear logic. Mathematical Structures in Computer Science 1 (1991) 159–178. 2. Barwise, J., Seligman, J.: Information Flow: the Logic of Distributed Systems. Cambridge University Press, Cambridge (1997). 3. Ganter, B., Wille, R.: Formal Concept Analysis: Mathematical Foundations. Springer, Heidelberg (1999). 4. Kent, R.: The Information Flow Foundation for Conceptual Knowledge Organization. Beghtol, C., Howarth, C., Williamson, N. (eds.): Dynamism and Stability in Knowledge Organization. Proceedings of the Sixth International ISKO Conference. Advances in Knowledge Organization 7 (2000) 111–117. Ergon Verlag, W¨ urzburg. 5. Kent, R.: The IFF Foundation for Ontological Knowledge Organization. Beghtol, C., Williamson, N. (eds.): Knowledge Organization and Classification in International Information Retrieval. Cataloging and Classification Quarterly. The Haworth Press Inc., Binghamton, New York (2003). 2
The truth classification of a first-order language L is the meta-classification, whose instances are L-structures, whose types are L-sentences, and whose classification relation is satisfaction. The truth concept lattice, the concept lattice of the truth classification, functions as the appropriate “lattice of ontological theories” for the SUO. A formal concept in this lattice has an intent that is a closed theory and an extent that is the collection of all models for that theory. The theory (intent) of the join or supremum of two concepts is the intersection of the theories (conceptual intents), and the theory (intent) of the meet or infimum of two concepts is the theory of the common models.
Distributed Conceptual Structures
123
6. Lawvere, W.: Metric Spaces, Generalized Logic and Closed Categories. Rendiconti del Seminario Matematico e Fisico di Milano 43 (1973) 135–166. 7. Mac Lane, S.: Categories for the Working Mathematician. Springer, New York (1971).
A
Appendix
According to Saunders Mac Lane [7], equivalences between categories are more general, and more useful, than isomorphisms between categories. We emphatically concur and we argue that the main reason for their usefulness is derived from the well-known observation that equivalent categories have the same limit/colimit structures. We make this precise by the following fact. Fact 1 If categories A and B are equivalent, then they have the same limit/colimit structures. More particularly, if B has limits (colimits) for all Cshaped diagrams, then A also has limits (colimits) for all C-shaped diagrams, and vice-versa. In particular, A is complete (co-complete) iff B is complete (cocomplete). Proof. Assume F : A → B and G : B → A are functors that mediate the equivalence through natural isomorphisms η : IdA ⇒ F ◦ G and : G ◦ F ⇒ IdB . This means the following. – η is the unit and is the counit of adjunction F G: ηF • F = IdF and Gη • G = IdG . – The natural transformations η are η −1 are inverse: η • η −1 = IdIdA and η −1 • η = IdF◦G . – The natural transformations are −1 are inverse: • −1 = IdG◦F and −1 • = IdIdB . −1 – is the unit and η −1 is the counit of adjunction G F: −1 G • Gη −1 = IdG and F−1 • η −1 F = IdF . Let D : C → A be any C-shaped diagram in A. Diagram D is mapped by F to D ◦ F : C → B, a C-shaped diagram in B. By assumption, there is a limiting cone λ : B ⇒ D ◦ F for D ◦ F in B with limit object B. We will show that λG • Dη −1 : G(B) ⇒ D is a limiting cone for D in A with limit object G(B). Let γ : A ⇒ D be any cone for D in A. Then γF : F(A) ⇒ D ◦ F is a cone for D ◦ F in B. Since λ is a limiting cone, there is a unique B-arrow g : F(A) → B . with g · λ = γF. Define A-arrow f = ηA · G(g) : A → G(F(A)) → G(B). We will show that f is the unique mediating A-arrow for cone λG • Dη −1 ; that is, f is the unique A-arrow satisfying the constraint f · λG • Dη −1 = γ. [Existence] f · G(λi ) · η −1 Di = ηA · G(g) · G(λi ) · η −1 Di = ηA · G(g · λi ) · η −1 Di = ηA · G(F(γi )) · η −1 Di = ηA · ηA −1 · γi = γi . [Uniqueness] Suppose that f˜ : A → G(B) is any A-arrow satisfying f˜ · (λG • Dη −1 ) = γ. Applying F, get F(f˜) · η −1 B · λi = F(f˜) · F(G(λi )) · η −1 F(Di ) = F(f˜) · F(G(λi )) · F(η −1 Di ) = F(f˜· G(λi ) · η −1 Di ) = F(γi ) = g · λi . By uniqueness F(f˜) · ηB −1 = F(f˜) · η −1 B = g. Applying G, get f˜ = ηA · G(F(f˜)) · G(ηB −1 ) = ηA · G(g).
A Computer Algebra Approach to Relational Systems Using Gr¨ obner Bases L.M. Laita1 , E. Roanes-Lozano2 , L. de Ledesma1 , T. Calvo3 , and L. Goz´alez-Sotos3 1
3
Universidad Polit´ecnica de Madrid, Dept. Artificial Intelligence, Campus de Montegancedo, Boadilla del Monte, 28660-Madrid, Spain 2 Universidad Complutense de Madrid, Dept. Algebra, Edificio “Almudena”, c/ Rector Royo Villanova s/n, 28040-Madrid, Spain Universidad de Alcal´ a, Dept. de Ciencias de la Computaci´ on, Escuela Polit´ecnica, Campus Universitario, 28871-Alcal´ a de Henares, Madrid, Spain Abstract. The aim of this paper is to suggest that Gr¨ obner bases can be applied to the study of reflexive and transitive relations. To illustrate our suggestion, we consider two examples. The first example is a Rule-BasedExpert-System: the reflexive and transitive relation is the implication arrow. The second example deals with railway interlocking systems: the reflexive and transitive relation is the accessibility to sections in a railway network. The study is generic; it can be applied to any reflexive and transitive relation.
1
Introduction
The aim of this paper is to suggest that some Computer Algebra theoretical results and their implementation using Gr¨ obner bases, can be applied to reflexive and transitive relations. To illustrate our suggestion, we consider two examples. The first example is a Rule-Based-Expert-System (denoted RBES) dealing with what in medical circles is known as “appropriateness criteria” for the application of certain techniques in medicine [5]. Considering the RBES as a RAS (Relational Attribute System) [4], the reflexive and transitive relation is the implication arrow. The second example deals with railway interlocking systems. Railway interlocking systems are designed to prevent conflicting actions (related to the position of switches and signals) during railway exploitation. A decision model (independent from the topology of the station) based on the use of polynomial ideals and Gr¨ obner bases is presented [9]. The reflexive and transitive relation is the accessibility to sections in a railway network.
2
The RAS-RBES “Appropriateness Criteria”
The next Computer Algebra theoretical result links “tautological consequence” in logic with an ideal membership problem in algebra using Gr¨ obner bases [3,6,7]. H. de Swart (Ed.): RelMiCS 2001, LNCS 2561, pp. 124–133, 2002. c Springer-Verlag Berlin Heidelberg 2002
A Computer Algebra Approach to Relational Systems Using Gr¨ obner Bases
125
Theorem 1. A formula A0 is a tautological consequence of a set of formulae {A1 , A2 , ..., Am }, in any p-valued logic for p a prime number, iff the polynomial translation of (the negation of ) A0 belongs to the ideal J + I, where J is the ideal generated by the polynomial translations of (the negations of ) A1 , A2 , ..., Am , and I is the ideal generated by the polynomials xp1 −x1 , xp2 −x2 , ..., xpn −xn . Intuitively written: N EG(A0 ) ∈< N EG(A1 ), N EG(A2 ), ..., N EG(Am ), xp1 − x1 , xp2 − x2 , ..., xpn − xn ) >
which can be denoted as N EG(A0 ) ∈ J + I . Particularly interesting is the application of this theorem to a type of RAS whose relations are reflexive and transitive: RBES represented in multi-valued and modal logic. In this case the formulae A1 , A2 , ..., Am , are the production rules and facts (and possibly any other information as integrity constraints). We have been able to apply the theorem to verification and extraction of new knowledge in RBES containing up to 200 propositional variables under bi-valued logic and up to 150 propositional variables under three-valued and modal logic. The process of verification of consistency of a certain ideal takes a few seconds. The same holds for the process of extraction of consequences. Next we give an outline of a particular example of RAS-RBES dealing with what is known in Medicine as “appropriateness criteria”. 2.1
Table Description
A set of medical data -effort test proof, one, two, or three blood vessels diseased, LVEF (Left Ventricle Ejection Fraction) value- were presented to a panel of ten experts on coronary diseases. They were asked about the appropriateness of performing or not the action of revascularization, which could have been done by means of two techniques: PTCA (Percutaneous Transluminal Coronary Angioplasty) and CABG (Coronary Artery Bypass Grafting). The experts’ opinions were set out in a table containing 260 information items. Items 1 to 24 are transcribed afterwards as illustration. (The patients are asymptomatic) 1. Effort test positive. 1.1. Left common trunk diseased 1.1.1. Surgical risk low-moderate % LVEF (F) F > 50 50 ≥ F > 30 30 ≥ F ≥ 20
Revascularization PTCA and CABG 1:1 2 3 4 5 6 7 81 ∗9 +A 2:1 2 3 4 5 6 7 81 ∗9 − + A 5:1 2 3 4 5 6 7 8 ∗10 +A 6:1 2 3 4 5 6 7 8 1 ∗9 − + A 9:1 2 3 4 5 6 7 81 ∗9 +A 10:1 2 3 4 5 6 71 81 ∗8 − + A
126
L.M. Laita et al.
1.1.2. Surgical risk high. % LVEF (F) Revascularization PTCA and CABG F > 50 3:1 2 3 4 5 6 7 84 ∗6 +A 4:1 2 3 4 51 6 71 81 ∗7 − + A 50 ≥ F > 30 7:1 2 3 4 5 6 7 83 ∗7 +A 8:1 2 3 4 5 61 71 8 2 ∗6 − + A 30 ≥ F ≥ 20 11:1 2 3 4 5 6 71 82 ∗7 +A 12:1 2 3 4 5 62 7 82 ∗6 − + A 1.2. Three blood vessels diseased 1.2.1. Surgical risk low-moderate % LVEF (F) Revascularization PTCA F > 50 13:1 2 3 4 51 61 71 82 ∗ 95 + A 14:1 2 3 50 ≥ F > 30 17:1 2 3 4 5 6 71 82 ∗7 +A 18:1 2 3 30 ≥ F ≥ 20 21:1 2 3 41 5 6 7 82 ∗7 +A 22:1 2 3 1.2.2. Surgical risk high. % LVEF (F) F > 50 50 ≥ F > 30 30 ≥ F ≥ 20
and CABG 4 51 61 71 82 ∗ 95 ? + A 4 51 6 72 82 ∗ 95 ? + A 4 51 61 72 ∗2 94 ? + A
Revascularization PTCA and CABG 15:11 2 3 4 51 6 72 ∗2 94 + A 16:1 2 33 42 5 ∗7 81 94 + +I 19:1 2 3 4 5 6 71 83 ∗6 +A 20:1 2 33 42 5 ∗ 7 8 2 93 + +D 23:1 2 3 41 5 6 7 83 ∗6 +A 24:1 22 31 41 51 ∗ 61 71 81 92 + +D
A bold-type digit (1, 2,..., 260) is assigned to each row of digits, symbols +, −, ∗, and letters A, D, I of the table. The rules R1, R2,..., R260 of the RBES that translates the table, are numbered according to the digits mentioned above. The experts were informed, for instance, in the six cases 1, 2, 5, 6, 9, 10 that, for a certain patient, the data were: effort test positive, suffers from left common trunk disease, surgical risk is low/moderate and LVEF is in a given percentage bracket. Once a set of data is given, the panelists are asked about the appropriateness of a therapeutic action: revascularization, which can be performed by means of two different techniques, PTCA and CABG. The panelists’ response is transcribed as a row of digits with or without superscripts, symbols +, −, ?, ∗, and letters A, D, I. The superscripts express the number of experts that have assigned a value from 1 to 9 to the appropriateness of an action. “∗” stands for the median. The symbols + , − (and ? used in other parts of the table) respectively mean that an action is, respectively, appropriate, inappropriate (and of undecided appropriateness). If the panelists refer to revascularization, only one of these symbols is transcribed. PTCA and CABG are referred to simultaneously by a pair of these symbols. The letter A at the end of a row means “agreement” about appropriateness, inappropriateness or undecided appropriateness. D means “disagreement”and I (which appears in other parts of the table) means “undecided agreement”. There is disagreement D if the ratings of three panelists rank between 1 and 3, and the ratings of another three rank between 7 and 9. There is agreement when there are no more than two opinions outside an interval of [1,3], [4,6] or [7,9] that contains the median. In any other case, there is undecided agreement I (i. e. big standard deviation).
A Computer Algebra Approach to Relational Systems Using Gr¨ obner Bases
2.2
127
Translation of the Tables into a Set of Production Rules
The tables can be translated into a set of production rules as follows. • Assign a propositional variable, denoted x[i] (i = 1, ..., 9, 13), y[1], y[2] (possibly preceded by a permissible combination of symbols ¬, 2, ♦), to each symptom and therapeutic action. Let us refer as illustration to a few of them, for instance: - Surgical risk high: ¬x[1] - Effort test positive: x[2] - ... - Effort test not done or not decisive: ♦¬x[2] - LVEF > 50%: x[7] - ... - Left common trunk disease: x[3] - ... - Anterior proximal descendent affected: x[13] - ... - PTCA: y[1] - CABG: y[2] The relation between a given set of symptoms and the respective experts’ response can be reinterpreted as a production rule subject to the following convention. • In production rules based on the three-valued and modal logic here illustrated, the propositional variables x[i] (i = 1, .., 9), x[13], y[1] and y[2] can be preceded by ¬, or/and the modal logic connectives 2, ♦ or by any permissible combinations of these symbols. For reasons given in [5], the symbols 2 and ♦ do not precede, except for the case “effort test not done or not decisive” (represented as said above as ♦¬x[2]), the variables x[i], i = 1, .., 9 and x[13] or negations of variables that represent symptoms. The relationship between the symptoms and the respective list of digits and symbols in 6, for instance, can be reinterpreted as the statement: “IF the effort test of a patient is positive AND he/she suffers from left common trunk disease AND his/her surgical risk is low/moderate AND his/her LVEF is less-equal 50% and over 30%, THEN the experts have assessed that PTCA is inappropriate but CABG is appropriate”; moreover, “there is agreement (A) on this assessment”. These “IF-THEN ” assertions can be translated into production rules as follows. • We write the symbol ∧ in the conclusion when at the end of the row of digits the pair of symbols composed of +, − and ? are −+ , +−, −−, −?, ?− , +? and ?+ (see item 4 below for instance). We write the symbol ∨ when the above pair are ++ (as in item 20 below) and ??. The reason for choosing these translations are given in [5].
128
L.M. Laita et al.
• The different degrees of appropriateness are translated as follows (for instance, if the experts assess that a therapeutic action y, is not appropriate under agreement, y − A, we propose the translation, “necessarily not y”: 2¬y) : y y y y y y y
− A ↔ 2¬y − I ↔ ¬y − D ↔ ♦¬y ? ↔ true or f alse, (see below) + D ↔ ♦y +I ↔y + A ↔ 2y
We translate the information items that contain “?” as follows. If the consequent refers to both PTCA and CABG, “?” is translated as “true” if the consequent is a conjunction and as false if it is a disjunction, except in the case where two “?”s appear, in which case the whole consequent is translated as “true”. The reason is that in this last case, the whole rule becomes a trivially true implication, whereas in the first and second cases, the consequent becomes PTCA and CABG whichever is not marked with “?”. For example, information item 4 is translated into a production rule as follows. R4: ¬x[1] ∧ x[2] ∧ x[3] ∧ x[7] → 2¬y[1] ∧ 2y[2]). Information 40 is translated as: R40: ¬x[1] ∧ x[2] ∧ x[5] ∧ ¬x[13] ∧ x[7] → true ∧ 2¬y[2]. An example of a rule that expresses disagreement (using ♦), is the one that translates item 20: R20: ¬x[1] ∧ x[2] ∧ x[4] ∧ x[8] → ♦y[1] ∨ ♦y[2]. Note that → is a reflexive and transitive relation. If denoted R, production rule R4 can be written as follows (¬x[1], x[2], x[3], x[7])R(2¬y[1], 2y[2]) . The set U is a set of tuples of symptoms that correspond to a set of classes of patients. The set of values Va is a set of ordered pairs of assessments of degrees of appropriateness y[...]. The relation admits a representation using three-valued and modal logic operators. The symbol ∨ is sometimes required, as in the relation that would correspond to R20. 2.3
CoCoA Implementation
The method follows the next three steps. (i)
To traduce IF-THEN statements into production rules. This has been already illustrated above.
A Computer Algebra Approach to Relational Systems Using Gr¨ obner Bases
129
(ii) To traduce production rules to polynomials. The theoretical process is described in [3,6,7]. We just give the results for three-valued logic with modal operators next: ¬x (NOT-x) is translated as 2 + 2x ♦x (POSSIBLY-x) is translated as 2x2 2x (NECESSARILY-x) is translated as x2 + 2x x ∨ y (x-OR-y) is translated as x2 y 2 + x2 y + xy 2 + 2xy + x + y x ∧ y (x-AND-y) is translated as 2x2 y 2 + 2x2 y + 2xy 2 + xy x → y (IF x-THEN-y) is translated as x2 y 2 + x2 y + xy 2 + 2x + 2. (iii) To apply Theorem 1. As advanced above, this theorem allows to perform the following two important steps. The foundations are in [1,10] and the tool used is the computer algebra language CoCoA 1 [2]. (1) To check automatically that the given set of logical formulae does not lead to inconsistencies. (2) To extract new information from the information contained in the mentioned set. Step (1) requires just to type a command (Gr¨ obner basis) in CoCoA . GBasis(I + J); If the output is [1], there is inconsistency. The inconsistencies can be – purely logical, as when one obtains, say, 2y[2] ∧ 2¬y[2] (“necessarily y[2]” and “necessarily-not y[2]”) from firing R4 and R40 – due to having found in the process of obtaining inferences, say some z[9] and some w[27], being these two pieces of information incompatible because the experts have determined so. This incompatibility is called an integrity constraint. Step (2) requires the use of Normal Forms. For instance, in order to check that, say, a propositional variable v[6] (or some formula, say x[8] → 2¬y[1]) follows from the RBES, it is enough to type, respectively: NF(NEG(v[6]), J + I); NF(NEG(IMP(x[8], NEC(NEG(y[1])))), J + I); Depending on whether the output is 0 or not, v[6] (resp. x[8] → 2¬y[1] ), follows or does not follow from the RBES. We have also implemented a program that exactly locates the inconsistencies. 1
CoCoA, a system for doing Computations in Commutative Algebra. Authors: A. Capani, G. Niesi, L. Robbiano. Available via anonimous ftp from: cocoa.dima.unige.it
130
3
L.M. Laita et al.
Railway Interlocking System
Clearance can be given simultaneously to more than one train. This action should not allow two trains to collide at any point (this is not a scheduling problem but a problem of compatibility of permissions). This is not a trivial problem if the network is not very small (Fig. 1). We shall consider that all trains are allowed to move at the same time in any direction at any speed, unless there is a signal forbidding the movement.
Fig. 1. Example of small station
The way the safety of the logical problem is treated is very similar to that used above to check consistency of RBES. In both cases what has to be checked is the degeneracy of an ideal of a polynomial residue class ring into the whole ring (what is done using Gr¨ obner bases). In this case the steps are railway situation → graph → (algebraic) decision model . Let us detail the model. Four oriented graphs (GD, GS, G∗ , G) are considered. In all four cases the vertices of the graph are the sections of the line, but the edges will be different, as explained below. Graph GD corresponds to the layout (and the position of the switches of the turnouts). There is an edge connecting section xi and section xj iff one of the following conditions holds: – sections xi and xj are consecutive in the line (i.e., they are two consecutive sections of a block-system) – there is a turnout connecting sections xi and xj and the switch is in the position that connects sections xi and xj – there is a turnout connecting sections xi and xj and the switch is in the position such that it is possible to pass from section xi to section xj trailing through this switch set against. Graph GS corresponds to semaphores. There is an edge connecting section xi with section xj iff there is a semaphore controlling the pass from section xi to section xj and it forbids such movement.
A Computer Algebra Approach to Relational Systems Using Gr¨ obner Bases
131
Therefore, a next section is accessible from another iff the layout (and the position of the switches of the turnouts) make it possible and the semaphores do not forbid it, i.e., iff there is an edge connecting them in G∗ = GD − GS. General accessibility is given by the transitive closure of the graph G∗ above, to be denoted G. Let us denote the sections by x1 , x2 , ..., xn (polynomial variables). The graph G will be interpreted as a polynomial ideal I ⊆ Q[x1 , x2 , ..., xn ] that is initialized as {0} . That it is possible to move from section xi to a (next) section xj will be represented by including the polynomial xi · (xi − xj ) in the ideal I. Trains will be denoted by positive integers. If train α is in section xi , xi − α will be added to the set where the position of trains is stored (P T ). Theorem 2. Safety: A situation of the switches and signals given by the ideal I and a position of trains given by the ideal < P T > is safe iff I+ < P T > = < 1 > (what can be checked using Gr¨ obner Bases). The idea in the background is that if, for example, train 11 is in section x3 and trains can move from section x3 to section x4 , then x3 − 11 = 0 and x3 · (x3 − x4 ) = 0. From both equations, x4 − 11 = 0, that is, section x4 is reachable by train 11. The whole package is implemented in the computer algebra language Maple [8] (the implementation is similar to that of section 3.3).
4
General Reflexive and Transitive Relations
Both models can be used to perform computations in the reflexive and transitive closure of general relations. 4.1
First Approach
The reflexive and transitive closure of a relation R, defined in a set C, can be approached through a polynomial model, the same way as a Boolean logic RBES was simulated using Theorem 1. Note that it is enough: – to treat the elements of C as the propositional variables of the RBES – to identify to be related to in C with the if-then statements of the RBES.
132
L.M. Laita et al.
Then the reflexive and transitive closure of R will behave exactly the same way as to be a tautological consequence in the RBES (although ∧ and ∨ do not have any use or translation). Moreover, in case the relation is an equivalence one, the quotient Boolean algebra can be identified with the corresponding (propositional) Lindenbaum algebra.
4.2
Second Approach
The reflexive and transitive closure of a relation R, defined in a set C, can also be approached through a polynomial model the same way as the railway interlocking system of the previous section. In this case it is enough: – to treat the elements of C as the sections of the line (identifying them with polynomial variables) – to identify to be related to in C with the edges of the main (and now only) oriented graph (G), by including the adequate xi · (xi − xj ) polynomial in the corresponding ideal. Then reflexive and transitive closure of R will behave exactly the same way as to be accesible. For instance, to check if a, b ∈ C are such that a is related to b in the reflexive and transitive closure of R, it is enough to add a − 1 to the ideal corresponding to G and to check whether b − 1 is also in it or not.
5
Conclusions
It has been illustrated, by using two examples, that Gr¨ obner Bases can be applied to the study of consistence of sets of reflexive and transitive relations. Automated extraction of knowledge, using Normal Forms can be performed in these sets. This information is related to appropriateness criteria in medicine in the first example, and to transit security in railway stations. Gr¨ obner Bases can be applied to any equivalence and partial order relations. In the case of RBES considered as RAS, it can be possible to extend the idea to the multi-valued information systems of [4] by using production rules as, say: x[1] ∧ x[2] ∧ x[3] ∧ ... ∧ x[k] → v[1] ∧ ... ∧ v[l] (allowing also the connective ∨ when necessary), where U is a set of tuples of single facts (that correspond to a class of objects, say sick persons) and {v[1], ..., v[l]} ∈ 2Va . The symbols for necessity and possibility can be used to translate the relations I and B.
A Computer Algebra Approach to Relational Systems Using Gr¨ obner Bases
133
References 1. V. Adams and P. Loustaunau, An Introduction to Gr¨ obner Bases. Graduate Studies in Mathematics 3. American Mathematical Society Press, Providence, RI (1994). 2. A. Capani, and G. Niesi, CoCoA User’s Manual (v. 3.0b), Dept. of Mathematics University of Genova (1996). 3. J. Chazarain, A. Riscos, J. A. Alonso and E. Briales, Multivalued Logic and Gr¨ obner Bases with Applications to Modal Logic. Journal of Symbolic Computation, 11 (1991) 181-194. 4. I. D¨ untsch, G. Gediga, E. Orlowska, Relational attribute systems. Preprint. 5. L. M. Laita, E. Roanes-Lozano, V. Maojo, L. de Ledesma, L. Laita, An Expert System for Managing Medical Appropriateness Criteria based on Computer Algebra Techniques. Computers and Mathematics with Applications, 42/12 (2001) 1505-1522. 6. L. M. Laita, E. Roanes-Lozano, L. de Ledesma and J. A. Alonso, A Computer Algebra Approach to Verification and Deduction in Many-Valued Knowledge Systems, Soft Computing 3-1 (1999) 7-19. 7. E. Roanes-Lozano, L. M. Laita and E. Roanes-Mac´ıas, A Polynomial Model for Multivalued Logics with a Touch of Algebraic Geometry and Computer Algebra, Mathematics and Computers in Simulation 45/1 (1998) 83-99. 8. E. Roanes-Mac´ıas and E. Roanes-Lozano, C´ alculos matem´ aticos por ordenador con Maple V.5. Ed. Rubi˜ nOS 1860, Madrid (1999). 9. E. Roanes-Lozano, E. Roanes-Mac´ıas, L.M. Laita, Railway Interlocking Systems and Gr¨ obner Bases, Mathematics and Computers in Simulation 51/5 (2000) 473482. 10. F. Winkler, Polynomial Algorithms in Computer Algebra. Springer, Wien (1996).
Fuzzy Relational Images in Computer Science Mike Nachtegael, Martine De Cock, Dietrich Van der Weken, and Etienne E. Kerre Fuzziness and Uncertainty Modelling Research Unit Department of Applied Mathematics and Computer Science Ghent University, Krijgslaan 281 (S9), 9000 Gent, Belgium {mike.nachtegael,martine.decock}@rug.ac.be, {dietrich.vanderweken,etienne.kerre}@rug.ac.be, http://fuzzy.rug.ac.be Abstract. Relations appear in many fields of mathematics and computer science. In classical mathematics these relations are usually crisp, i.e. two objects are related or they are not. However, many relations in real-world applications are intrinsically fuzzy, i.e. objects can be related to each other to a certain degree. With each fuzzy relation, different kinds of fuzzy relational images can be associated, all with a very practical interpretation in a wide range of application areas. In this paper we will explicite the formal link between well known direct and inverse images of fuzzy sets under fuzzy relations on one hand, and different kinds of compositions of fuzzy relations on the other. Continuing from this point of view we are also able to define a new scale of so-called double images. The wide applicability in mathematics and computer science of all these fuzzy relational images is illustrated with several examples.
1
Introduction
For many relations in the real world it is very difficult to obtain a justifiable crisp partition of the universe into those objects satisfying the relationship and those who do not. For example [18] suppose we have some set of patients P = {p1 , p2 , ..., pn } and some set of symptoms S = {s1 , s2 , ..., sm }. The first step in a diagnosis problem consists in identifying the set of symptoms related to each patient, i.e. in establishing a relation R from P to S. One of the symptoms may be for example fever (denoted f ). Classically only two situations may appear: a given patient p has fever f or not, depending on the temperature of his body. This presupposes that the concept of fever has been defined in a very crisp way, for example: p has f if and only if temperature(p) ≥ 38, and hence (p, f ) ∈ R if and only if temperature(p) ≥ 38 (p, f ) ∈ / R if and only if temperature(p) < 38 Using the same symbol R to denote the characteristic mapping of R, we can write (p, f ) ∈ R as R(p, f ) = 1, while (p, f ) ∈ / R can be written as R(p, f ) = 0. Now we know that there is a great difference between a body temperature of H. de Swart (Ed.): RelMiCS 2001, LNCS 2561, pp. 134–151, 2002. c Springer-Verlag Berlin Heidelberg 2002
Fuzzy Relational Images in Computer Science
135
38.2◦ C and one of 41◦ C. Nevertheless both situations are handled in the same way according to the model above. This unrealistic black-or-white distinction can be made more realistic by introducing a continuous scale for degrees of relationship. We may define for instance R(p, f ) = A(temperature(p)) with if x < 37 0 (x−37)2 2 9 if 37 ≤ x ≤ 38.5 A(x) = (1) 2 1 − 2 (x−40) if 38.5 ≤ x ≤ 40 9 1 if 40 < x for all x in the universe of temperatures T . I.e. we may define R as a P × S − [0, 1] mapping that associates with every patient-symptom couple (p, s) a number between 0 and 1, interpreted as the degree to which p and s are related. Such a mapping is called a fuzzy relation from P to S. The mapping A defined in (1) associates with every temperature x the degree between 0 and 1 to which x is considered to be “feverous”. Therefore this mapping is called the membership function of the fuzzy set “feverous”. Like relations are fundamental tools in classical set theory (also called crisp set theory), fuzzy relations are a basic concept of fuzzy set theory [28]. They appear in a wide range of fuzzy systems: for instance various kinds of fuzzy relations are used in fuzzy relational database systems (see e.g. [4], [27]); fuzzy tolerance relations, similarity relations and document description relations are used in information retrieval systems (see e.g. [5], [32]); preference relations are used in multicriteria decision support systems (see e.g. [14]); in approximate reasoning systems IF-THEN rules are modelled by means of fuzzy relations (see e.g. [24]), etc. Since fuzzy relations are so widely used it is not surprising that most tutorials on fuzzy set theory include a definition of the concept of a fuzzy relation as well as of some kind of composition of fuzzy relations (see e.g. [16], [18], [21], [33]). For instance if X, Y and Z are three universes, the sup-min composition of a fuzzy relation R1 from X to Y , and a fuzzy relation R2 from Y to Z is a fuzzy relation from X to Z denoted by R1 ◦ R2 and defined by (R1 ◦ R2 )(x, z) = sup min(R1 (x, y), R2 (y, z)), y∈Y
(2)
for all (x, z) in X × Z. However only few textbooks also mention some concept of image of a fuzzy set under a fuzzy relation [16], [18]. Since, as we will show, fuzzy relational images can and are used all over the place, it is quite remarkable that they are not more formalized and studied as a concept an sich. One of the oldest applications of fuzzy relational images is the compositional rule of inference [29], [30]. If A is a fuzzy set in X, R is a fuzzy relation from X to Y , and v1 and v2 are variables on X and Y respectively, this rule dictates v1 is A (*) (v1 , v2 ) is R (**) v2 is B (***)
136
M. Nachtegael et al.
i.e. from fact (*) and fact (**) we can derive fact (***) in which B is the fuzzy set in Y defined by (3) B(y) = sup min(A(x), R(x, y)), x∈X
for all y in Y . In [29], [30] B is quite ad hoc called “the composition of a unary fuzzy relation A and a binary fuzzy relation R”, somehow referring to (2). But rather than some composition, formula (3) is a straightforward generalization of the classical direct image R(A) of a crisp set A under a crisp relation R which is defined by R(A) = {y ∈ Y |(∃x ∈ X)(x ∈ A ∧ (x, y) ∈ R)} Although meanwhile a fuzzification of this image has been realized (see [6], [16], [17]), in recent works regarding approximate reasoning many authors (e.g. [1], [26], [31], [33]) still hang on to the compositional notation A ◦ R introduced in [29], [30]. In this paper we will explicite a mathematically correct connection between the direct image of A under R and the composition of fuzzy relations. It is inspired by an idea proposed in [15] on one hand, and the concept of cylindrical extension [30] on the other. We will do the same exercise for three other kinds of direct images, as well as for four kinds of inverse images. Continuing in the same line, we will also introduce several new kinds of double images of fuzzy sets under fuzzy relations. We will show that all these images are very powerful tools that can be used in many applications of computer science. The study of their definition and properties is therefore more than worth the while. This paper is structured as follows: in Section 2 we recall the definition of four direct and four inverse fuzzy relational images and some of their properties. In Section 3 we recall the definition of four kinds of fuzzy relational compositions and we are able to draw the mathematically correct link with the images of Section 2 due to the newly introduced concepts left and right extension of a fuzzy set. These extensions will also allow us to define several kinds of double images in Section 4. We end with examples from application fields in Section 5.
2
Fuzzy Relational Images
Throughout this paper X, Y and Z denote universes of discourse, i.e. classical (or “crisp”) sets containing the objects we want to say something about. In realworld applications, relations between sets of ‘objects’ are not always crisp by nature. For example, as clearly illustrated in the introduction of this paper, the patient-fever relation is an intrinsically vague relation. In order to cope with the vagueness involved in these kind of relations, the notion of fuzzy relations has been introduced, as a special kind of fuzzy set [28]. Definition 1 (Fuzzy set). A fuzzy set A in a universe X is characterized by a X − [0, 1] mapping, also denoted by A, that associates with every element x in X the degree to which x belongs to A. A(x) is therefore called the membership
Fuzzy Relational Images in Computer Science
137
degree of x in the fuzzy set A. The class of all fuzzy sets in X is denoted by F(X). Given two fuzzy sets A and B in X, their (Zadeh)-union A ∪ B and (Zadeh)intersection A ∩ B are fuzzy sets in X, defined by, for all x in X: (A ∪ B)(x) = max(A(x), B(x)) (A ∩ B)(x) = min(A(x), B(x)). The algebraic structure (F(X), ∩, ∪) is a lattice with the ordering defined by A ⊆ B if and only if (∀x ∈ X)(A(x) ≤ B(x)) for all A and B in F(X). A fuzzy set A is normal if there exists an x in X such that A(x) = 1. A fuzzy set in which 0 and 1 are the only membership values, is called a crisp set or a classical set. Definition 2 (Fuzzy relation). A fuzzy relation R from X to Y is a fuzzy set in X × Y . It is characterized by a X × Y − [0, 1] mapping, also denoted by R, that associates with every pair (x, y) in X × Y the degree to which x and y are related by R. If X = Y , then R is called a binary fuzzy relation in X. The inverse fuzzy relation of R is the fuzzy relation R−1 from Y to X defined by R−1 (y, x) = R(x, y) for all x in X and y in Y . Definition 3 (Foreset, afterset). [3] Let R be a fuzzy relation from X to Y . For all x in X, the R-afterset xR of x is the fuzzy set on Y defined as xR(y) = R(x, y), for all y in Y . Likewise, for all y in Y , the R-foreset Ry of y is the fuzzy set on X defined as Ry(x) = R(x, y), for all x in X. An important role in the definitions of fuzzy relational images is played by logical operators on [0, 1] which are an extension of logical conjunction and implication on {0, 1}. Definition 4 (Conjunctor). A conjunctor on [0, 1] is a [0, 1] × [0, 1] − [0, 1] mapping T that satisfies T (0, 0) = T (0, 1) = T (1, 0) = 0 and T (1, 1) = 1, and that is increasing in both arguments. It is called a semi-norm if T (a, 1) = T (1, a) = a, for all a in [0, 1]. A semi-norm is called a t-norm if it is also commutative and associative. ukasiewicz t-norm The minimum TM , the algebraic product TP and the L TW (also called bounded sum) are very popular t-norms. They are given by TM (a, b) = min(a, b), TP (a, b) = a·b and TW (a, b) = max(0, a+b−1), respectively. Definition 5 (Implicator). An implicator on [0, 1] is a [0, 1] × [0, 1] − [0, 1] mapping I that satisfies I(0, 0) = I(0, 1) = I(1, 1) = 1 and I(1, 0) = 0, and that is decreasing in the first and increasing in the second argument. It is called a border implicator if I(1, a) = a, for all a in [0, 1]. The L ukasiewicz implicator IL , the Kleene-Dienes implicator IKD and the Reichenbach implicator IR are among the most popular border implicators. They
138
M. Nachtegael et al.
are respectively given by IL (a, b) = min(1, 1 − a + b), IKD (a, b) = max(1 − a, b) and IR (a, b) = 1 − a + a · b. The first condition in the definitions of conjunctors and implicators guarantees that these operators are extensions of their classical counterparts. The monotonicity conditions guarantee that the operators have a logical interpretation. Using a conjunctor T , the notions of reflexivity, symmetry and T -transitivity can be defined for binary fuzzy relations in the following way: Definition 6. Let R be a binary fuzzy relation in X. R is called – reflexive iff R(x, x) = 1, for all x in X – symmetric iff R(x1 , x2 ) = R(x2 , x1 ), for all x1 and x2 in X – sup-T transitive iff R(x1 , x2 ) ≥ sup T (R(x1 , z), R(z, x2 )), z∈X
for all x1 and x2 in X
Definition 7 (Degree of overlap). Let T be a conjunctor. For two fuzzy sets A and B in X, the degree of T -overlap of A and B is given by sup T (A(x), B(x))
x∈X
Definition 8 (Degree of inclusion). [2] Let I be an implicator. For two fuzzy sets A and B in X, the degree of I-inclusion of A in B is given by inf I(A(x), B(x))
x∈X
Definition 9 (Degree of equality). [2] Let T be a conjunctor and I an implicator. For two fuzzy sets A and B in X, the degree of (T , I)-equality of A and B is given by T ( inf I(A(x), B(x)), inf I(B(x), A(x))) x∈X
x∈X
If R is a crisp relation from X to Y , it can map a crisp subset A of X onto a crisp subset B of Y in many different ways. Interesting ones are: – The direct image of A under R, which is the set of objects in Y that are related to at least one object of A. – The subdirect image of A under R, which is the set of objects in Y that are related to all objects of A. – The superdirect image of A under R, which is the set of objects in Y that are related only to objects of A. – The square direct image of A under R, which is the set of objects in Y that are related to all and only to objects of A. In other words it is the set of objects typically related to A. Fuzzification of these images leads to the following definition [6], [17]: Definition 10 (Direct images). For T a conjunctor, I an implicator, R a fuzzy relation from X to Y , and A a fuzzy set in X, the direct image RT (A), the subdirect image RI (A), the superdirect image RI (A) and the square direct image RT ,I (A) of A under R are the fuzzy sets in Y defined by, for all y in Y ,
Fuzzy Relational Images in Computer Science
139
– RT (A)(y) = sup T (A(x), R(x, y)) (direct image) x∈X
– RI (A)(y) = inf I(A(x), R(x, y)) (subdirect image) x∈X
– RI (A)(y) = inf I(R(x, y), A(x)) (superdirect image) x∈X
– RT ,I (A)(y) = T (RI (A)(y), RI (A)(y)) (square direct image) That Definition 10 corresponds to the semantics of the images described above becomes more clear when viewing R(x, y) as Ry(x), i.e. the degree to which x belongs to the R-foreset of y. Indeed Ry is actually the fuzzy set of objects related to y. So when determining the degree to which y belongs to the direct image of A under R, i.e. the degree to which y is related to at least one object of A, one looks at the degree of overlap of A and Ry. Likewise when determining the degree to which y belongs to the subdirect image of A under R, i.e. the degree to which y is related to all objects of A, one looks at the degree to which A is included in Ry etc. Note that it is a changement in the position of the arguments of I in the definition of subdirect image which gives rise to that of superdirect image. A similar thing can be done for the direct image, leading to another image RT (A) defined by RT (A)(y) = sup T (R(x, y), A(x)) x∈X
for all y in Y . This image was included in the study in [11]. Most common conjunctors T are however commutative. One can easily verify that for these kind of conjunctors RT (A) = RT (A). If R is a relation from X to Y and B is a set in Y it also makes sense to talk about the set of objects of X related to at least one object of B, related to all objects of B, only related to objects of B, etc. This gives rise to the definition of several inverse images [6]. Definition 11 (Inverse images). For T a conjunctor, I an implicator, R a fuzzy relation from X to Y , and B a fuzzy set in Y , we say that – – – –
(R−1 )T (B) is the inverse image of B under R (R−1 )I (B) is the subinverse image of B under R (R−1 )I (B) is the superinverse image of B under R (R−1 )T ,I (B) is the square inverse image of B under R
The above definitions have a certain degree of freedom: for every choice of the conjunctor T and the implicator I, a fuzzy relational image is obtained. This freedom of choice is reflected in the notation. In theoretical studies one can focus on the direct fuzzy relational images, since the inverse images w.r.t. R can be regarded as direct images w.r.t. R−1 . The fuzzy relational images defined in Definition 10 and 11 have an important characteristic which was pointed out in [6], namely the lack of non-emptiness conditions. Indeed if the fuzzy set A is almost empty, i.e. all the membership degrees in A are small, then RI (A)(y) tends to be high. In the extreme case when A = ∅, the subdirect image of A under every imaginable fuzzy relation
140
M. Nachtegael et al.
R will be the universe for most common implicators I. A similar phenomenon occurs with the superdirect image, and of course then also with the square direct image. To avoid this the following images have been defined [6]: Definition 12 (Kerre direct images). For T a t-norm, I an implicator, R a fuzzy relation from X to Y , and A a fuzzy set in X, the Kerre subdirect image RTk,I (A), the Kerre superdirect image RTk,I (A) and the Kerre square direct image RTk,I (A) of A under R are the fuzzy sets in Y defined by – RTk,I (A) = RI (A) ∩ RT (A) (Kerre subdirect image) – RTk,I (A) = RI (A) ∩ RT (A) (Kerre superdirect image) – RTk,I (A) = RT ,I (A) ∩ RT (A) (Kerre square direct image) The definition in [6] requires T to be a t-norm. If T would have been a conjunctor in general it would be more natural to involve the alternative direct image RT (A) defined above in the definition of the Kerre subdirect image and square direct image. Proposition 1. Let T be a t-norm and I a border implicator. Furthermore let R be a fuzzy relation from X to Y and A a fuzzy set in X. 1. If A is normal then RTk,I (A) = RI (A). 2. If all R-foresets are normal then RTk,I (A) = RI (A). 3. If A is normal and all R-foresets are normal then RTk,I (A) = RT ,I (A) Proof. If A is normal there is an x0 in X such that A(x0 ) = 1. For all y in Y : RI (A)(y) = inf I(A(x), R(x, y)) ≤ I(A(x0 ), R(x0 , y)) x∈X
≤ T (A(x0 ), R(x0 , y)) ≤ sup T (A(x), R(x, y)) = RT (A)(y) x∈X
Therefore RI (A) ⊆ RT (A) and hence 1. If all R-foresets are normal likewise we can prove that RI (A) ⊆ RT (A) and hence 2. If both the conditions of 1 and 2 are fulfilled we have for all y in Y : T (RI (A)(y), RI (A)(y)) ≤ min(RI (A)(y), RI (A)(y)) ≤ RT (A)(y) and hence 3.
The conditions stated in Proposition 1 are rather weak. For example, the condition on R will be satisfied if R is a reflexive binary fuzzy relation in X, and most common conjunctors and implicators are indeed t-norms and border implicators. As we illustrate in Section 5, many applications live by the conditions of Proposition 1, so throughout the remainder of this paper we will focus on the fuzzy relational images defined in Definition 10. The properties of these images are studied in detail in [6], [11], [12]. We summarize some of the results with interesting interpretation for the applications. Proposition 2 (Monotonicity). Let A be a fuzzy set in X, R a fuzzy relation from X to Y , T a conjunctor and I an implicator.
Fuzzy Relational Images in Computer Science
141
– The direct and the square direct image are increasing w.r.t. the conjunctor. The subdirect, the superdirect and the square direct image are increasing w.r.t the implicator. I.e. if T1 and T2 are conjunctors such that T1 ≤ T2 and I1 and I2 are implicators such that I1 ≤ I2 , then: RT1 (A) ⊆ RT2 (A) RI 1 (A) ⊆ RI 2 (A)
RI 1 (A) ⊆ RI 2 (A) RT 1 ,I (A) ⊆ RT 2 ,I (A) RT ,I1 (A) ⊆ RT ,I2 (A). – The direct and the subdirect image are increasing w.r.t. the fuzzy relation, while the superdirect image is decreasing w.r.t. the fuzzy relation. I.e. if R1 and R2 are fuzzy relations from X to Y such that R1 ⊆ R2 , then: (R1 )T (A) ⊆ (R2 )T (A) (R1 )I (A) ⊆ (R2 )I (A)
(R1 )I (A) ⊇ (R2 )I (A).
– The direct and the superdirect image are increasing w.r.t. the fuzzy set, while the subdirect image is decreasing w.r.t the fuzzy set. I.e. if A1 and A2 are fuzzy sets in X such that A1 ⊆ A2 , then: RT (A1 ) ⊆ RT (A2 ) RI (A1 ) ⊆ RI (A2 )
RI (A1 ) ⊇ RI (A2 ).
Proposition 2 expresses that we can strengthen or weaken the direct images by adjusting the logical operators. It also expresses that more specific relations or more specific fuzzy sets lead to less or more specific direct images. For example, let A1 be the set of male corpulent patients, and let A2 be the set of corpulent patients (male or female). The superdirect image RI (A1 ) is the fuzzy set of symptoms that only occur with the male corpulent patients, while RI (A2 ) is the fuzzy set of symptoms that only occur with male or female corpulent patients. It is clear that the latter set will not be smaller than the first one. From the proof of Proposition 1 we know that the superdirect image is a subset of the direct image if all R-foresets are normal. As we mentioned above this is the case for every reflexive binary fuzzy relation R. The direct images of a fuzzy set A in X under such a fuzzy relation R again are fuzzy sets in X, which makes it meaningful to study a possible ordering between the original fuzzy set A and its images. In particular we have the following proposition: Proposition 3 (Expansiveness and restrictiveness). Let T be a semi-norm and I a border implicator. Furthermore let R be a binary fuzzy relation in X and A a fuzzy set in X. If R is reflexive, then: RT ,I (A) ⊆ RI (A) ⊆ A ⊆ RT (A)
142
M. Nachtegael et al.
Proposition 3 shows that the square and the superdirect images are smaller (i.e. more specific) than the original fuzzy set, but that the direct image is larger (i.e. less specific) than the original fuzzy set. Proposition 4 (Interaction with union of fuzzy sets). Let T be a conjunctor and I be an implicator. For a fuzzy relation R from X to Y and fuzzy sets A1 and A2 in X it holds that: RT (A1 ∪ A2 ) = RT (A1 ) ∪ RT (A2 ) RI (A1 ∪ A2 ) ⊆ RI (A1 ) ∪ RI (A2 ) RI (A1 ∪ A2 ) ⊇ RI (A1 ) ∪ RI (A2 )
The direct image of the union of two fuzzy sets can be expressed as the union of the direct images of the separate fuzzy sets. For the other types of direct images only a containment relation w.r.t. the union of the separate direct images holds. The property RT (A1 ∪ A2 ) = RT (A1 ) ∪ RT (A2 ) is very interesting from a practical and computational point of view. For example, let A1 be the set of corpulent male patients, and let A2 be the set of corpulent female patients. One can then compute the direct images RT (A1 ) and RT (A2 ) as respectively the fuzzy set of symptoms that occur with corpulent male patients and the fuzzy set of symptoms that occur with corpulent female patients. Without many other calculations, one can now easily compute the direct image RT (A1 ∪ A2 ), which is the fuzzy set of symptoms that occur with corpulent patients (both male or female).
3
Fuzzy Relational Compositions
Definition 13 (Compositions). [3], [6], [28] For R in F(X × Z) and S in F(Z × Y ), the sup-T -composition and the Bandler-Kohout compositions of R and S are fuzzy relations from X to Y defined as, for all x in X and all y in Y : – R ◦T S (x, y) = sup T (R(x, z), S(z, y)) (sup-T -composition) z∈Z
– R I S (x, y) = inf I(R(x, z), S(z, y)) (subproduct) z∈Z
– R I S (x, y) = inf I(S(z, y), R(x, z)) (superproduct) z∈Z
– R T ,I S (x, y) = T (R I S(x, y), R I S(x, y)) (squareproduct) Following the principle of cylindrical extension we define the extension operators ← and . .
→ .
Definition 14 (Left extension, right extension). The left extension oper→ ← ator . and the right extension operator . are F(X) − F(X × X) mappings that →
←
turn every fuzzy set A in X into the binary fuzzy relations A and A in X defined by
Fuzzy Relational Images in Computer Science
143
←
2 A (x, y) = A(y), for all (x, y) in X
→
2 A (x, y) = A(x), for all (x, y) in X
←
→
A and A are called the left and the right extension of A respectively.
Characterization 1 The binary fuzzy relations R1 and R2 in X are respectively the left and the right extensions of the fuzzy set A in X if and only if xR1 = A, for all x in X R2 y = A, for all y in X Since the left extension of a fuzzy set A in X is a binary fuzzy relation in X, it makes sense to study its compositions with a fuzzy relation R from X to Y . Likewise we can study the compositions of R with the right extension of a fuzzy set B in Y . Proposition 5. Let T be a conjunctor and I an implicator, R in F(X × Y ), A in F(X) and B in F(Y ). For all x in X and y in Y it holds that: ←
(1) A ◦T R(x, y) = RT (A)(y) ← (2) A I R(x, y) = RI (A)(y) ←
→
(5) R◦T B (x, y) = (R−1 )T (B)(x) → (6) RI B (x, y) = (R−1 )I (B)(x) →
(3) A I R(x, y) = RI (A)(y) (7) RI B (x, y) = (R−1 )I (B)(x) ← → (4) A T ,I R(x, y) = RT ,I (A)(y) (8) RT ,I B (x, y) = (R−1 )T ,I (B)(x)
Proof. The proposition above can be proven by applying the definitions of the concepts involved. As an example we prove (1) and (6). ←
←
A ◦T R(x, y) = sup T (A (x, z), R(z, y)) z∈X
= sup T (A(z), R(z, y)) = RT (A)(y) →
z∈X
→
RI B (x, y) = inf I(R(x, z), B (z, y)) z∈X
= inf I(R(x, z), B(z)) = (R−1 )I (B)(x) z∈X
Hence the compositions of the left extension of A and R give rise to the direct, the subdirect, the superdirect and the square direct image of A under R, while the compositions of R and the right extension of B result in the inverse, the superinverse, the subinverse and the square inverse image of B under R. Note → that it is the subproduct of B and R that results in the superinverse image → of B under R, and that it is the superproduct of B and R that results in the subinverse image of B under R. Furthermore note that while the left hand sides of the equalities in Proposition 5 seem to be depending on x and y the right hand sides only depend on either x or y.
144
4
M. Nachtegael et al.
Double Images
If A is a fuzzy set in X and B is a fuzzy set in Y their left and right extension will respectively be a binary fuzzy relation in X and a binary fuzzy relation in ← Y . The compositions of A with a fuzzy relation R from X to Y result in fuzzy → relations from X to Y which we can once again compose with B . Likewise the → compositions of R and B lead to fuzzy relations from X to Y which can be ← composed with A. This is precisely what we study in this section. Proposition 6. Let T be a conjunctor and I an implicator. Furthermore let R be a fuzzy relation from X to Y , A a fuzzy set in X and B a fuzzy set in Y . For all x in X and y in Y it holds that: ← → (1)((A ◦T R)◦T B )(x, y) = sup T (RT (A)(z), B(z)) ←
z∈Y
→
(2)(A ◦T (R◦T B ))(x, y) = sup T (A(z), (R−1 )T (B)(z)) ←
z∈X
→
(3)((A I R)I B )(x, y) = inf I(RI (A)(z), B(z)) ←
z∈Y
→
(4)(A I (RI B ))(x, y) = inf I(A(z), (R−1 )I (B)(z)) ←
z∈X
→
(5)((A I R)I B )(x, y) = inf I(B(z), RI (A)(z)) ←
z∈Y
→
(6)(A I (RI B ))(x, y) = inf I((R−1 )I (B)(z), A(z)) ←
z∈X
→
(7)((A T ,I R)T ,I B )(x, y) = T ( inf I(B(z), RT ,I (A)(z)), inf I(RT ,I (A)(z), B(z))) ←
z∈Y
z∈Y
→
(8)(A T ,I (RT ,I B ))(x, y) = T ( inf I(A(z), (R−1 )T ,I (B)(z)), inf I((R−1 )T ,I (B)(z), A(z))) z∈X z∈Y Proof. All equalities can be proven using Definition 10 and 11 and Proposition 5. As an example we prove (2) and (7). ←
→
→
(A ◦T (R◦T B ))(x, y) = (R◦T B ))T (A)(y) → = sup T (A(z), R◦T B (z, y)) z∈X
= sup T (A(z), (R−1 )T (B)(z)) ←
z∈X
→
←
((A T ,I R)T ,I B )(x, y) = ((A T ,I R)−1 )T ,I (B)(x) ← ← = T (((A T ,I R)−1 )I (B)(x), ((A T ,I R)−1 )I (B)(x)) ←
←
= T ( inf I(B(z), A T ,I R(x, z)), inf I(A T ,I R(x, z), B(z))) z∈Y
z∈Y
= T ( inf I(B(z), RT ,I (A)(z)), inf I(RT ,I (A)(z), B(z))) z∈Y
z∈Y
All compositions presented above result in constant fuzzy relations. In [15] ←
→
(A ◦T R)◦T B
is called the double image of A and B under R. It is “the extent to which A and B are related under R”. The analogy with the preceding section inspires the following definition:
Fuzzy Relational Images in Computer Science
145
Definition 15 (Double images). For T a conjunctor, I an implicator, R a fuzzy relation from X to Y , A a fuzzy set in X, and B a fuzzy set in Y we say that ←
→
– (A ◦T R)◦T B is the double direct image of A and B under R ←
→
– A ◦T (R◦T B ) is the double inverse image of A and B under R ←
→
– (A I R)I B is the double subdirect image of A and B under R ←
→
– A I (RI B ) is the double subinverse image of A and B under R ←
→
– (A I R)I B is the double superdirect image of A and B under R ←
→
– A I (RI B ) is the double superinverse image of A and B under R ←
→
– (A T ,I R)T ,I B is the double square direct image of A and B under R ←
→
– A T ,I (RT ,I B ) is the double square inverse image of A and B under R The following question which naturally comes to mind is whether it is possible to combine two kinds of fuzzy relational composition into one new kind of double image. As can be expected this leads to mixed double images such as the ones listed below, derived using Definition 10 and 11 as well as Proposition 5: ←
→
←
→
– ((A ◦T R)I B )(x, y) = inf I(RT (A)(z), B(z)) z∈Y
– ((A ◦T R)I B )(x, y) = inf I(B(z), RT (A)(z)) ←
→
z∈Y
– (A I (RI B ))(x, y) = inf I(A(z), (R−1 )I (B)(z)) – etc.
z∈Y
The main idea for all the double images is similar: A is a fuzzy set in X, B is a fuzzy set in Y , and R is a fuzzy relation from X to Y . If you want to know the degree to which A and B are related under R, first you transform either A into a fuzzy set on Y (using one of the direct images of Definition 10), or B into a fuzzy set on X (using one of the inverse images of Definition 11). Then you are left with two fuzzy sets on the same universe, which you can compare by looking to their degree of overlap, their degree of inclusion or their degree of equality.
5
A Look at Some Applications
So far we have illustrated the use of the fuzzy relational images in a medical data mining setting. In this section we will briefly discuss some other kinds of applications that can be described using fuzzy relational images, in particular fuzzy morphology, fuzzy rough sets, linguistic modifers and an application of the double images. An application in the context of modal logic has already been discussed in [13].
146
M. Nachtegael et al.
5.1
Fuzzy Morphology
Mathematical morphology is a theory for the processing and analysis of images, using operators based on geometrical concepts [25]. Different theories have been developed for binary images (represented as subsets of Rn ) and gray-scale images (represented as fuzzy subsets of Rn ). In the past years, fuzzy morphology has been introduced as an extension of binary morphology to gray-scale morphology, using techniques from fuzzy set theory [8], [19], [20]. The basic morphological operations are dilation and erosion. These operations transform an image A into a new image A by means of a structuring element S. Definition 16 (Fuzzy dilation and erosion). Let T be a t-norm, I a border implicator, and A and S fuzzy sets (gray-scale images) in Rn . The fuzzy dilation DT (A, S) and the fuzzy erosion EI (A, S) are defined by, for all y ∈ Rn : DT (A, S)(y) = sup T (S(x − y), A(x)) x∈Rn
EI (A, S)(y) = infn I(S(x − y), A(x)) x∈R
In many image processing applications it is required that S(0) = 1, i.e. the origin completely belongs to the structuring element. Now, if V is the substraction in Rn , i.e. V (x, y) = x − y for all x and y in Rn , then this implies that the relation R = V ◦ S is a reflexive binary fuzzy relation in Rn . In that case we have: DT (A, S) = RT (A), EI (A, S) = RI (A). Note the nice interpretation of Proposition 4: the fuzzy dilation DT (A, S) can be obtained by partitioning A into subsets Ai (i = 1, ..., n), by (simultaneously) computing the fuzzy dilations DT (Ai , S) (i = 1, ..., n), and by combining the obtained results. This property can be used to decrease computation time. 5.2
Fuzzy Rough Sets
Rough set theory was introduced in the field of knowledge representation for modelling incomplete information [22]. Later on it was generalized to fuzzy rough set theory which can be used to model imprecise information as well (see e.g. [23]). Given a reflexive binary fuzzy relation R on a universe X of objects, the pair (X, R) is called a fuzzy approximation space. The basic operations in fuzzy rough set theory are the upper and lower fuzzy rough approximation. Definition 17 (Upper and lower fuzzy rough approximation). Let T be a t-norm, I a border implicator, (X, R) a fuzzy approximaton space and A a T fuzzy set in X. The upper fuzzy rough approximation Fas (A) and the lower fuzzy rough approximation FasI (A) of A are given by, for all y ∈ X: T
Fas (A)(y) = sup T (R(x, y), A(x)), x∈X
FasI (A)(y) = inf I(R(x, y), A(x)). x∈X
Fuzzy Relational Images in Computer Science
147
T
Fas (A) is the fuzzy set of objects that possibly belong to A while FasI (A) is the fuzzy set of objects that necessarily belong to A. The link with the fuzzy direct and the fuzzy superdirect image is obvious, namely: T
Fas (A) = RT (A), FasI (A) = RI (A). Proposition 3 guarantees that the lower fuzzy rough approximation of A is a fuzzy subset of A, while A is a fuzzy subset of its upper fuzzy rough approximation. 5.3
Linguistic Modifiers
If a linguistic term is represented by a fuzzy set A in a universe X, then a linguistic modifier (i.e. an adverb) can be modelled by a F(X) − F(X) mapping, called a fuzzy modifier. Applying a fuzzy modifier to the fuzzy set modelling a linguistic term then yields the representation of the modified linguistic term; for instance applying a fuzzy modifier for very onto a fuzzy set for tall gives rise to the membership function for very tall. In fuzzy systems two kinds of interpretations of linguistic modifiers such extremely, very, more or less and roughly are common. In the first one, called the inclusive interpretation, it is assumed that extremely A ⊆ very A ⊆ A ⊆ more or less A ⊆ roughly A
(4)
In [9] it is suggested to represent the linguistic modifiers by means of resemblance relations, i.e. reflexive binary fuzzy relations that model approximate equality. Let E1 and E2 be two resemblance relations in X and E1 ⊂ E2 (hence the approximate equality modelled by E1 is stronger than the approximate equality modelled by E2 ). For T a t-norm and I a border implicator the following representations are proposed (for all A in F(X)) : more or less A = (E1 )T (A) very A = (E1 )I (A) roughly A = (E2 )T (A) extremely A = (E2 )I (A) These representations can be interpreted like “y belongs to more or less A if y resembles to an object of A” and “y belongs to very A if y resembles only to objects of A”. Propositions 2 and 3 guarantee that the inclusions in (4) hold. In the non-inclusive interpretation which is often used in fuzzy control applications, a modified linguistic term does not denote a subset nor a superset of the original term. The original term and the modified term denote two different (possibly overlapping categories). If the universe X is numerical this situation can be modelled by shifting fuzzy modifiers that shift the original fuzzy set to the left or to the right. In [9] it was demonstrated that shifting modifiers can be seen as operators that take direct images under a fuzzy relation Rα defined as 1 if x = y − α Rα (x, y) = 0 otherwise for all x and y in X. The value of the numerical parameter α determines the size of the shift as well as its direction.
148
5.4
M. Nachtegael et al.
Applications of the Double Images
As far as we know until now the power of double images for applications in computer science has not yet been investigated. We suggest some examples: – If X is a universe of people, A and B are fuzzy sets in X and the binary fuzzy relation R in X expresses the degree of friendship between two people, then the double direct image of A and B under R can be interpreted as the degree to which group A and group B are friends. – If R is a fuzzy patient-symptom relation, A is the fuzzy set of people that live nearby a nuclear powerplant and B is the set of fuzzy set of symptoms for cancer then the subdirect double image of A and B under R expresses the degree to which the fuzzy set of symptoms that occur with all patients that live nearby a plant is included in the fuzzy set of symptoms of cancer. – If R is a fuzzy relation between the universe of terms X and the universe of documents Y in a document retrieval system, if A is the fuzzy set of terms present in a profile for a research vacancy and B is the set of publications of a particular researcher, then the double direct image of A and B under R expresses the degree to which this researcher matches the vacancy. A particularly interesting case arises when we denote the crisp equality on X by E, i.e. E(x, y) = 1 if x = y and E(x, y) = 0 otherwise. Indeed then one can easily verify that ET (A) = A for all A in F(X). For all A and B in F(X) it then holds that ←
→
– (A ◦T E)◦T B is the degree of T -overlap of A and B ← → – (A ◦T E)I B is the degree of I-inclusion of A in B ←
→
– (A ◦T E)T ,I B is the degree of (T , I)-equality of A and B By (slightly) changing the values of the membership degrees in E, new and refined measures of similarity and inclusion can be defined. One option is to define E(x, x) = 0 for the objects x in an uninteresting part of the universe. Another option is to use a fuzzy relation instead of a crisp one to model equality (see [10]). The advantages of such an approach become clear in the following example: suppose we want to compare the first two pictures depicted in Figure 1. When representing them as fuzzy sets A and B, all pixels that have membership degree 1 in A (namely the white pixels) have degree 0 in B (black pixels) and vice versa. Hence for all T and I the degree of (T , I)-equality of A and B will be 0. However if we model approximate equality on the pixel level by a fuzzy relation E1 such that every pixel is equal to itself to degree 1 and neighbouring pixels are equal to each other to degree 0.5, using min as a t-norm, the direct image of picture a of Figure 1 under E1 looks like picture c. The white pixels of the original picture remain white but the black pixels of the original picture have gotten a higher membership degree in the new one (namely 0.5) and therefore ←
→
look gray. (A ◦min E1 )T ,I B might be greater than 0. For example, using the minimum operator as a t-norm and the Kleene-Dienes implicator, this mixed double image of A and B under E1 is 0.5. If needed we can even use a fuzzy
Fuzzy Relational Images in Computer Science
149
relation such that we obtain an equality degree of 1. Such a result is desirable in the context of measuring similarity between images (two images that have been shifted w.r.t. each other are similar images). To achieve this result we can choose a fuzzy relation such that the direct image acts like a shifting modifier as briefly discussed in the preceding subsection.
Fig. 1. Pictures a, b and c all containing 11 by 17 pixels
References 1. H. Bandemer, S. Gottwald (1995). Fuzzy Sets, Fuzzy Logic, Fuzzy Methods. John Wiley and Sons, Chisester. 2. W. Bandler, L. J. Kohout (1980). Fuzzy Power Sets and Fuzzy Implication Operators. Fuzzy Sets and Systems, 4, p. 13–30. 3. W. Bandler, L. J. Kohout (1980). Fuzzy Relational Products as a Tool for Analysis and Synthesis of the Behaviour of Complex Natural and Artificial Systems. In: Fuzzy Sets: Theory and Application to Policy Analysis and Information Systems S. K. Wang, P. P. Chang (eds.). Plenum Press, New York and London, p. 341–367. 4. G. Chen (1993). Fuzzy Data Modeling: Perspectives, Problems and Solutions. In: [18]. 5. F. Crestani, G. Pasi (eds.) (2000). Soft Computing in Information Retrieval, Techniques and Applications. Studies in Fuzziness and Soft Computing Vol. 50. PhysicaVerlag Heidelberg/New York. 6. B. De Baets (1995). Solving Fuzzy Relational Equations: an Order Theoretical Approach. Ph.D. Thesis, Ghent University (in Dutch). 7. B. De Baets, E. E. Kerre (1993). A Revision of Bandler-Kohout Compositions of Relations. Mathematica Pannonica, 4, p. 59–78. 8. B. De Baets (1997). Fuzzy Morphology: a Logical Approach. In: Uncertainty Analysis in Engineering and Sciences: Fuzzy Logic, Statistics, and Neural Network Approach. Kluwer Academic Publishers, Boston, p. 53–67. 9. M. De Cock, E. E. Kerre (2000). A New Class of Fuzzy Modifiers. In: Proceedings of 30th IEEE International Symposium on Multiple-Valued-Logic (ISMVL’2000), IEEE Computer Society, p. 121-126.
150
M. Nachtegael et al.
10. M. De Cock, E. E. Kerre (2002). On (Un)suitable Fuzzy Relations to Model Approximate Equality. To appear in: Fuzzy Sets and Systems. 11. M. De Cock, M. Nachtegael, E. E. Kerre (2000). Images under Fuzzy Relations: a Master-Key to Fuzzy Applications. In: Intelligent Techniques and Soft Computing in Nuclear Science and Engineering (D. Ruan, H. A. Abderrahim, P. D’hondt, E. E. Kerre (eds.)). World Scientific Publishing, Singapore, p. 47–54. 12. M. De Cock, M. Nachtegael, E. E. Kerre (2000). Images under Fuzzy Relations: Technical Report FUM-2000-02 (22 pages). Fuzziness and Uncertainty Modelling Research Unit, Ghent University (in Dutch). 13. M. De Cock, A. M. Radzikowska, E. E. Kerre (2002). A Fuzzy-Rough Approach to the Representation of Linguistic Hedges. In: Technologies for Constructing Intelligent Systems, Part 1: Tasks (B. Bouchon-Meunier, J. Gutierrez-Rios, L. Magdalena, R. R. Yager (eds.)). Springer-Verlag, Heidelberg, p. 33–42. 14. J. Fodor, M. Roubens (1994). Fuzzy Preference Modelling and Multicriteria Decision Support. Kluwer Academic Publishers Dordrecht/Boston/London. 15. J. A. Goguen (1967). L-Fuzzy Sets. Journal of Mathematical Analysis and Applications, 18, p. 145–174. 16. S. Gottwald (1993). Fuzzy Sets and Fuzzy Logic. Vieweg, Braunschweig. 17. E. E. Kerre (1992). A Walk through Fuzzy Relations and their Applications to Information Retrieval, Medical Diagnosis and Expert Systems. In: Analysis and Management of Uncertainty: Theory and Applications (B. M. Ayyub, M. M. Gupta, L. N. Kanal (eds.)). Elsevier Science Publishers, p. 141–151. 18. E. E. Kerre (ed.) (1993). Introduction to the Basic Principles of Fuzzy Set Theory and Some of its Applications. Communication and Cognition, Gent. 19. M. Nachtegael, E. E. Kerre (1999). Different Approaches towards Fuzzy Mathematical Morphology. In: Proceedings of EUFIT’99 (on CD-ROM), Aachen. 20. M. Nachtegael, E. E. Kerre (2000). Classical and Fuzzy Approaches towards Mathematical Morphology. In: Fuzzy Techniques in Image Processing (E. E. Kerre, M. Nachtegael (eds.)). Springer-Verlag, Heidelberg, p. 3–57. 21. V. Novak (1986). Fuzzy Sets and Their Applications. Adam Hilger, Bristol. 22. Z. Pawlak (1982). Rough sets. International Journal of Computer and Information Science, 11, p. 341–356. 23. A. M. Radzikowska, E. E. Kerre (2000). A Comparative Study of Fuzzy Rough Sets. To appear in Fuzzy Sets and Systems. 24. D. Ruan, E. E.Kerre (eds.) (2000). Fuzzy IF-THEN Rules in Computational Intelligence. Kluwer Academic Publishers Boston/Dordrecht/London. 25. J. Serra (1982). Image Analysis and Mathematical Morphology. Academic Press, London. 26. J. F.-F. Yao, J.-S. Yao (2001). Fuzzy Decision Making for Medical Diagnosis based on Fuzzy Number and Compositional Rule of Inference. Fuzzy Sets and Systems, 120, p. 351–366. 27. A. Yazici, R. George (1999). Fuzzy Database Modeling. Studies in Fuzziness and Soft Computing, Vol. 26. Physica-Verlag, Heidelberg. 28. L. A. Zadeh (1965). Fuzzy Sets, Information and Control, 8, p. 338–353. 29. L. A. Zadeh (1975). Calculus of Fuzzy Restrictions. In: Fuzzy Sets and Their Applications to Cognitive and Decision Processes L. A. Zadeh, K.-S. Fu, K. Tanaka, M. Shimura (eds.). Academic Press Inc., New York, p. 1–40. 30. L. A. Zadeh (1975). The Concept of a Linguistic Variable and its Application to Approximate Reasoning I, II, III. Information Sciences, 8, p. 199–249, p. 301–357, and 9, p. 43–80.
Fuzzy Relational Images in Computer Science
151
31. L.A. Zadeh, J. Kacprzyk (eds.) (1999). Computing with Words in Information/Intelligent Systems 1. Studies in Fuzziness and Soft Computing, Vol. 33. Physica-Verlag, Heidelberg. 32. R. Zenner, R. De Caluwe, E. E. Kerre (1984). Practical determination of document description relations in document retrieval systems. In: Proceedings of the Workshop on the Membership Function, EIASM, Brussels, p. 127–138. 33. H.-J. Zimmerman (1996). Fuzzy Set Theory and its Applications. Kluwer Academic Publishers, Boston.
A Completeness Theorem for Extended Order Dependencies on Relational Attribute Models in Dedekind Categories Hitomi Okuma and Yasuo Kawahara Department of Informatics, Kyushu University 33, Fukuoka 812-8581, Japan {okuma, kawahara}@i.kyushu-u.ac.jp
Abstract. Order dependencies in relational database due to Ginsburg and Hull are relationships between attributes with domains of ordered values. The basic notions in this paper are comparison systems and relational attribute models in Dedekind categories. A comparison system constitutes a formal structure of possible orders for attribute domains. A relational attribute model is a system of relations on an object (of tuples or records) in Dedekind categories, which can be constructed by a suitable relational interpretation of comparing symbols. Generalizing order dependencies as well as functional dependencies, the paper introduces extended order dependencies, and their satisfactory relations for relational attribute models in Dedekind category. Then we give a simple proof that a revised set of inference rules is sound and complete.
1
Introduction
Relational database theory has been studied since Codd introduced relational database models. Dependency theory relates to relationships (constraints) between attributes in database relations. One of important questions with dependencies is that of (logical) implication: What other dependencies are necessarily satisfied by an instance when we know that a set of dependencies is satisfied by this instance? Inference rules are used in order to show implication between dependencies. The appropriateness of a set of inference rules is expressed by “soundness” and “completeness”. Armstrong and Beeri et al. [2] have proved the completeness theorems for functional and multivalued dependencies. Dependency theory has been studied with relational methods. Orlowska [10] proposed a relational formulation of functional, multivalued and other dependencies, and Buszkowski and Orlowska [3] developed an axiomatic relational calculus for dependency theory. Schmidt and Str¨ ohlein [12] explained a basic relational feature of functional dependency for relational models of databases. Jaoua et al. [6,11] studied difunctional dependencies in relational databases and gave some inference rules. MacCaull [7,8] investigated a relational formulation for functional and multivalued dependencies and association rules, and proved soundness and completeness for the implication problem of these H. de Swart (Ed.): RelMiCS 2001, LNCS 2561, pp. 152–170, 2002. c Springer-Verlag Berlin Heidelberg 2002
A Completeness Theorem for Extended Order Dependencies
153
dependencies with a Rasiowa/Sikorski-style tableaux method. The foundations and recent applications of relational methods in computer science are excellently summarized in [1]. In [5], Ginsburg and Hull introduced order dependencies to incorporate semantic information involving order on attribute domains. For example, we consider the information concerning the checks written by a specific individual (see Example 1). We assume that the checks have been written in sequence in ascending order, an order dependency can describe the statement that a check having larger check number would be written later than the smaller one. They defined a notion of marked attributes to describe the order between attribute values, and presented a sound and complete set of inference rules for order dependencies. They also discussed the complexity of determining logical implication for order dependency. The aim of this paper is to show the soundness and the completeness theorems of a set of inference rules for extended order dependencies, based on theory of Dedekind categories. To discuss order dependencies in our general framework, we define comparison systems, which generalize marked attributes in sense of [5], and comparators, which correspond to sets of marked attributes. Then we introduce a notion of relational attribute models, which is an extension of those in [9] or indiscernibility relations in [10]. A satisfactory relation for extended order dependencies on relational attribute models in Dedekind categories is defined with the same formalism as that for functional dependencies in [9]. The extended order dependencies succeed in extending order dependencies in [5]. Almost all arguments in this paper are point-free except for comparison systems and comparators. Finally we show the soundness and the completeness of a set of inference rules. The remainder of this paper is organized as follows: In Section 2, we review the definition of Dedekind categories, a kind of relational category, and some basic properties of Heyting algebra. In Section 3, first we recall the original definition of marked attribute and order dependency by Ginsburg and Hull [5]. Then the foundation of comparison systems to describe the relationships between attribute values is given. We present examples of comparison systems for functional dependencies and for order dependencies. In Section 4, we introduce a notion of relational attribute models in order to discuss extended order dependencies in Dedekind categories. Then we present basic properties of relational attribute models. In Section 5, we define extended order dependency and its satisfactory relation in Dedekind categories, and show some properties related to the propriety and the logical implication for extended order dependencies by using properties of Heyting algebra. In Section 6, a modified set of inference rules for extended order dependencies is presented and shown to be sound. In Section 7, properties of univalent comparators are shown, and finally we give a completeness theorem for the set of inference rules.
154
2
H. Okuma and Y. Kawahara
Dedekind Categories
In this section we recall the definition of a kind of a relation category which we will call Dedekind categories following Olivier and Serrato (1980). Dedekind categories are equivalent to locally complete division allegories, in particular they are equivalent to locally complete distributive allegories introduced in [4]. Throughout this paper, a morphism α from an object X into an object Y in a Dedekind category (which will be defined below) will be denoted by a half arrow α : X Y , and the composite of a morphism α : X Y followed by a morphism β : Y Z will be written as αβ : X Z. Also we will denote the identity morphism on X as idX . Definition 1. A Dedekind category D is a category satisfying the following: D1. [Complete Heyting Algebra] For all pairs of objects X and Y the hom-set D(X, Y ) consisting of all morphisms of X into Y is a complete Heyting algebra (namely, a complete distributive lattice) with the least morphism 0XY and the greatest morphism ∇XY . Its algebraic structure will be denoted by D(X, Y ) = (D(X, Y ), , , , 0XY , ∇XY ). That is, (a) is a partial order on D(X, Y ), (b) ∀α ∈ D(X, Y ) :: 0XY α ∇XY , (c) λ∈Λ αλ α iff αλ α for all λ ∈ Λ, (d) α λ∈Λ αλ iff α αλ for all λ ∈ Λ, and (e) α (λ∈Λ αλ ) = λ∈Λ (α αλ ). D2. [Converse] There is given a converse operation : D(X, Y ) → D(Y, X). That is, for all morphisms α, α : X Y , β : Y Z, the following laws hold: (a) (αβ) = β α , (b) (α ) = α, and (c) If α α , then α α . D3. [Dedekind Formula] For all morphisms α : X Y , β : Y Z and γ : X Z the Dedekind formula αβ γ α(β α γ) holds. D4. [Residue] For all morphisms β : Y Z and γ : X Z the residue (or division or weakest precondition) γ ÷ β : X Y is a morphism such that αβ γ if and only if α γ ÷ β for all morphisms α : X Y . A morphism f : X Y such that f f idY (univalent) and idX f f (total) is called a function and may be introduced as f : X → Y . In what follows the word relation is a synonym for morphism of a Dedekind category. Each hom-set D(X, Y ) has pseudo-complement, that is, for any two relations α and β in D there is a relation α ⇒ β in D(X, Y ) such that α γ β iff γ α ⇒ β for all relations γ. The next proposition lists some basic properties of relations in Dedekind categories related to Heyting algebra. Proposition 1. Let α, α , β, β , γ : X Y be relations in a Dedekind category D. Then the following hold:
A Completeness Theorem for Extended Order Dependencies (a) (b) (c) (d) (e) (f) (g) (h)
β α ⇒ β. α (α ⇒ β) β. If α α and β β , then α ⇒ β α ⇒ β . α ⇒ β (γ α) ⇒ (γ β). (α ⇒ β) = α ⇒ β . (α ⇒ β) (β ⇒ γ) α ⇒ γ. If α β, then α (β ⇒ γ) = α γ. (α ⇒ γ) (β ⇒ γ) (α β) ⇒ γ.
155
An object I of a Dedekind category D is called a unit if 0II = idI = ∇II . In this paper we assume that a Dedekind category D has a unit I. An I-point x of X is a function x : I → X.
3
Comparison Systems
Ginsburg and Hull [5] introduced a dependency which incorporate information involving order, called order dependency. They defined a notion of marked attributes to describe a certain relationship between attributes with ordered domains of values. In this section, as a foundation of a relational treatment for order dependencies in Dedekind categories, we first introduce comparison systems, a generalization of the marked attribute [5]. Then we define a notion of comparators which correspond to sets of marked attributes. At the end of the section we list two examples of comparison systems for functional dependencies and for extended order dependencies. Before presenting the notions of comparison systems and comparators, we recall the original definition of marked attribute and order dependency in [5]. Definition 2. Let U be a finite set of attributes and Ta a domain of values for each attribute a ∈ U . A tuple is a function u from U such that u(a) ∈ Ta for each attribute a ∈ U . T is the set of all tuples over U and an instance R over U is a subset of T . For each attribute a ∈ U , marked attributes are formal symbols a= , a≤ , a≥ , < a , a> , ain , used in the following way: Let u, v be tuples in T . Write: (a) u[a= ]v if u(a) =a v(a); (b) u[a≤ ]v if u(a) ≤a v(a); (c) u[a≥ ]v if u(a) ≥a v(a); (d) u[a< ]v if u(a) ]v if u(a) >a v(a); (f) u[ain ]v if u(a) and v(a) are incomparable under ≤a , where =a is an equality on Ta and ≤a is a partial ordering on Ta . An order dependency is an expression of the form X → Y , where X and Y are sets of marked attributes. We say that an instance R satisfies order dependency X → Y if for all tuples u and v in R, u[a∗ ]v for each marked attribute a∗ ∈ X implies u[b∗ ]v for each marked attribute b∗ ∈ Y . We next present an example of order dependencies in Checking-account database [5].
156
H. Okuma and Y. Kawahara
Example 1. Consider the information concerning the checks written by a specific individual. Let U be a set of four attributes N , D, P and A. The domain of N is the set of natural numbers representing the number of a check. The domain of D is the set of dates representing the date the check was written. The domain of P is the set of names of individuals and corporations, representing to whom the check was paid. The domain of A is the set of dollar amounts representing the amount for which the check was paid. We assume that the checks have been written in sequence in ascending order. Then the order dependencies N ≤ → D≤ and D< → N < hold. As a foundation of the relational treatment for order dependency in Dedekind categories, we now define a notion of comparison systems, which generalize the marked attributes. Definition 3. A comparison system L is a quartet (U, C, t , q) of a finite set U of attributes, a finite set C of comparing symbols, a transposal mapping t : C → C and an attribute mapping q : C → U satisfying the following: (a) stt = s and q(st ) = q(s) for each s ∈ C, (b) For each a ∈ U there is a distinguished element ea ∈ C such that eta = ea and q(ea ) = a. Set Ca = {s ∈ C | q(s) = a} (the inverse image of a by q) for each a ∈ U . Note that q is surjective and each Ca is a nonempty set, because q(ea ) = a and so ea ∈ Ca . Definition 4. Let L = (U, C, t , q) be a comparison system. A comparator over L is a binary relation X : U C (namely, a subset X ⊆ U × C) such that for each a ∈ U and s ∈ C, if (a, s) ∈ X, then q(s) = a. Note that a binary relation X : U C is a comparator iff Xq ⊆ idU iff X ⊆ q in a relational notation. For a comparator X over L we denote by X(a) a subset {s ∈ C | (a, s) ∈ X} of Ca . Clearly X uniquely corresponds to a set of subsets {X(a) | a ∈ U }. A comparator X over L is univalent if X(a) has at most one element for every a ∈ U , and is total if X(a) is nonempty for every a ∈ U . It is easy to see that the union X ∪ Y and the intersection X ∩ Y of comparators X and Y over L are also comparators over L. The transposal mapping t : C → C can be extended for relations from U into C as follows: ∀(a, s) : (a, s) ∈ X t ⇔ (a, st ) ∈ X. It is trivial that if X is a comparator over L then so is X t . A comparator U · over L is defined by U · (a) = {ea } for all attributes a ∈ U . For an attribute a ∈ U and a subset x ⊆ Ca we defined a comparator ax over L by ax (b) = x if b = a, and ax (b) = Cb otherwise. The comparator ax is a generalization of marked attributes in sense of [5]. Note that the comparator U · over L is always total and univalent by the definition.
A Completeness Theorem for Extended Order Dependencies
157
To maintain the consistency with the traditional notation in database theory we turn the natural inclusion of comparators upside down. That is, we write X ⊆ Y , X ∩ Y and X ∪ Y as X ⊇· Y , X ∪· Y and X ∩· Y , respectively. We now straightforwardly have the following proposition by the definition of comparators. Proposition 2. Let L = (U, C, t , q) be a comparison system, a ∈ U , x, y subsets of Ca and X a comparator over L. Then the following hold: (a) ax ∪· ay = ax∩y and ax ∩· ay = ax∪y . (b) X = ∪·a∈U aX(a) . Proof. (a) It is easy to see that (ax ∪· ay )(b) = x∩y if b = a, and (ax ∪· ay )(b) = Cb otherwise, and (ax ∩· ay )(b) = x ∪ y if b = a, and (ax ∩· ay )(b) = Cb otherwise. (b) For all b ∈ U we have (∪·a∈U aX(a) )(b) = ∩a∈U aX(a) (b) = X(b) ∩ {∩b =a aX(a) (b)} = X(b) ∩ Cb = X(b). The following proposition gives a basic property of containment between comparators, which is useful result to show the completeness theorem for order dependencies. Proposition 3. Let L = (U, C, t , q) be a comparison system and X and M comparators over L. Then the following hold: (a) If X is univalent and M ⊇· X, then either M = X or M is not total. (b) If X is univalent and X ⊇· M , then X ∪· M is not total. Proof. (a) Assume M is total. Then both X and M are functions, since M ⊇· X. Therefore M = X. (b) Assume X∪· M is total. Then X and X∪· M are functions, since X ⊇ X∩M = X ∪· M . Hence X = X ∪· M . Therefore X ⊇· M , which is a contradiction. We now show a trivial example of comparison systems for functional dependencies. Example 2. When we consider only functional dependencies in a database, the comparison system is simple, that is, a set C of comparing symbols is just a cartesian product U × {=, =} (where U is a given finite set of attributes), a mapping t : C → C is the identity mapping on C and a mapping q : C → U is the first projection from C onto U (see Fig. 1). Set ea = a × {=} for each a ∈ U . For a subset S of U we define a comparator S · : U C by S · = ∪·a∈S aea . Then it is easily seen that S · ⊆· U · (which is equivalent to U · ⊆ S · ) for all S ⊆ U , and that S ⊆ T ⇔ S · ⊆· T · . The traditional functional dependency pays attention only to comparators X such that X = S · for some subset S ⊆ U . This is a reason why we use the symbols ⊆· , ∪· and ∩· .
158
H. Okuma and Y. Kawahara Cb
Ca C
=a g
t
=b g
t
=ad
t
=bd
t
q
? a
U
···
b
···
Fig. 1. A comparison system for functional dependency
Next we introduce an example of comparison systems for order dependencies. Example 3. Consider the checking-account database mentioned in Example 1. Let C be a set of fifteen comparing symbols N , =N , D , =D , =P , =P , A2 , A1 , A0 , =A (see Fig. 2). The orders associated with each of the attribute domains for attributes N , D and P are obvious, and for example, for A it is the restriction of a divided-order for the real numbers: r 0 which, by Lemma 2(viii’), means that [R ]⊕ [R ]⊕ X(u0 ) [R ]⊕ X(u0 ). Hence [ R ]⊕ [R ]⊕ X ⊆L [R ]⊕ X. The second equivalence can be shown in the similar way. Acknowledgements. The authors are grateful to Petr Hajek, Peter Jipsen and the anonymous referee for helpful suggestions. This paper was partially supported by the KBN Grant No 8T11C01617. The work was carried out in the framework of the COST Action 274 on Theory and Applications of Relational Structures as Knowledge Instruments.
References 1. G. Allwein and M. Dunn (1993). “Kripke models for linear logic”. Journal of Symbolic Logic 58, No 2, pp. 514-545.
188
E. Orlowska and A.M. Radzikowska
2. Ph. Balbiani and E. Orlowska (1999). “A hierarchy of modal logics with relative accessibility relations”. Journal of Applied Non-Classical Logics 9, No 2-3, pp. 303–328, special issue in the memory of George Gargov. 3. K. Blount and C. Tsinakis (2001). “The structure of residuated lattices”. Preprint. 4. L. Farinas del Cerro and H. Prade (1986) “Rough sets, fuzzy sets and modal logic”. Fuzziness in indiscernibility and partial information. In: A. Di Nola and A. ¨ Rheinland. G. Ventre (eds), The Mathematics of Fuzzy Systems, Verlag TUV 5. C. C. Chang (1958) “Algebraic analysis of many-valued logics”. Transactions of the American Mathematical Society 88, pp. 467–490. 6. S. Demri, E Orlowska and D. Vakarelov (1999). “Indiscernibility and complementarity relations in information systems”. In: J. Gerbrandy, M. Marx, M. de Rijke and Y. Venema (eds) JFAK. Essays Dedicated to Johan van Benthem on the Occasion of his 50th Birthday. Amsterdam University Press. 7. S. Demri and E Orlowska (2002). Incomplete Information: Structure, Inference, Complexity. EATCS Monographs in Theoretical Computer Science, Springer, forthcoming. 8. R. P. Dilworth and N. Ward (1939) “Residuated lattices”. Transactions of the American Mathematical Society 45, pp. 335–354. 9. I. D¨ untsch and E. Orlowska (2000). “Logics of complementarity in information systems”. Mathematical Logic Quarterly 46, pp. 267–288. 10. I. D¨ untsch and E. Orlowska (2000). “Beyond modalities: sufficiency and mixed algebras”. In: E. Orlowska and A. Szalas (eds) Relational Methods for Computer Science Applications, Physica Verlag, Heidelberg, pp. 263–285. 11. I. D¨ untsch and E. Orlowska (2001). “Algebraic structures for qualitative reasoning”. Alfred Tarski Centenary Conference, Warsaw, Poland. 12. F. Esteva and L. Godo (2001) “Monoidal t–norm based logic: towards a logic for left–continuous t–norms”. Fuzzy Sets and Systems 124, pp. 271–288. 13. F. Esteva and L. Godo (2001). “On complete residuated many-valued logics with tnorm conjunction”. Proceedings of the 31st International Symposium on MultipleValued Logic, Warsaw, Poland, pp. 81–86. 14. P. Flondor, G. Georgescu, and A. Iorgulecu (2001). “Psedo-t-norms and pseudo-BL algebras”. Soft Computing 5, No 5, pp. 355-371. 15. S. Gottwald (2001). A Treatise on Many-Valued Logics, Studies in Logic and Computation 9, Research Studies Press: Baldock, Hertfordshire, England. 16. J. A. Gougen (1967). “L–fuzzy sets”. Journal of Mathematical Analysis and Applications 18, pp. 145–174. 17. P. Hajek (1998). Metamathematics of Fuzzy Logic, Kluwer, Dordrecht. 18. J. B. Hart, L. Rafter, and C. Tsinakis (2001). “The structure of commutative residuated lattices”. Preprint. 19. C. A. R. Hoare and H. Jifeng (1986). “The weakest prespecification”. Part I: Fundamenta Informaticae IX, pp. 51–84. Part II: Fundamenta Informaticae IX, pp. 217–252. 20. U. H¨ ohle and U.P. Klement (eds) (1996). Non–Classical Logics and their Applications to Fuzzy Subsets, Kluwer, Dordrecht. 21. U. H¨ ohle (1996). “Commutative, residuated l-monoids”. In [20], pp. 53–106. 22. I. Humberstone (1983). “Inaccessible words”. Notre Dame Journal of Formal Logic 24, pp. 346–352. 23. P. Jipsen (2001). “A Gentzen system and decidability for residuated lattices”. Preprint. 24. B. Konikowska (1997). “A logic for reasoning about relative similarity”. Studia Logica 58, pp. 185–226.
Double Residuated Lattices and Their Applications
189
25. R. K. Meyer and R. Routley (1972). “Algebraic analysis of entailment”, Logique et analyse 15, pp. 407–428. 26. E. Orlowska (1988). “Kripke models with relative accessibility and their application to inferences from incomplete information”. In: Mirkowska, G. and Rasiowa, H. (eds), Mathematical Problems in Computation Theory. Banach Center Publications 21, pp. 329–339. 27. E. Orlowska (ed) (1998). Incomplete Information – Rough Set Analysis, Studies in Fuzziness and Soft Computing, Springer-Verlag. 28. E. Orlowska (1999). “Many–valuedness and uncertainty”. Multiple-Valued Logic 4, pp. 207–227. 29. E. Orlowska and A. M. Radzikowska (2001). “Information relations and operators based on double residuated lattices”. In H. de Swart (ed), Proceedings of the 6th Seminar on Relational Methods in Computer Science RelMiCS’2001, pp. 185–199. 30. A. M. Radzikowska and E. E. Kerre (2002) “A comparative study of fuzzy rough sets”. Fuzzy Sets and Systems 126 No 2, pp. 137–155. 31. A. M. Radzikowska, E. E. Kerre (2002) “A general calculus of fuzzy rough sets”. Submitted. 32. A. M. Radzikowska and E. E. Kerre (2001). “On some classes of fuzzy information relations”. Proceedings of the 31st International Symposium on Multiple-Valued Logic, Warsaw, Poland, pp. 75–80. 33. A. M. Radzikowska and E. E. Kerre (2001) “Towards studying of fuzzy information relations”. To appear in Proceedings of EUSFLAT-2001, Leicester, UK. 34. H. Rasiowa and R. Sikorski (1970). The Mathematics of Metamathematics, Warszawa. 35. C. Rauszer (1974). “Semi-Boolean algebras and their applications to intuitionistic logic with dual operations”. Fundamenta Mathematicae 83, pp. 219-249. 36. B. Schweizer and A. Sklar (1983). Probabilistic Metric Spaces. North Holland, Amsterdam. 37. E. Turunen (1999) Mathematics Behind Fuzzy Logic, Springer–Verlag. 38. D. Vakarelov (1998). “Information systems, similarity relations and modal logics”. In [27], pp. 492–550.
Interval Bilattices and Some Other Simple Bilattices Agata Pilitowska Warsaw University of Technology, Department of Mathematics Plac Politechniki 1, 00-661 Warsaw, Poland
[email protected] Abstract. In a number of papers M.Ginsberg introduced algebras called bilattices having two separate lattice structure and one additional basic unary operation. They originated as an algebraization of some nonclassical logics that arise in artificial intelligence and knowledge-based logic programming. In this paper we introduce some new class of bilattices which originate from interval lattices and show that each of them is simple. A known simple lattices are used to give other examples of simple bilattices. We also describe simple bilattices satisfying some additional identities so called P-bilattices (or interlaced bilattices).
1
Introduction
In a number of papers M.Ginsberg [6], [7] introduced algebras called bilattices having two separate lattice structure and one additional basic unary operation acting on both lattices in a very regular way. Bilattices originated as an algebraization of some non-classical logics that arise in artificial intelligence and knowledge-based logic programming. The two lattice orderings of a bilattice may be viewed as the degrees of truth and knowledge of possible events. In general, both lattice structures of a bilattice are ”almost” not connected to each other. However bilattices appearing in applications usually satisfy some additional conditions. Bilattices were also investigated by Fitting, Romanowska, Avron, Mobasher, Pigozzi, Slutzki, Voutsadakis, Pynko and others. In [5] M.Fitting described a structure based on intervals of a lattice with one lattice and one semilattice orderings. The main purpose of this paper is to introduce a new class of bilattices which also originates from interval lattices and to show that such bilattices are simple. In Sect. 1 we collect some useful facts about interval lattices, and bilattices. Interval bilattices, a new class of bilattices which originate from interval lattices, are described in Sect. 2. In Sect. 3 we show that all interval bilattices are simple. In Sect. 4 a known simple lattices are used to give other examples of simple bilattices. Finally, we describe simple P -bilattices (bilattices satisfying some additional identities which were introduced and investigated in [13]. P-bilattices were also investigated under the name of ”interlaced” bilattices by other authors.).
The paper was written within the framework of COST Action 274.
H. de Swart (Ed.): RelMiCS 2001, LNCS 2561, pp. 190–196, 2002. c Springer-Verlag Berlin Heidelberg 2002
Interval Bilattices and Some Other Simple Bilattices
2
191
Preliminaries
Let L = (L, ∨, ∧) be a lattice with the ordering relation ≤. The sublattice [a, b] = {x ∈ L|a ≤ x ≤ b} of L is called an interval with the end-points a and b. The empty subset of L is also considered as an interval [ , ]. The family of all intervals in L is denoted by IntL. IntL is a lattice IntLk = (IntL, ◦, +) under set inclusion ≤k . We have that [a, b] ≤k [c, d] if and only if a ≥ c and b ≤ d. The lattice operations are defined by the following formulas: [a ∨ c, b ∧ d] if a ∨ c ≤ b ∧ d [a, b] ◦ [c, d] := (1) [, ] otherwise [a, b] + [c, d] := [a ∧ c, b ∨ d] .
(2)
The lattice IntLk will be called an interval lattice. Note that the empty interval [ , ] is the least element in this lattice. The interval lattices were studied by W.Duthie [2] and by V.Igoshin [8]. Definition 1. A bilattice B = (B, ∧, ∨, ◦, +, ¬, 01 , 11 , 02 , 12 ) is an algebra such that 1. (B, ∧, ∨, 01 , 11 ) and (B, ◦, +, 02 , 12 ) are bounded lattices. 2. A negation ¬ : B → B is an unary operation on B satisfying for all x, y ∈ B the identities: ¬¬x = x, (3) ¬(x ∨ y) = ¬x ∧ ¬y
(4)
¬(x ∧ y) = ¬x ∨ ¬y
(5)
¬(x + y) = ¬x + ¬y
(6)
¬(x ◦ y) = ¬x ◦ ¬y .
(7)
Definition 2. A bilattice B = (B, ∧, ∨, ◦, +, ¬, 01 , 11 , 02 , 12 ) satisfying the following identities: ((x ∧ y) ◦ z) ∧ (y ◦ z) = (x ∧ y) ◦ z (8) ((x ◦ y) ∧ z) ◦ (y ∧ z) = (x ◦ y) ∧ z
(9)
((x ∧ y) + z) ∧ (y + z) = (x ∧ y) + z
(10)
((x + y) ∧ z) + (y ∧ z) = (x + y) ∧ z
(11)
will be called P-bilattice. The following results were proved in [13]. (See also [1], [4] and [12].)
192
A. Pilitowska
Theorem 1. Let L = (L, ∧, ∨, 0, 1) be a bounded lattice. On a set L × L we define an algebra B(L) = (L × L, ∧, ∨, ◦, +, ¬, 01 , 11 , 02 , 12 ) as follows: for all (x, x ), (y, y ) in L × L, (x, x ) ∧ (y, y ) := (x ∧ y, x ∨ y )
(12)
(x, x ) ∨ (y, y ) := (x ∨ y, x ∧ y )
(13)
(x, x ) ◦ (y, y ) := (x ∧ y, x ∧ y )
(14)
(x, x ) + (y, y ) := (x ∨ y, x ∨ y )
(15)
¬(x, y) := (y, x)
(16)
01 := (0, 1), 11 := (1, 0), 02 := (0, 0), 12 := (1, 1) .
(17)
The algebra B(L) is a P-bilattice. The bilattice B(L) will be called a product bilattice associated with the lattice L. Theorem 2. An algebra B = (B, ∧, ∨, ◦, +, ¬, 01 , 11 , 02 , 12 ) is a P-bilattice if and only if there is a bounded lattice L = (L, ∧, ∨, 0, 1), such that B is isomorphic to the product bilattice B(L) associated with the lattice L.
3
A Construction of Interval Bilattices
Let L = (L, ∧, ∨, ¬, 0, 1) be a bounded lattice with a polarity ¬ : L → L and ordering relation ≤. Let Int0 L be the family of all intervals of L excluding the empty interval. In Int0 L define two binary operations as follows: [a, b] ∧ [c, d] := [a ∧ c, b ∧ d]
(18)
[a, b] ∨ [c, d] := [a ∨ c, b ∨ d] .
(19)
Proposition 1. The algebra Int0 Lt = (Int0 L, ∧, ∨, [0, 0], [1, 1]) is a bounded lattice. Proof. It is evident that if a ≤ b and c ≤ d then a ∧ c ≤ b ∧ d and a ∨ c ≤ d ∨ d. Therefore [a, b] ∨ [a, b] = [a ∨ a, b∨, b] = [a, b], [a, b] ∨ [c, d] = [a ∨ c, b ∨ d] = [c ∨ a, d ∨ b] = [c, d] ∨ [a, b], ([a, b] ∨ [c, d]) ∨ [e, f ] = [a ∨ c, b ∨ d] ∨ [e, f ] = [a ∨ c ∨ e, b ∨ d ∨ f ] = = [a, b] ∨ [c ∨ e, d ∨ f ] = [a, b] ∨ ([c, d] ∨ [e, f ]), [a, b] ∨ ([a, b] ∧ [c, d]) = [a ∨ b] ∨ [a ∧ c, b ∧ d] = [a ∨ (a ∧ c), b ∨ (b ∧ d)] = [a, b]. Simillarly, we can show that the second operation ∧ is idempotent, commutative and associative. Moreover, for all [a, b] in Int0 L: [a, b] ∨ [1, 1] = [a ∨ 1, b ∨ 1] = [1, 1], [a, b] ∧ [0, 0] = [a ∧ 0, b ∧ 0] = [0, 0]. It follows that Int0 Lt is a bounded lattice with the least element [0, 0] and the greatest element [1, 1].
Interval Bilattices and Some Other Simple Bilattices
193
Corollary 1. The ordering relation ≤t of a lattice Int0 Lt is defined by [a, b] ≤t [c, d] if and only if a ≤ c and b ≤ d .
(20)
Now let IntLt be the extension of the partially ordered set (Int0 L, ≤t ) by the empty interval [ , ]. By definition [ , ] covers [0, 0], is covered by [1, 1] and is not comparable with any other interval of L. In IntL we define one unary operation ¬ : IntL → IntL by ¬[a, b] := [¬b, ¬a]
(21)
¬[ , ] := [ , ] .
(22)
Note that if a ≤ b then ¬b ≤ ¬a, whence [¬b, ¬a] is an interval in L. Proposition 2. The algebra BIntL = (IntL, ∧, ∨, ◦, +, ¬, [0, 0], [1, 1], [ , ], [0, 1]) is a bilattice. Proof. Since the lattice L = (L, ∧, ∨, ¬, 0, 1) is bounded, the lattice IntLk has the greatest element [0, 1]. Hence IntLt = (IntL, ∧, ∨, [0, 0], [1, 1]) and IntLk = (IntL, ◦, +, [ , ], [0, 1]) are bounded lattices. Note that the unary operation ¬ : IntL → IntL is an involution ¬¬[a, b] = ¬[¬b, ¬a] = [a, b] .
(23)
It reverses the order ≤t and preserves the order ≤k . Indeed, if [a, b] ≤t [c, d] then a ≤ c and b ≤ d implying ¬a ≥ ¬c and ¬b ≥ ¬d. Hence [¬b, ¬a] ≥t [¬d, ¬c] and ¬[a, b] ≥t ¬[c, d]. If [a, b] ≤k [c, d] then [a, b] + [c, d] = [a ∧ c, b ∨ d] = [c, d] implying ¬[a ∧ c, b ∨ d] = [¬(b ∨ d), ¬(a ∧ c)] = [¬b ∧ ¬d, ¬a ∨ ¬c] = = [¬b, ¬a] + [¬d, ¬c] = ¬[a, b] + ¬[c, d] = ¬[c, d]. Hence, ¬[a, b] ≤k ¬[c, d]. It follows that BIntL is a bilattice. Definition 3. The bilattice BIntL = (IntL, ∧, ∨, ◦, +, ¬, [0, 0], [1, 1], [ , ], [0, 1]) will be called an interval bilattice. Example 1. Let ({0, 1}, ∧, ∨, ¬, 0, 1) be two element chain. In this case IntL = {[ , ], [0, 0], [1, 1], [0, 1]} and the interval bilattice BIntL is isomorphic to four element distributive bilattice B4 . Lemma 1. Let L = (L, ∧, ∨, ¬, 0, 1) be an arbitrary bounded polarity lattice with more than two elements. Then the interval bilattice BIntL = (IntL, ∧, ∨, ◦, +, ¬, [0, 0], [1, 1], [ , ], [0, 1]) is not a P-bilattice. Proof. Let a = 0 and a = 1 be an element in L. We have (([ , ] ∧ [0, a]) ◦ [a, 1]) ∧ ([0, a] ◦ [a, 1]) = ([0, 0] ◦ [a, 1]) ∧ [0 ∨ a, a ∧ 1] = = [0 ∨ a, 0 ∧ 1] ∧ [a, a] = [ , ] ∧ [a, a] = [0, 0] and ([ , ] ∧ [0, a]) ◦ [a, 1] = [0, 0] ◦ [a, 1] = [ , ]. It follows that identity ((x ∧ y) ◦ z) ∧ (y ◦ z) = (x ∧ y) ◦ z
(24)
is not satisfied in the interval bilattice BIntL. Hence BIntL is not a P-bilattice.
194
4
A. Pilitowska
Congruence Lattices of Interval Bilattices
In this section we give a characterization of the lattice of congruence relations of an interval bilattice. Let BIntL be an interval bilattice obtained from some bounded polarity lattice L= (L, ∧, ∨, ¬, 0, 1). The lattice of congruence relations of BIntL will be denoted by ConBIntL. Theorem 3. For any bounded polarity lattice L= (L, ∧, ∨, ¬, 0, 1), the congruence lattice ConBIntL is isomorphic to the 2-element chain. Proof. Let a, b, c, d be in L, d = 0, d = 1 and θ be a non-zero congruence relation of BIntL. First note that if [a, b] ≤k [c, d] then c ≤ a ≤ b ≤ d and [a, b] ◦ [d, d] = [a ∨ d, b ∧ d] = [ , ] and [c, d] ◦ [d, d] = [c ∨ d, d ∧ d] = [d, d]. Therefore it is easy to see that if [a, b] ≤k [c, d] and [a, b] and [c, d] are in θ then [d, d] and [ , ] are in θ too. It is well known that in a lattice (L, ∧, ∨), elements x and y are in some congruence relation ϕ if and only if x ∧ y and x ∨ y are in ϕ. Since by definition intervals [d, d] and [ , ] are not comparable with respect to ≤t , it is clear that if [d, d] and [ , ] lie in θ, then θ also contains [0, 0] and [1, 1], and hence all other intervals. Moreover, if [1,1] and [ , ] (or [0,0] and [ , ]) are in θ, then ¬[1, 1] = [0, 0] and ¬[ , ] = [ , ] (or ¬[0, 0] = [1, 1] and [ , ]) are in θ too. It follows that θ is the largest trivial congruence relation. Corollary 2. For any bounded polarity lattice L, the interval bilattice BIntL is simple.
5
Further Examples of Simple Bilattices
In this section we give further examples of simple bilattices. Example 2. Let n, i and j be natural numbers and let Qn := {(0, 0), (1, 1)} ∪ {(2−i , 0)|0 ≤ i ≤ n − 1} ∪ {(0, 2−i )|0 ≤ i ≤ n − 1}∪ ∪{(2−i , 2−j )|1 ≤ i, j ≤ n and |i − j| ≤ 1}. n n Let Qn k = (Q , ≤k ) be the set Q with the order ≤k defined by (a1 , b1 ) ≤k (a2 , b2 ) iff a1 ≤ a2 and b1 ≤ b2 .
(25)
n n And let Qn t = (Q , ≤t ) be the set Q with the order ≤t defined by
(a1 , b1 ) ≤t (a2 , b2 ) iff a1 ≤ a2 and b1 ≥ b2 .
(26)
n Obviously, Qn k and Qt are lattices. n The lattices Qk and Qn t with the unary operation ¬(a, b) := (b, a) form the bilattice Qn . First note that the unary operation ¬ : Qn → Qn is an involution:
Interval Bilattices and Some Other Simple Bilattices
195
¬¬(a, b) = (a, b). Moreover, it reverses the order ≤t and preserves the order ≤k , because (a1 , b1 ) ≤k (a2 , b2 ) ⇔ a1 ≤ a2 and b1 ≤ b2 ⇔ ⇔ ¬(a1 , b1 ) = (b1 , a1 ) ≤k (b2 , a2 ) = ¬(a2 , b2 ) and (a1 , b1 ) ≤t (a2 , b2 ) ⇔ a1 ≤ a2 and b1 ≥ b2 ⇔ ⇔ ¬(a1 , b1 ) = (b1 , a1 ) ≥t (b2 , a2 ) = ¬(a2 , b2 ). By R.Wille [14] the lattice Qn k is simple. Because if a lattice reduct of a bilattice is simple then the bilattice in question is always simple, too, the bilattice Qn is simple. n n Example 3. Let n ≥ 2 be a natural number and Q2n 0 := Q ∪P −{(0, 1), (1, 0)}, n n where Q is the set defined in Ex. 2 and P := {(1 − x, 1 − y)|(x, y) ∈ Qn }. In Q2n 0 define partial order ≤t as follows:
(a, b) ≤t (c, d) iff a ≤ c and b ≤ d .
(27)
2n Now let Q2n t be the extension of the partially ordered set (Q0 , ≤t ) by a single −n −n element (0,1). By definition (0,1) covers (2 , 2 ), is covered by (1 − 2−n , 1 − 2−n ) and is not comparable with any other element of the set Q2n 0 different from (1,1) and (0,0). It is clear that Q2n t is a bounded lattice with (1,1) as the greatest element and (0,0) as the least element. Moreover in Q2n := Q2n 0 ∪ {(0, 1)} define two binary operations as follows: (max(a, c), max(b, d)) if (a, b), (c, d) ∈ Qn − {(0, 1), (1, 1)} (a, b) ◦ (c, d) := (min(a, c), min(b, d)) if (a, b), (c, d) ∈ P n − {(0, 1), (0, 0)} ( 12 , 12 ) otherwise (min(a, c), min(b, d)) if (a, b), (c, d) ∈ Qn − {(0, 1), (1, 1)} (a, b) + (c, d) := (max(a, c), max(b, d)) if (a, b), (c, d) ∈ P n − {(0, 1), (0, 0)} . (0, 1) otherwise 1 1 2n It is easy to see that Q2n k = (Q , ◦, +, ( 2 , 2 ), (0, 1)) is a bounded lattice. 2n On the set Q we define the unary operation
¬(a, b) := (1 − b, 1 − a) .
(28)
2n It is easy to note that Q2n k and Qt with the above unary operation is a bilattice 2n Q . As was shown by R. Wille [14], the lattice Q2n t is simple. Hence the bilattice Q2n is simple too.
The following result was proved in [13]. Lemma 2. Let B = (B, ∧, ∨, ◦, +, ¬, 01 , 11 , 02 , 12 ) be a P-bilattice isomorphic to the product bilattice B(L) associated with the lattice L = (L, ∧, ∨). Then the lattice ConB is isomorphic to the lattice ConL. As an easy consequence of this lemma we obtain the following corollary. Corollary 3. A P-bilattice B(L) is simple if and only if the lattice L is simple.
196
A. Pilitowska
References 1. Avron, A.: The structure of interlaced bilattices. Math. Structures Comput. Sci. 6 (1996) 287-299 2. Duthie, W.: Segments of order sets. Trans. Amer. Math. Soc. 51(1942) 1-14 3. Fitting, M.: Bilattices and the theory of truth. Journal of Philosophical Logic 18(1989) 225-256 4. Fitting, M.: Bilattices in logic programming. The Twentieth International Symposium on Multiple-Valued Logic (ed. Epstein, G.) IEEE (1990) 238-246 5. Fitting, M.: Kleene’s logic, generalized. J. Logic Computat. 1 (1991) 797-810 6. Ginsberg, M.: Multi-valued logics. Proc. AAAI-86, Fifth National Conference on Artificial Intelligence, Morgan Kaufmann Publishers (1986) 243-247 7. Ginsberg, M.: Multivalued logics: A uniform approach to inference in artificial intelligence. Computational Intelligence 4(1988) 265-316 8. Igoshin, V.: Algebraic characteristic of interval lattices/Russian/. Uspekhi Mat. Nauk 40(1985) 205-206 9. McKenzie, R., McNulty, G., Taylor, W.: Algebras, Lattices, Varieties. The Wadsworth and Brooks, Monterey (1987) 10. Mobasher, B., Pigozzi, D., Slutzki, G., Voutsadakis, G.: A Duality theory for bilattices. Algebra Universalis 43(2000) 109-125 11. Odintsov, V.: Congruences on lattices of intervals /Russian/. Mat. Zap. 14(1988) 102-111 12. Pynko, A.: Regular bilattices. Journal of Applied Non-Classical Logics 10(2000) 93-111 13. Romanowska, A., Trakul, A.(Pilitowska A.): On the structure of some bilattices. In: Hakowska, K., Stawski, B.(eds.) Universal and Applied Algebra, World Scientific (1989) 235-253 14. Wille, R.: A note on simple lattices. Col. Math. Soc. Janos Bolyai, 14. Lattice Theory, Szeged (1974) 455-462
Interactive Systems: From Folklore to Mathematics Gheorghe S ¸ tef˘ anescu Department of Computer Science, National University of Singapore
[email protected] Abstract. This paper gives a brief introduction to finite interactive systems, an abstract mathematical model of agents’ behavior and their interaction. The paper contains the definition of finite interactive systems, examples, and a few simple results. No examples modeling real interactive systems are included, but there are many pointers suggesting how this can be done.
1
Introduction
This paper gives a brief introduction to finite interactive systems (fis’s, for short), an abstract mathematical model of agents’ behavior and their interaction. The paper contains the definition of finite interactive systems, examples, a few simple results, and pointers suggesting how fis’s may be use to model real interactive systems. The agents we are talking about are those specified in standard concurrent object-oriented languages. The key point is the observation that grids may be seen as interaction running patterns for programs written in these languages much in the same way as paths represent possible executions of classical sequential programs. Then, finite interactive systems may be seen as a kind of two-dimensional finite automata melting together a state transforming automaton with a behavior transforming one (this latter transformation is responsible for modeling the interaction of the threads generated by the first automaton). The main advances of the present paper are: – it introduces finite interactive systems as a machine-like device for recognizing grid languages; – it presents a large collection of toy examples illustrating the power of the mechanism; and – it gives pointers suggesting how finite interactive systems may be use to model real interactive systems as those described by concurrent objectoriented languages. Parts of this paper are included in the survey paper [Ste02] published in the volume dedicated to 2001 Marktoberdorf Summer School.
On leave from Departement of Fundamentals of Computer Science, University of Bucharest
H. de Swart (Ed.): RelMiCS 2001, LNCS 2561, pp. 197–211, 2002. c Springer-Verlag Berlin Heidelberg 2002
198
2 2.1
G. S ¸ tef˘ anescu
Grids and Regular Expressions Grids
A grid (or planar word) is a rectangular two-dimensional area filled in with atomic letters in a given alphabet V . Their set is denoted by V † . Each letter in V is to be seen as a two-dimensional atom having its own northern, southern, western and eastern border. We do not consider any typing mechanism here, hence all atoms’ borders are similar, denoted by 1; this border information is naturally extended to general grids specifying the number of atoms lying on that border (rectangle’s dimensions). For a grid p, we let n(p), s(p), w(p) and e(p) denote the dimension of its northern, southern, western, and eastern border, respectively. In the literature, the term picture is often used as a substitute for grid or twodimensional word, especially in the context of picture languages. Useful surveys on two-dimensional languages may be found in [InT91], [GiR97], or [LMN98]. However, our view is more semantical, hence we prefer to use the term ‘grid’ considered as a two-dimensional version of ‘path’. Sometimes we will also use the term ‘planar word’ (as a two dimensional version of ‘word’), but never ‘picture’. 2.2
Simple Regular Expressions
We present here a simple extension of classical regular expressions [Kle56,Con71] to cope with grids. Simple two-dimensional regular expressions. The signature of simple twodimensional regular expressions consists of two collections of usual regular operators, sharing the additive part. Specifically, it is +, 0, ·, , |, , † , − where (+, 0, ·, , |) is a usual Kleene signature to be used for the vertical dimension, while (+, 0, , † , −) is another Kleene signature to be used for the horizontal dimension. Notice that ‘0’, ‘|’ and ‘−’ are zero-ary operators (constants), ‘ ’ and ‘† ’ are unary, while the remaining ones are binary. The following names are used for the above symbols: ‘+’ is sum, ‘0’ is zero, ‘·’ is vertical composition, ‘ ’ is vertical star (or iterated vertical composition), ‘|’ is vertical identity, ‘’ is horizontal composition, ‘† ’ is horizontal star (or iterated horizontal composition) and ‘−’ is horizontal identity. Finally, simple two-dimensional regular expressions over an alphabet V (consisting of two-dimensional letters/atoms) are obtained starting from atoms and iteratively applying the operations in the described signature. Formally, E ::= a | 0 | E + E | E · E | E | | | E E | E † | − Their set is denoted by 2RegExp(V ).
Interactive Systems: From Folklore to Mathematics
199
Axiomatization, one-dimensional case. There are many elegant axiomatizations for (one-dimensional) regular languages, including the one presented by Kozen in [Koz94]. A particular class of Kleene algebras is defined by the following set of identities (where we use matrices to simplify the form of the star axioms), see for instance Ch.8 of [Ste00]. Let M = (M(m, n)m,n , +, ·, ∗, 0m,n , In ) be a family of matrices with operations: (1) 0m,n ∈ M(m, n), (2) In ∈ M(n, n); (3) ‘+’ : M(m, n) × M(m, n) → M(m, n); (4) ‘·’ : M(m, n) × M(n, p) → M(m, p); (5) ‘∗’ : M(n, n) → M(n, n). M is a Kleene algebra if – M is a semiring of matrices, namely: (1) with +, 0 it is a commutative monoid (a + b) + c = a + (b + c); a + 0m,n = 0m,n + a = a; a + b = b + a; (2) with ·, Ip it is a category (a · b) · c = a · (b · c); a · In = a = Im · a; (3) distributivity: a · (b + c) = (a · b) + (a · c); (a + b) · c = (a · c) + (b · c); and (4) zero-laws: 0p,m · a = 0p,n ; a · 0n,p = 0m,p . – It is idempotent, i.e., a + a = a. – The following axioms for star holds: (I) (In )∗ = In (S) (a + b)∗ = (a∗ · b)∗ · a∗ (P) (a · b)∗ = In + a · (b · a)∗ · b (Inv) a · ρ = ρ · b ⇒ a∗ · ρ = ρ · b∗ , where ρ is matrix over 0,1. These axioms for Kleene algebra give a correct and complete axiomatization for regular languages. It is highly desirable to have a similar set of simple and powerful identities for regular grid languages, too. From expressions to grids. To each expression in 2RegExp(V ) one may associate a language of grids over V . To this end, we first describe how composition operations act on grids and give the meaning of the identity operators. If v, w are grids, then their horizontal composition is defined only if e(v) = w(w) and the result v w is the word obtained putting v on left of w. Their vertical composition is defined only if s(v) = n(w) and v · w is the word obtained putting v on top of w. For each natural number k one may associate two ‘empty’ grids: the vertical identity k having w(k ) = e(k ) = 0 and n(k ) = s(k ) = k and the horizontal identity λk with w(λk ) = e(λk ) = k and n(λk ) = s(λk ) = 0. Now, the interpretation | | : 2RegExp(V )
→
P(V † )
from expressions to languages of grids is defined by: – – – –
|a| = {a}; |0| = ∅; |E + F | = |E| ∪ |F |; |E · F | = {v · w : v ∈ |E| & w ∈ |F |}; |E | = {v1 · . . . · vk : k ∈ IN & v1 , . . . , vk ∈ |E|}; ||| = {0 , . . . , k , . . .};
200
G. S ¸ tef˘ anescu
– |E F | = {v w : v ∈ |E| & w ∈ |F |}; – |E † | = {v1 . . . vk : k ∈ IN & v1 , . . . , vk ∈ |E|}; – |−| = {λ0 , . . . , λk , . . .}. Before giving some examples, we define a useful flattening operator mapping languages of grids to languages of usual, one-dimensional words. The flattening operator. The flattening operator : P(V † )
→
P(V )
maps sets of grids to sets of strings representing their topological sorting. In more details it is defined as follows: – Each grid w may be considered as an acyclic directed graph g(w) drawing: (1) horizontal edges from each letter to its right neighbor, if this exists, and (2) vertical edges from each letter to its bottom neighbor, if this exists. – Being acyclic, g(w) and all of its subgraphs have minimal vertices, i.e., vertices without incoming arrows. – The topological sorting procedure selects a minimal vertex, then deletes it and its outgoing edges and repeats this as long as possible. This way a onedimensional word containing all the atoms of w is obtained. – Varying the minimal vertex selection in the topological sorting procedure in all possible ways one gets a set of words that is the value of the flattening operator applied to the grid. – Finally, to define on a language L, take the union of (w), for w ∈ L. To have an example, let us start with a grid abcd efgh ; there is only one minimal bcd ; now there are two minimal elements element a and after its deletion we get efgh cd ; next, b and e to choose from and suppose we choose b; what remains is efgh
from the minimal elements c and e, we choose e and get
cd fgh ; and so on; finally
a usual word, say abecfgdh, is obtained. Actually, ( abcd efgh ) = {abcdefgh, abcedfgh, abcefdgh, abcefgdh, abecdfgh, abecfdgh, abecfgdh, abefcdgh, abefcgdh, aebcdfgh, aebcfdgh, aebcfgdh, aebfcdgh, aebfcgdh}. It is one of our main claims that this flattening operator is responsible for the well-known state-explosion problem which occurs in the verification of concurrent object-oriented systems. And one of our main hopes is that the lifting of the verification techniques from paths to the grids is possible and may avoid this problem. A very nice quantitative analysis of the state-explosion generated by the flattening operator is presented in [BCGS01]: If w is a partial grid (i.e., a part of a grid which remains after some letters are already selected/deleted) and
Interactive Systems: From Folklore to Mathematics
201
k1 , . . . , kp are the numbers representing for each cell the sum of the cells on top of it and on left of it plus 1, then the number of the flattened words is p! k1 . . . kp Applied to full, square grids of size n, this show that the state-explosion gener2 ated by the flattening operator is of order O(nn ). Convention: we use terminal type letters a, b, ... in grids; they should be identified with the corresponding italic version, used elsewhere. Examples. A few simple examples of regular expressions are presented below. 1. The expression (a · d · g) (b · e · h)† (c · f · i) represents the language of grids ad be .. .. be cf . Notice that the same language may be represented by the . . d g
. . e h
. . . .
. . . .
. . e h
. . f i
expression (a b† c) · (d e† f ) · (g h† i). 2. The language associated with a† · b† consists of grids aa..a bb..b . Its flattened † † ∗ version (a · b ) = {w ∈ {a, b} : |w|a = |w|b & ∀w = w w : |w |a ≥ |w |b } is context-free, but not regular (|w|a denotes the number of the occurrences of a in w). The slightly extended expression a† · b† · c† represents the language of grids aa..a . Notice that its flattened version is (a† · b† · c† ) = {w ∈ {a, b, c}∗ : |w|a = bb..b cc..c
|w|b = |w|c & ∀ w = w w : |w |a ≥ |w |b ≥ |w |c }. This latter language is not even context-free. 3. Notice that (a+b)† = (a+b)† . This shows that vertical and horizontal stars may commute, provided simple atoms are involved. (This explain our notation V † for the set of grids over V : if V = {a1 , . . . , an }, then V † = (a1 +. . .+an )† .) † † On the other hand, aa bb ∈ (a + b b) \ (a + b b) , showing that, in general, the stars do not commute. 4. Finally, notice that a† · b† = a b , but (a† · b† ) = (a b ). This example shows that the flattening mechanism may lose some information by passing from planar to usual languages. 2.3
Extended Regular Expressions
The regular expressions previously described are simple and natural, but they have a number of shortcomings. For instance, their generative power is quite
202
G. S ¸ tef˘ anescu
limited, comparing with finite interactive systems. In this subsection we roughly describe a few possible extensions of simple two-dimensional regular expressions. The separate vertical and horizontal iteration operators are not strong enough to represent some natural languages, for instance the languages Lsq consisting of squares of a’s. One way to increase the power of the iterating operators is to allows for more freedom within the iteration process. For Lsq one may use an alternating horizontal/vertical concatenation within the iteration process. Indeed, starting with a and iterating a horizontal concatenation on the right with a vertical string a and then a vertical concatenation on the bottom with a horizontal string a† a one precisely gets the language of squares of a’s. In other words, Lsq is the least solution of the equation X = a + (X a ) · (a† a). Another example considers the language Lspir consisting of ‘spiral words’ x 2aa 2aaaa . . .. A corresponding equation describing this language is 2x1 22aa1 bb1
22x11 2bb11 bbbb1
X = x + (2 a† ) · (2 X 1 ) · (b† 1). While in the above example the iterative process uses concatenation at right and bottom only, in the present case concatenation on all sides is used within the iteration process.
3 3.1
Finite Interactive Systems Some Motivation
A finite interactive system may be seen as a kind of two-dimensional automaton mixing together a state-transforming machinery with a machinery used for the interaction of different threads of the first automaton. Syntactically, they may be described by graphs as the ones in Fig.1. While this way of presentation may be somehow useful, it is also misleading. The first point is that there are not two different automata to be combined, but these two views are melted together to give this new concept of finite interactive systems. A next point is that the term ‘automaton’ is not the right one to consider. While there may be some other interpretations, an automaton is generally considered as a state-transforming device. A state is a temporal section of a running system: it gives the information related to the current values of the involved variables at a given temporal point. This information is used to compose the systems, to get more complex behaviors out of simpler ones. Then, using a term as ‘two-dimensional automaton’ is more or less as considering automata, but with a more complicate, say two-dimensional, state structure. The situation with finite interactive systems is by far different: one dimension (by convention, the vertical one) is used to model this state transformation, but the other dimension is considered to be a behavior transforming device. Orthogonal to the previous case of states, a behavior gives a spatial section of the running system: it considers information on all the actions a spatially located
Interactive Systems: From Folklore to Mathematics
203
agent has made during its evolution. This information is used at the temporal interfaces by which the agents interact, to get more complicate intelligent systems out of simpler ones. An agent is considered as a job-transforming device: given a job-request at its input temporal interface, the agent acts, transforms it, and passes the transformed job-request at its output temporal interface. One feature which is implicitly present in this view is that the systems are open: to run such a system one has to give an appropriate initial state, but also an initial job-request, actually a place where the starting user interacts with the system. The result of the running is a pair consisting of a final state of the system and a trace of the jobs/messages the system passes at an output user. Some more motivation along this line may be found in Sec.12.5 of [Ste00] dedicated to space-time duality. That section was written before the discovery of finite interactive systems. However, in some sense, it contains in the nutshell the seeds of this concept.
3.2
Finite Interactive Systems – Formal Definitions
After the above general considerations let us give a precise definition of interactive systems and explain it by a concrete example. Definition. A finite interactive system (fis) is a finite hyper-graph with two types of vertices and one type of (hyper) edges: — the first type of vertices is distinguished using a labeling with numbers (or lower case letters); such a vertex denotes a state (memory of variables); — the second type of vertices is distinguished using a labeling with capital letters; such a vertex is considered to be a class (of job requests); — the actions/transitions are labeled by letters denoting atoms of grids and are such that: (1) each transition has two incoming arrows: one from a class vertex, the other from a state vertex, and (2) each transition has two outgoing arrows: one to a class vertex, the other to a state vertex. Some classes/states may be initial (in the graphical representation this situation is specified by using small incoming arrows) or final (in this case the double circle representation is used).
An example is given in Fig.1(a). This fis has four vertices: two classes A and B and two states 1 or 2; moreover, A is an initial class, B is a final class, 1 is an initial state, and 2 is a final state. It also has three transitions labeled by a, b, and c. Sometimes it may be more convenient to have a simple non-graphical representation for fis’s. One possibility is to represent a fis by its transitions and to specify which states/classes are initial or final. (In this ‘cross’ representation of transitions the incoming vertices are those placed at north and west, while the outgoing ones are those from east and south.) For the given example, the interactive system is represented by:
204
G. S ¸ tef˘ anescu
Fig. 1. Examples of finite interactive systems
Interactive Systems: From Folklore to Mathematics 1
2
1
2
2
1
205
— A a B , A c A , and B b B and the information that — A, 1 are initial states/classes and B, 2 are final. 1
The execution of a transition A a B has two aspects: By firing action a the 2
state of the corresponding thread is changed from 1 to 2. However, this firing is possible only if a different thread requires the execution of a via class A. After the firing of a, our current thread may requires via class B other transitions to fire. One may see this as a highly abstract mechanism capturing the atomic parts (methods’ invocation) of the running of concurrent object oriented programs. As captured by fis’s, this mechanism of methods’ invocation is non-blocking. This means, after the passing of requirements at B the current thread is free to continue its evolution. Notice that the states are locals, each computation thread having its own set of variables. Similarly, the classes are local, each interactive object having a finite number of methods for its interaction interface. A simple example dealing with the ‘8 queens problem’ is given in Sec.2.4 of [Ste02]. Running patterns, accepting criteria. Interactive systems may be used to recognize grids. The procedure is the following. Definition. (1) Suppose we are given: a fis S, a grid w with m lines and n columns, a horizontal string with initial states bn = s1 . . . sn , and a vertical string with initial classes bw = C1 . . . Cm . Then: – Insert the states s1 , . . . , sn at the northern border of w (from left to right) and the classes C1 , . . . , Cm at the western border of w (from top to bottom). – Parse the grid w selecting minimal1 still unprocessed atoms. – For each such minimal unprocessed atom a, suppose s is the state inserted at its northern border and C is the class inserted at its western border. s
Then, choose a transition C a C from S (if any), insert the state s at s
the southern border of a and the class C at its eastern border, and consider this atom to be already processed. – Repeat the above two steps as long as possible. If finally all the atoms of w were processed, then look at the eastern border be and the southern border bs of the result. If they contain only final classes and final states, respectively, then consider this to be a successful running for w with respect to the border conditions bn , bw , be , bs . 1
The notion of minimal atoms we are using here is the one described when the flattening operator has been introduced.
206
G. S ¸ tef˘ anescu
(2) Given a fis S and four regular languages Bn , Bw , Be , Bs , a grid w is recognized by S with respect to the border conditions Bn , Bw , Be , Bs if there exists some strings bn ∈ Bn , bw ∈ Bw , be ∈ Be , bs ∈ Bs and a successful running for w with respect to the border conditions bn , bw , be , bs . Let us denote by L(S; Bn , Bw , Be , Bs ) their set. (3) Finally, a language of grids L is recognized by S if there are some regular languages Bn , Bw , Be , Bs such that L = L(S; Bn , Bw , Be , Bs ). (4) We simply denote the associated language by L = L(S) when the border conditions are superfluous, i.e., the border languages are the corresponding full languages and actually no border condition is imposed.
Example. A concrete example may be useful. Consider the fis Sq in Fig. 1(a) and suppose that no border conditions are imposed (that is, Bn = {1}∗ , Bw = {A}∗ , Be = {B}∗ , Bs = {2}∗ ). A running is: a b b
1 1 1 Aa b b Ac a b
1 1 1 AaBb b 2 Ac a b
1 1 1 AaBbBb 2 1 Ac a b
c a b c c a
Ac c a
Ac c a
Ac c a
w
w0
w1
w2
1 1 1 1 1 1 AaBbBb . . . AaBbBbB 2 1 1 2 1 AcAaBbB AcAa b 2 2 1 2 AcAcAaB Ac c a 2 2 2
w3
w9
The above sequence starts with a grid w. (1) The word obtained bordering w with initial states/classes is w0 . (2) At this stage all atoms are unprocessed, but only one is minimal, i.e., the a in the left-upper corner. In S there is only 1
one transition of the type A a
1
, i.e., A a B . After its application one gets 2
w1 . (3) Now there are two minimal unprocessed elements: b (position (1,2) in w) and c (position (2,1) in w). Suppose we choose b. Again, there is only one 1
appropriate transition B b B . After its application one gets w2 . (4) Next, the 1
minimal unprocessed elements are b (position (1,3) in w) and c (position (2,1) in 2
w). Suppose we choose c and the corresponding unique transition A c A . Then 2
we get w3 . (5) We can continue and finally an accepting running is obtained, leading to w9 . One may easily see that the recognized language L(S) consists of square words with a’s on the main diagonal, the top-right half filled in with b’s and the bottom-left half filled in with c’s. Finally, one may look at some border conditions. There is only one initial/final state/class, hence the border languages are regular languages over a one letter alphabet. To have an example, if Bn = (111)∗ , Bw = (AAAA)∗ , Be = B ∗ and Bs = (22)∗ , then L(S; Bn , Bw , Be , Bs ) is the sub-language of the above L(S) consisting of those squares whose dimension k is a multiple of 12.
Interactive Systems: From Folklore to Mathematics
207
State projection and class projection. Some familiar nondeterministic finite automata (nfa’s) may be obtained from fis’s neglecting one dimension. The automaton obtained from a fis S neglecting the class transforming part is call the state projection nfa associated with S and it is denoted by s(S). Similarly, the class projection nfa, denoted c(S), is the nfa obtained neglecting the state transforming part of S. The class projection nfa c(S) may be used to define macro-step transitions in the original interactive system S. Dually, the state projection nfa s(S) may be used to define agent behaviors (seen as job class transformers) in the original interactive system. (See Ch.12.5 of [Ste00] for more comments on this ‘job class transformers’ view.) Let us return to the fis S in Fig.1(a). Its class projection c(S) is given c
a
b
by the transitions A −→ B, A −→ A, B −→ B and the information that A is initial and B is final. The state projection s(S) is defined by transitions a
c
b
1 −→ 2, 2 −→ 2, 1 −→ 1 with 1 initial and 2 final. The fis S in Fig.1(a) has an interesting intersection property L(S) = L(s(S)) ∩ L(c(S)) showing that, in some sense, the interaction between the state and the class transformations of S is not so strong.2 3.3
Examples
This subsection contains some more toy examples of finite interactive systems. There is no claim that they describe some interesting, real interactive systems. They are merely used to illustrate the mathematical power of the mechanism. While syntactically some of them look quite similar, actually they recognize very unrelated languages. Squares. Let Lsq be the language
a
aa aa
aaa aaa aaa
aaaa aaaa aaaa aaaa
. . . consisting
of square grids of a’s. An example of a fis recognizing Lsq is Sq illustrated in Fig.1(a). Two examples of running patterns for Sq are shown below (the first one is successful; the second one is unsuccessful due to the fact that the southern border has an occurrence of the non-final state 1): 1 1 1 AaBaBaB 2 1 1 AaAaBaB 2 2 1 AaAaAaB 2 2 2
1 1 1 1 AaBaBaBaB 2 1 1 1 AaAaBaBaB 2 2 1 1 AaAaAaBaB 2 2 2 1
The same language is recognized by Sq2, drawn in the same figure. Making distinction between the various occurrences of a in Sq2 one gets another fis Sq1 2
To avoid some confusions, we emphasize here that the usual languages associated with the projection nfa’s are extended to the two-dimensional case. For instance, a grid w is in L(s(S)) if every column in w is recognized by s(S) in the usual sense. Similarly for L(c(S)), but now the lines are taken into account.
208
G. S ¸ tef˘ anescu
that recognizes the language of square grids with a’s on the main diagonal, the top-right triangle filled in with b’s, and the bottom-left triangle filled in with c’s. (Sq1 consists of the initial state 1, the initial class A, the final state 2, the final 1
2
1
2
2
2
1
2
classes B, and the following transitions A a B , A c A , B b B , B d B .)
Exponential function. Another interactive system is P ow, drawn in Fig.1(d). 10 100 . . ., grids whose number of lines is It recognizes the language 1 01 010 11
exponential (more precisaly, 2n − 1).
110 001 101 011 111
Pascal triangle. The fis P as in Fig.1(e) recognizes the set of grids over {0, 1} whose northern and western borders have alternating sequences of 0 and 1 (starting with 0) and, along the secondary diagonals, it satisfies the recurrence rule of the Pascal triangle, modulo 2 (that is, the value in a cell (i, j) is the sum of the values of the cells (i − 1, j) and (i, j − 1), modulo 2). In this case, the projection automata (on states and classes, respectively) are reset automata where all states/classes are final, hence they recognize any word. The intersection of their extensions to grids is the full language of grids, very far from the grids with Pascal-triangle property. This may be interpreted as a strong interaction between the state transformation and the class transformation within P as. It also shows that the ‘intersection’ property is not generally valid for the full class of finite interactive systems. Spiral grids. The finite interactive system drawn in Fig.1(c) recognizes the x 2aa 2aaaa . . .. language of ‘spiral’ grids 2x1 22aa1 bb1
3.4
22x11 2bb11 bbbb1
Recognizable Two-Dimensional Languages
It is known that the following statements are equivalent for a grid language L (called recognizable two-dimensional language): 1. L is recognized by a on-line tessellation automaton; 2. L is defined by a simple regular expression with intersection and adding homomorphisms; 3. L is defined by a local lattice language plus homomorphisms; 4. L is defined by an existential monadic second order formula.
Interactive Systems: From Folklore to Mathematics
209
The proof of this and many other informations may be found in the survey papers [GiR97,LMN98]. Alternatively, one may browse for papers starting from the following web page http://math.uni-heidelberg.de/logic/bb/2dpapers.html. Notice that regular expressions and local lattice languags do not cover the class of recognizable languages, but reach this using homomorphisms. This may be seen as an inconvenience of these mechanisms. On the other hand, this problem does not occur for on-line tessellation automata or for our finite interactive systems, as they are closed under homomorphisms. Our finite interactive systems represent a natural extension of finite automata to two dimensions and, as one may see below, capture recognizability. Find appropriate classes of two-dimensional regular expressions to capture recognizability without using homomorphisms is still an open question. Whether deterministic and (general) finite interactive systems are equivalent is also currently an open question. Local Lattice Languages (lll’s, for short), also called tile systems, are specified using a finite set of finite (rectangular) grids that are allowed as sub-grids of the recognized grids. If one combines this with (letter-to-letter) homomorphisms, then the power of the mechanism does not change by restricting to 2 × 2 pieces, or even to horizontal and vertical dominoes (i.e., 1 × 2 and 2 × 1 pieces). It is not too difficult to show that the class of grid languages specified by fis’s coincides with the class of recognizable two-dimensional languages: – One may simulate an lll by a fis and the class of languages recognized by fis’s is closed under letter-to-letter homomorphisms. From these simple observations it follows that the class of languages recognized by finite interactive systems contains the class of recognizable two-dimensional languages. – The converse inclusion is also valid: One can easily see that the running patterns of a fis S, completed with a special symbols for the blank space, are recognized by a finite tile system. Then, one may use an homomorphism to hide the state/class/blank information around the letters in order to capture L(S). Hence we get the following theorem: A set of grids is recognizable by a finite interactive system if and only if it is recognizable by a tiling system. This connection gives a way of inheriting many results known for two dimensional languages. Two important ones are: (1) if one consider the restriction to the top line of the grid languages recognized by fis’s, then one precisely gets context-sensitive string languages; and (2) the emptiness problem for finite interactive systems is undecidable.
4
Closing Comments
The concept of finite interactive systems presented here has evolved from a study of Mixed Network Algebras, see [Ste00]. Our finite interaction systems may be seen as an extension of finite transition systems [Arn94], Petri nets [GaR92], data-flow networks [BrS01], or asynchronous automata [Zie87], in the sense that a potentially unbound number of
210
G. S ¸ tef˘ anescu
processes may interact. In the mentioned models, while the number of processes may be unbounded, their interaction is bounded by the transitions’ breadth. This seems to be a key feature for the ability to pass from concurrent to concurrent, object-oriented agents. Our model has some similarities with the zero-safe nets model studied by Bruni and Montanari (see, e.g., [Bru99], [BrM97]), but the precise relationship still has to be investigated. We also expect that a logic similar to the spatial logic of [CaG00] may be associated with interaction systems and use for the verification of concurrent, object-oriented systems.
References [Arn94] [Bor99] [BCGS01] [BrS01] [Bru99] [BrM97] [CaG00] [Con71] [GaR92] [GiR97] [GrLS00] [GrSB98]
[InT91] [Kle56] [Koz94]
A. Arnold. Finite transition systems. Prentice-Hall, 1994. Borchert, B. http://math.uni-heidelberg.de/logic/bb/2dpapers.html M. Broy, G. Ciobanu, R. Grosu, and G. Stefanescu. Finite interactive systems: A unified model for agents’ behaviour and their interaction. Draft, December, 2001. M. Broy and G. Stefanescu. The algebra of stream processing functions. Theorertical Computer Science, 258:95–125, 2001. R. Bruni. Tile Logic for Synchronized Rewriting of Concurrent Systems. PhD thesis, Dipartimento di Informatica, Universita di Pisa, 1999. Report TD-1/99. R. Bruni and U. Montanari. Zero-safe nets, or transition synchronization made simple. Electronic Notes in Theoretical Computer Science, vol. 7(20 pages), 1997. L. Cardeli and A. Gordon. Anytime, anywhere: modal logics for mobile ambients. In: POPL-2000, Symposium on Principles of Programming Languages, Boston, 2000. ACM Press, 2000. J.H. Conway. Regular Algebra and Finite Machines. Chapman and Hall, 1971. V. Garg and M.T. Ragunath. Concurrent regular expressions and their relationship to Petri nets. Theoretical Computer Science, 96:285–304, 1992. D. Giammarresi and A. Restivo. Two-dimensional languages. In G. Rozenberg and A. Salomaa, editors, Handbook of formal languages. Vol. 3: Beyond words, pages 215–265. Springer-Verlag, 1997. R. Grosu, D. Lucanu, and G. Stefanescu. Mixed relations as enriched semiringal categories. Journal of Universal Computer Science, 6(1):112– 129, 2000. R. Grosu, G. Stefanescu, and M. Broy. Visual formalism revised. In Proceeding of the CSD’98 (International Conference on Application of Concurrency to System Design, March 23-26, 1998, Fukushima, Japan), pages 41–51. IEEE Computer Society Press, 1998. K. Inoue and I. Takanami. A survey of two-dimensional automata theory. Information Sciences, 55:99–121, 1991. S.C. Kleene. Representation of events in nerve nets and finite automata. In: C.E. Shannon and J. McCarthy, eds., Automata Studies, Annals of Mathematical Studies, vol. 34, 3–41. Princeton University Press, 1956. D. Kozen. A completeness theorem for Kleene algebras and the algebra of regular events. Information and Computation 110:366-390, 1994.
Interactive Systems: From Folklore to Mathematics [LMN98] [Ste98] [Ste00] [Ste01] [Ste02]
[Zie87]
211
K. Lindgren, C. Moore, and M. Nordahl. Complexity of two-dimensional patterns. Journal of Statistical Physics 91:909–951, 1998. G. Stefanescu. Reaction and control I. Mixing additive and multiplicative network algebras. Logic Journal of the IGPL, 6(2):349–368, 1998. G. Stefanescu. Network algebra. Springer-Verlag, 2000. G. Stefanescu. Kleene algebras of two dimensional words: A model for interactive systems. Dagstuhl Seminar on ”Applications of Kleene algebras”, Seminar 01081, 19-23 February, 2001. G. Stefanescu. Algebra of networks: Modeling simple networks, as well as complex interactive systems. In: H. Schwichtenberg and R. Steibr¨ ugen, eds., Proof and System Reliability, 49-78. Kluwer Academic Publishers, 2002. W. Zielonka. Notes on finite asynchronous automata. Theoretical Informatics and Applications, 21:99–135, 1987.
Relational Constructions in Goguen Categories Michael Winter Department of Computer Science University of the Federal Armed Forces Munich 85577 Neubiberg, Germany
[email protected] Abstract. Goguen categories were introduced as a suitable calculus for L-fuzzy relations, i.e., for relations taking values from an arbitrary complete Brouwerian lattice L instead of the unit interval [0, 1] of the real numbers. Such a category may provide some relational constructions as products, sums or subobjects. The aim of this paper is to show that under an assumption on the lattice L one may require without loss of generality that the related relations are crisp, i.e., all entries are either the least element 0 or the greatest element 1 of L.
1
Introduction
Since the mid-1970’s it has become clear that the calculus of relations is a fundamental conceptual and methodological tool in computer science just as much as in mathematics. While computer science applications are evolving rapidly in several areas as in communication, programming, software, data or knowledge engineering, mathematical foundations are needed to provide a basis for existing methods. One important application is the treatment of uncertain or incomplete information. To handle this kind of information, Zadeh [15] introduced the concept of fuzzy sets. Later on, Goguen [4] generalized this concept to L-fuzzy sets and relations for an arbitrary complete Brouwerian lattice L (or complete Heyting algebra) instead of the unit interval [0, 1] of the real numbers. Such structures constitute a Dedekind category introduced in [8]. In [7] the concept of fuzzy relation algebras was introduced as an algebraic formalization of mathematical structures formed by fuzzy relations, e.g., relations with coefficients from the unit interval, on a set with sup-min composition. Those algebras are equipped with a semi-scalar multiplication, i.e., an operation mapping an element from [0, 1] and a fuzzy relation to a fuzzy relation. Using this operation, one may characterize when a relation is 0-1 crisp, i.e., all entries are either the least element 0 or the greatest element 1. Unfortunately, such an operation does not exist if the unit interval [0, 1] is replaced by an arbitrary complete Brouwerian lattice. On the other hand, it was shown in [13] that a suitable algebraic formalization for arbitrary L-fuzzy relations demands an extra operator. In particular, it was shown that there is no formula in the theory of Dedekind categories expressing the fact that a given L-fuzzy relation is 0-1 crisp. Therefore, the concept of Goguen categories was introduced. It was shown that the standard model of H. de Swart (Ed.): RelMiCS 2001, LNCS 2561, pp. 212–227, 2002. c Springer-Verlag Berlin Heidelberg 2002
Relational Constructions in Goguen Categories
213
L-fuzzy relations, i.e., the matrix algebra with coefficients from L, is indeed a Goguen category. A Goguen category may provide some relational constructions as products, sums or subobjects (cf. [11,12]). Such a construction is given by an object together with a set of relations fulfilling some equations. For example, a relational product of two objects A and B is an object A × B together with two relation π : A × B → A and ρ : A × B → B such that π ; π IA ,
ρ ; ρ IB
π ; ρ = AB ,
π; π ρ; ρ = IA×B .
One may expect that the pairing Q; π R; ρ of two crisp relations Q and R is crisp again. If the projections are crisp then this is the case since the class of crisp relations is closed under the usual relational operations. Especially in applications, such a property seems to be essential. For example, a fuzzy controller may be described using the following picture. Linguistic Linguistic Variables Variables (Input) (Output) 1 1 Q1 S1 + +@ @ Q2 1 1 HS2 @ + H @ @ HH@ @ + R Q3 S j H 3 -B @ R crisp 1 1 AH YH * H + HH + Q4 1 1 S4 .. . + .. . + .. .. . . The linguistic variables are modeled by a disjoint union of several copies of a singleton set. The relation R corresponds to the rulebase of the controller, and the relations Qi and Si are the fuzzy sets corresponding to every linguistic variable. The controller itself is described by the relational term (
i∈I
Q i ; ιi ); R; (
ι j ; Sj ),
j∈J
where ιi denotes the injection into the disjoint union, followed by a suitable defuzzification. Now, if the injections are non-crisp or A or B is a product with non-crisp projection there would be a fuzzification which is not an integral part of the controller. This fuzzification arises from the specific choice of the product or sum. In this case, reasoning about the controller using the term above seems to be impossible or at least difficult. Unfortunately, there may exists such noncrisp injections or projections. Consider the Boolean lattice {0, a, b, 1} and the relations
214
M. Winter
10 a b π1 := b a, 01
10 b a ρ1 := a b , 01
10 1 0 π2 := 0 1, 01
10 0 1 ρ2 := 1 0. 01
Both pairs (π1 , ρ1 ) and (π2 , ρ2 ) constitute a product of two copies of a set with two elements, i.e., they fulfill the equations above. The first pair of relations is not crisp. But, this example also shows that there is a crisp version (π2 , ρ2 ) of the product, i.e., there are crisp relations between the same objects fulfilling the same set of equations. In our example, one may require without loss of generality that the projections are crisp. The aim of this paper is to show that under an assumption on the lattice L such a crisp version always exists. We assume that the reader is familiar with the basic concepts of allegories [3].
2
Lattices and Antimorphisms
In [14], it was shown that every Goguen category is isomorphic to the category of antimorphisms from a suitable lattice to the full subcategory of crisp relations. We will use this isomorphism throughout this paper. Therefore, we will recall some results from lattice theory (cf. [1]). We denote a lattice and the induced ordering by (L, , ) resp. . An atomic lattice is a lattice such that every element of x ∈ L is equal to the union of atoms less than or equal to x. We denote the set of all atoms by A. An element x of a complete lattice L is called completely irreducible iff for all subsets of M ⊆ L the property M = x implies x ∈ M . Notice, that an atom is completely irreducible which implies that every finite lattice has at least one completely irreducible element. Suppose f : L → L is a monotone function, i.e., x y implies f (x) f (y) for all x, y ∈ L, on a complete lattice L. If a is an element from L such that a f (a) then there exists a least fixedpoint µf (a) of f greater or equal to a, i.e., f (µf (a)) = µf (a), a µf (a) and if f (b) = b and a b then we have µf (a) b. A predicate P is called admissible (or continuous) iff for every set M ⊆ L the property P(x) for all x ∈ M implies P( M ). The so-called principle of fixedpoint induction states that for every admissible predicate P, monotone function f : L → L, and a ∈ L with a f (a) we may conclude from P(a) (basis step) and P(b) implies P(f (b)) for every b ∈ L (induction step) that P(µf (a)) holds. For n-ary predicates we get a similar result by taking the n-ary product of the underlying lattices. The set of all functions between two lattices (L1 , , ) and (L2 , , ) may be ordered pointwise. This structure is again a lattice, and will be denoted by (L1 → L2 , , ). This lattice inherits interesting properties from L2 , i.e., if L2 is complete Brouwerian then so is L1 → L2 . A function f : L1 → L2 is called antitonic iff x y implies f (y) f (x) for
all x, y ∈ L1 . The set of all antitone functions L1 → L2 constitutes a sublattice of L1 → L2 . Again, this sublattice inherits interesting properties from L1 → L2 .
Relational Constructions in Goguen Categories
215
A function f : L1 → L2 is called an antimorphism iff f ( M ) = f (M ) forall subsets M of L1 . Every antimorphism is antitonic, and we have f (0) = f ( ∅) = ∅ = 1. Given an element z ∈ L2 we denote with z˙ the quasi-constant antimorphism induced by z, which is defined by z(x) ˙ :=
1 iff x = 0, z iff x = 0.
anti
The set of all antimorphisms L1 → L2 need not be a sublattice of L1 → L2 or L1 → L2 . But this set forms a closure system, i.e., is closed under arbitrary ˙ x → 1. Therefore, it induces intersection and contains the greatest function 1, the following closure operation τ (f ) :=
{h | f h and h antimorphism}. anti
τ is a monotone mapping from L1 → L2 to L1 → L2 , and we have f τ (f ) and anti τ 2 (f ) = τ (f ) for all f . Now, L1 → L2 together with the operations · fi := i∈I fi forms a complete lattice. τ ( fi ) and i∈I
i∈I
Unfortunately, the definition of τ gives us no information about the image of τ (f ) for a given value x ∈ L1 . Therefore, consider the following operation ϕ(f )(x) :=
M ⊆L1 M =x
f (y). y∈M
This ϕ is monotone, expanding and we have the following. Lemma 1. Let L1 be a complete Brouwerian lattice, L2 a complete lattice and f : L1 → L2 be antitonic. Then we have τ (f ) = µϕ (f ). A proof may be found in [14]. The last lemma gives allows us to prove properties of τ using fixedpoint induction.
3
Dedekind Categories
In the remainder this paper, we use the following notations. To indicate that a morphism R of a category R has source A and target B we write R : A → B. The collection of all morphisms R : A → B is denoted by R[A, B] and the composition of a morphism R : A → B followed by a morphism S : B → C by R; S. The identity morphism on A is denoted by IA . In this section we recall some fundamentals on Dedekind categories [8,9]. Dedekind categories are called locally complete division allegories in [3].
216
M. Winter
Definition 1. A Dedekind category R is a category satisfying the following: 1. For all objects A and B the collection of all morphisms (or relations) R[A, B] is a complete Brouwerian lattice. Meet, join, the induced ordering, the least and the greatest element are denoted by , , , ⊥ ⊥AB , AB , respectively. 2. There is a monotone operation (called converse) such that for all relations Q : A → B and R : B → C the following holds: (Q; R) = R ; Q and (Q ) = Q. 3. For all relations Q : A → B, R : B → C and S : A → C the modular law Q; R S Q; (R Q ; S) holds. 4. For all relations R : B → C and S : A → C there is a relation S/R : A → B (called the left residual of S and R) such that for all X : A → B the following holds: X; R S ⇐⇒ X S/R. Corresponding to the left residual, we define the right residual by Q\R := (R /Q ) . This relation is characterized by Q; Y R ⇐⇒ Y Q\R. We will use some basic properties of relations in a Dedekind category throughout the paper without mentioning. These properties and their proofs may be found in [2,3,10,11,12]. Later on, we will investigate whenever a given equation holds in several Dedekind categories. For this purpose, we have to define a language and its meaning. This language is given by the equational theory of a distributive allegory [3], i.e., a relational category which need not be complete nor provide residuals. We require a set of object variables and a set of typed relation variables, i.e., every relation variable is of the form r : a → b where a and b are object variables.
Definition 2. The set of terms of type a → b is defined inductively as follows: 1. 2. 3. 4. 5. 6. 7.
Every relation variable r : a → b is a term of type a → b. If a is an object variable then Ia is a term of type a → a. If a and b are object variables then ⊥ ⊥ab and ab are terms of type a → b. If t is a term of type a → b then t is a term of type b → a. If t1 and t2 are terms of type a → b then t1 t2 is a term of type a → b. If t1 and t2 are terms of type a → b then t1 t2 is a term of type a → b. If t1 and t2 are terms of type a → b resp. b → c then t1 ; t2 is a term of type a → c.
We usually omit the type of a relation variable, for brevity. We will compute the value of a term in a distributive allegory. Such an allegory need not be a Dedekind category. For such a structure R, an environment σ over R is a function mapping every object variable a to an object A of R and every relation variable r : a → b to a relation R : σ(a) → σ(b). The update σ[R/r] of σ at the relation variable r : a → b with the relation R : σ(a) → σ(b) is defined for object variables a and relation variables s by σ(s) iff r = s σ[R/r](a) := σ(a), σ[R/r](s) := R iff r = s
Relational Constructions in Goguen Categories
217
As usual we denote a sequence of updates σ[Q/q][R/r][S/s] on an environment σ by σ[Q/q, R/r, S/s]. Definition 3. The value VR (t)(σ) of a term t of type a → b is defined inductively as follows: 1. 2. 3. 4. 5. 6. 7. 8.
VR (r)(σ) := σ(r), VR (Ia )(σ) := Iσ(a) , VR (⊥ ⊥ab )(σ) := ⊥ ⊥σ(a)σ(b) , VR ( ab )(σ) := σ(a)σ(b) , VR (t )(σ) := (VR (t)(σ)) , VR (t1 t2 )(σ) := VR (t1 )(σ) VR (t2 )(σ), VR (t1 t2 )(σ) := VR (t1 )(σ) VR (t2 )(σ), VR (t1 ; t2 )(σ) := VR (t1 )(σ); VR (t2 )(σ),
Obviously, the value of a term depends only on the values of variables occurring in it. Let S be a set of equations, i.e., a set such that every element is of the form t1 = t2 with t1 and t2 terms of equal type. Suppose the variables of S are within {r1 , . . . , rn }. We say that the relations R1 , . . . , Rn fulfill S, denoted by R1 , . . . , Rn |= S, iff for all equations t1 = t2 ∈ S and all environments σ over R the following holds VR (t1 )(σ[R1 /r1 , . . . , Rn /rn ]) = VR (t2 )(σ[R1 /r1 , . . . , Rn /rn ]). In some sense a relation of a Dedekind category may be seen as an L-relation. The lattice L may equivalently be characterized by the ideal relations, i.e., a BB = J, or by the scalar relations. relation J : A → B satisfying AA ; J; Definition 4. A relation α : A → A is called a scalar on A iff α IA and AA . The set of all scalars on A is denoted by ScR (A). AA ; α = α; The notion of ideal elements was introduced by J´ onsson and Tarski [5] and the notion of scalars by Furusawa and Kawahara [6]. Notice that the collection of ideal elements on A is isomorphic to the collecAA . tion of scalars on A via the mappings φ(J) := J IA and φ−1 (α) := α;
4
Goguen Categories
In [13], it was shown that we need an additional concept to define a suitable algebraic theory of L-fuzzy relations. Our approach introduces two operations mapping every relation to the greatest crisp relation it contains resp. to the least crisp relation it is included in. We now give the abstract definition. AB Definition 5. A Goguen category G is a Dedekind category with ⊥ ⊥AB = for all objects A and B together with two operations ↑ and ↓ satisfying the following:
218
M. Winter
1. R↑ , R↓ : A → B for all R : A → B. 2. (↑ , ↓ ) is a Galois correspondence between (G[A, B], ) and (G[A, B], ), i.e., S R↑ ⇐⇒ R S ↓ for all R, S : A → B. ↑ 3. (R ; S ↓ ) = R↑ ; S ↓ for all R : B → A and S : B → C. 4. If α = ⊥ ⊥AA is a nonzero scalar then α↑ = IA . anti ↑ 5. For all antimorphisms f : ScG (A) → G[A, B] such that f (α) = f (α) for all α ∈ ScG (A) and all R : A → B the following equivalence holds ↓ R α; f (α) ⇐⇒ (α\R) f (α) for all α ∈ ScG (A). α:A→A α scalar
In agreement with our intuition, we define crispness in an arbitrary Goguen category as follows. Definition 6. A relation R : A → B of a Goguen category is called crisp iff R↑ = R (or equivalently R↓ = R). The crisp fragment G ↑ of G is defined as the collection of all crisp relations of G. In [13] it was shown that the sets ScG (A) of scalars on A are isomorphic via AB IB . For this reason we identify those sets. We the mapping α → BA ; α; call this set the underlying lattice of G and denote it by Sc[G]. Furthermore, it was shown that G ↑ is a subcategory of G and together with the operations of G a simple Dedekind category. Theorem 1. Let L be an arbitrary complete Brouwerian lattice and R be a simple Dedekind category. Then RL defined by 1. 2. 3. 4. 5.
the objects of RL are the objects of R, RL [A, B] is the set of all antimorphisms from L to R[A, B], ˙ AB and ˙ AB , the constants are given by the antimorphisms I˙ A , ⊥ ⊥ intersection are defined componentwise, and transposition · fi := τ ( fi ) where fi is defined componentwise, i∈I
i∈I
i∈I
6. f ··, g :=τ (f ; g) where f ; g is defined componentwise, 7. f ·/h := · {k | k antimorphism and k ··, h f }, 8. f ↑ := R˙ where R := f (y), ↓
y=0
9. f := S˙ where S := f (1). is a Goguen category. A proof may be found in [14]. Notice, that the crisp relations in RL are those functions which are of the form R˙ for an R in R. Since the class of crisp relations is closed under all operations union and composition of crisp relations in RL can also be computed componentwise. Furthermore, in [14] the following pseudo-representation theorem was shown. Theorem 2 (Pseudo-Representation Theorem). Let G be a Goguen cateSc[G] Sc[G] gory. Then G ↑ is again a Goguen category and G and G ↑ are isomorphic.
Relational Constructions in Goguen Categories
219
For technical reasons, we will use the structure where (2) of Theorem 1 is replaced by the set of all antitone functions, union and composition are defined componentwise. Notice, that this structure is a distributive allegory. A Goguen category of the form RL is always a subclass of structure defined above.
L. Therefore, we will denote this structure by R Lemma 2. Let L be an arbitrary complete Brouwerian lattice and R be a De dekind category. For all antitone functions f, g, fi : L → R[A, B] for all i ∈ I
and h : L → R[B, C] we have (1) τ (f ) τ (g) = τ (f g), (2) τ ( τ (fi )) = τ ( fi ), i∈I
(4) τ (f ; τ (h)) = τ (f ; h), (5) τ (f ; h) = τ (h ; f ).
i∈I
(3) τ (τ (f ); h) = τ (f ; h), Proof. (1),(3)-(5) were shown in [14]. The inclusion in (2) is trivial. The other inclusion follows from τ (fi ) τ ( fi ) ∀i ∈ I : τ (fi ) τ ( fi ) ⇒ i∈I
i∈I
⇒
τ(
i∈I
τ (fi )) τ 2 (
i∈I
fi ) = τ (
i∈I
fi ).
i∈I
For atomic lattices, τ reduces to an operation that is easy to handle. Lemma 3. Let L1 be an atomic complete Brouwerian lattice, L2 a complete f (a). lattice and f : L1 → L2 be antitonic. Then we have τ (f )(x) = a∈A ax
Proof. : From f (a) τ (f )(a) we conclude that f (a) ax
τ (f )(a) = τ (f )( ax
a) = τ (f )(x)
ax
since τ (f ) is an antimorphism. : The other inclusion is shown by fixedpoint induction. Therefore, we define the predicate P(h) :⇔ ∀x : h(x) f (a). This predicate is admissible since ax
P(hi ) for all i ∈ I implies hi (x) ⇒
f (a) for all i ∈ I and x ∈ L ax
hi (x)
i∈I
⇔ P(
i∈I
f (a) for all x ∈ L ax
hi ).
220
M. Winter
The basis step P(f ) is trivial, and the induction step follows from ϕ(h)(x) = h(y)
=
M =x y∈M
f (a)
M =x y∈M ay M =x a M
f (a)
f (a).
= ax
From the principle of fixedpoint induction we conclude that τ (f )(x) for all x.
ax
f (a)
The last lemma shows that τ (f )(a) = f (a) for every atom of an atomic lattice. Later on, this property will allow us to construct from given relations fulfilling a set of equations a crisp relation fulfilling the same set of equations.
5
Complete Prime Filters and Proper Lattices
In this section we focus on the problem to find for a given complete Brouwerian lattice L a suitable atomic lattice L , i.e., the τ operation for L should be a canonical extension of the τ operation for L. Unfortunately, the usual embedding ψ of L into the powerset of all prime filters does not work. This embedding is not necessarily (upward) continuous such that the least upper bound of a subset M ⊆ L and union of the set ψ(y) for y ∈ M need not coincide. As a consequence, ϕ resp. τ for L is not a canonical extension of ϕ resp. τ for L. We have to switch to a special class of prime filters. Definition 7. A subset F ⊆ L of a complete Brouwerian lattice L is called a complete prime filter iff 1. 0 ∈ F , 2. x y ∈ F iff x ∈ F and y ∈ F for all x, y ∈ L, 3. M ∈ F iff ∃y ∈ M : y ∈ F for all subsets M ⊆ L. We will denote the set of all complete prime filters of L by FL . If FL = ∅ we call L proper. First, we want to study the class of proper lattices. Theorem 3. 1. Every linear ordering is proper. 2. If L has at least one completely irreducible element then L is proper. 3. The class of proper lattices is closed under arbitrary products. 4. A complete atomless Boolean algebra is not proper.
Relational Constructions in Goguen Categories
221
Proof. 1. The set F := L \ {0} for a linear ordering L is obviously a complete prime filter. 2. We prove the following equivalence which implies our assertion. A principal filter (x) = {y ∈ L | x y} is a complete prime filter iff x is completely irreducible. ⇒: Suppose M = x ∈ (x). Then there is a y ∈ M ∩ (x) since (x) is a complete prime filter. We conclude y x from y ∈ M and x y from y ∈ (x). This implies x = y ∈ M . ⇐: It is sufficient to show that M ∈ (x) implies thatthere is ay ∈ M with (x y) = x y. Define Mx := {x y | y ∈ M }. Then we have Mx = y∈M x M = x. This implies x ∈ Mx since x is completely irreducible. We conclude x = x y y for a y ∈ M . is a complete prime filter of the proper lattice Li for all i ∈ I. 3. Suppose Fi Fi is a complete prime filter of Li . Then F := i∈I
i∈I
4. We just give the sketch of the proof since we did not introduce the notions of prime and ultrafilters. Suppose F is a complete prime filter of a complete atomless Boolean algebra B. Then F is a prime filter and hence an ultrafilter of B. Since ultrafilters are maximal filters it can be shown that F is an atom or 0. Since B is atomless, F = 0 follows. Furthermore, F is an ultrafilter implies that F := {x F } is a maximal ideal and F ∩ F = ∅. Last but | x ∈ x = F = 1 ∈ F which shows that F is not a not least, we have F = x∈F
complete prime filter.
Notice, that property (2) of the last theorem implies that every finite lattice is proper. Furthermore, from (1) we conclude that the unit interval [0, 1] of the real numbers is also proper. Now, we define a function ψ : L → P(FL ) by ψ(x) := {F ∈ FL | x ∈ F }. This function is not necessarily injective. But we have the following. Lemma ψ(x y) = ψ(x) ∩ ψ(y), 4. 1. 2. ψ( M ) = ψ(y). y∈M
Proof. 1. Consider the following computation F ∈ ψ(x y) ⇐⇒ x y ∈ F ⇐⇒ x ∈ F and y ∈ F F is a prime filter ⇐⇒ F ∈ ψ(x) and F ∈ ψ(y) ⇐⇒ F ∈ ψ(x) ∩ ψ(y). 2. The assertion follows immediately from F ∈ ψ( M ) ⇐⇒ M ∈F ⇐⇒ ∃y ∈ M : y ∈ F
F is a complete prime filter
222
M. Winter
⇐⇒ ⇐⇒
∃y ∈ M
: F ∈ ψ(y) F ∈ ψ(y).
y∈M
For a function f : L1 → L2 we define fˆ : P(FL1 ) → L2 by fˆ(M) := f (x). x∈L1 M⊆ψ(x)
Let RL be a Goguen category. Then the structure RP(FL ) is again a Goguen category and ˆ maps every element of RL to an element of RP(FL ) . Lemma 5. Let L be an arbitrary complete Brouwerian lattice and R be a De dekind category. For all antitone functions f, g, fi : L → R[A, B] for all i ∈ I
and h : L → R[B, C] we have ˆ = fˆ , (1) fi , (3) f fi = i∈I
i∈I
(2) f g = fˆ gˆ,
(5) fˆ is antitonic, f (x). (6) fˆ({F }) =
ˆ (4) f ; h = fˆ; h,
x∈F
Proof. 1. Consider the following computation ( fˆi )(M) = fˆi (M) i∈I
i∈I
=
fi (x)
i∈I M⊆ψ(x)
=
fi (x)
M⊆ψ(x) i∈I
=
(
fi )(x)
M⊆ψ(x) i∈I
=( 2. First, we show that
M⊆ψ(x)
fi )(M).
i∈I
(f (x) g(x)) = (
M⊆ψ(x)
f (x)) (
M⊆ψ(x)
g(x)).
The inclusion is trivial. Suppose M ⊆ ψ(x) and M ⊆ ψ(y). Then M ⊆ ψ(x) ∩ ψ(y) = ψ(x y). Furthermore, f (x) g(y) f (x y) g(x y) since f and g are antitonic. We conclude that ( f (x)) ( g(x)) = (f (x) ( g(x)) M⊆ψ(x)
M⊆ψ(x)
M⊆ψ(x)
=
M⊆ψ(x)
(f (x) g(y))
M⊆ψ(x) M⊆ψ(y)
M⊆ψ(xy)
(f (x y) g(x y))
Relational Constructions in Goguen Categories
=
223
(f (x) g(x)).
M⊆ψ(x)
Now, consider the following computation (f g)(M) = (f g)(x) M⊆ψ(x)
=
(f (x) g(x))
M⊆ψ(x)
=(
f (x)) (
M⊆ψ(x)
g(x))
M⊆ψ(x)
= fˆ(M) gˆ(M) = (fˆ gˆ)(M). 3. The assertion follows immediately from )(M) = (f
f (x)
M⊆ψ(x)
=(
f (x))
M⊆ψ(x) = fˆ(M) = (fˆ )(M).
4. Analogously to (2). 5. If M ⊆ N we get {x | N ⊆ ψ(x)} ⊆ {x | M ⊆ ψ(x)} and hence fˆ(N) = f (x) f (x) = fˆ(M). N⊆ψ(x)
M⊆ψ(x)
6. We immediately conclude that f (x) = fˆ({F }) = {F }⊆ψ(x)
F ∈ψ(x)
f (x) =
f (x).
x∈F
The next lemma will show the key property we will need for our main theorem. Notice, that in the proof of this lemma it is essential that ψ is continuous. Lemma 6. If f : L1 → L2 is antitonic, then we have τ (τ (f )) = τ (fˆ). Proof. : Since f τ (f ) we get fˆ τ (f ) by Lemma 5(1) and hence τ (fˆ) τ (τ (f )). : Consider the property (∗)
τ (f ) τ (fˆ).
224
M. Winter
From (∗) we immediately conclude that τ (τ (f )) τ 2 (fˆ) = τ (fˆ). We proof (∗) ˆ τ (fˆ). by fixedpoint induction. Therefore, we define the predicate P(h) :⇔ h This predicate is admissible since P(hi ) for all i ∈ I implies
∀i ∈ I : hˆi τ (fˆ) ⇒
hˆi τ (fˆ)
i∈I
hi τ (fˆ) by Lemma 5(1)
⇔
i∈I
⇔ P(
hi ).
i∈I
The basis step P(f ) is trivial. Let M be a subset of FL , x such that M ⊆ ψ(x) and M a set with M = x. Then we define P := {M ∩ ψ(y) | y ∈ M }. Using Lemma 4(2) we conclude that
P =
(M ∩ ψ(y)) = M ∩
y∈M
ψ(y) = M ∩ ψ( M ) = M.
y∈M
Furthermore, we have h(y) h(z) for all y ∈ M which implies M∩ψ(y)⊆ψ(z) h(y) h(z). We conclude that
y∈M
N∈P N⊆ψ(z)
(∗∗)
M⊆ψ(x)
h(y)
M =x y∈M
h(z).
P =M N∈P N⊆ψ(z)
Now, the induction step follows from
ϕ(h)(M) =
ϕ(h)(x)
M⊆ψ(x)
M⊆ψ(x)
=
=
h(z)
by (∗∗)
ˆ h(N)
P =M N∈P
P =M N∈P N⊆ψ(z)
h(y)
M =x y∈M
τ (fˆ)(N)
induction hypothesis
P =M N∈P
= ϕ(τ (fˆ))(M) = τ (fˆ)(M). From the principle of fixedpoint induction we conclude P(µϕ (f )) and hence property (∗).
Relational Constructions in Goguen Categories
6
225
Equations in Goguen Categories
In this section we will investigate when a given equation holds in several Dedekind/Goguen categories introduced in the previous sections. Furthermore, we will show that if a set of equations is fulfilled by some relations of a Goguen category then there are crisp relations fulfilling the same set of equations. Lemma 7. Let RL be a Goguen category, t be a term, σ be an environment
L . Then we have over RL and σ be an environment over R 1. VRL (t)(σ) = τ (VR L (t)(σ)), (r), 2. VR (t)(σ ) where σ (r) := σ L (t)(σ ) = V RP(F ) 3. if x = 0 then (VR L (t)(σ ))(x) = VR (t)(σ ) where σ (r) := σ (r)(x). Proof. 1. We prove the assertion by structural induction. If t is a constant or a variable, the assertion is trivial since the corresponding function are antimorphisms. From Lemma 2 we conclude that a) τ (f ) τ (g) = τ (f g), b) τ (f ) · τ (g) = τ (f g), c) τ (f ) ··, τ (h) = τ (f ; h), d) τ (f ) = τ (f ), such that the assertion for all other cases follows immediately. 2. Again, the assertion is proved by structural induction. If t is a constant or a variable, the assertion is trivial. From Lemma 5(1)-(4) the assertion follows for all other cases immediately. 3. The assertion is trivial since all operations are defined componentwise and ˙ (x) = ⊥ ˙ for x = 0 we have I(x) = I and ⊥ ⊥ ⊥. Now, we are ready to prove our main theorem. Theorem 4. Let RL be a Goguen category with a proper lattice L, S be a set of equations with variable within {r1 , . . . , rn } and f1 , . . . , fn be elements of RL such that f1 , . . . , fn |= S. Then there are relations U1 , . . . , Un from R with U˙ 1 , . . . , U˙ n |= S. Proof. Suppose t1 = t2 ∈ S and σ is an environment over RL . For brevity, let ˜ (r) := σ ˜ (r), σ ˜ (r) := σ ˜ (r)({F }) with F ∈ FL and σ ˜ := σ[f1 /r1 , . . . , fn /rn ], σ σ ) for i = 1, 2. Then we conclude that hi := VR L (ti )(˜ σ ) = VRL (t2 )(˜ σ) VRL (t1 )(˜ ⇔ τ (h1 ) = τ (h2 ) ⇒ τ (h1 ) = τ (h2 ) ⇒ τ (τ (h1 )) = τ (τ (h2 )) ˆ 1 ) = τ (h ˆ 2) ⇔ τ (h (t1 )(˜ σ )) = τ (V (t2 )(˜ σ )) ⇔ τ (V RP(FL ) RP(FL ) (t1 )(˜ σ ))({F }) = (V (t2 )(˜ σ ))({F }) ⇒ (V RP(FL ) RP(FL ) σ ) = VR (t2 )(˜ σ ) ⇔ VR (t1 )(˜
Lemma 7(1)
Lemma Lemma Lemma Lemma
6 7(2) 3 7(3).
226
M. Winter
Now, let Ui := σ ˜ (ri ) = σ ˜ (ri )({F }) = (σ ˜ (ri ))({F }) = fˆi ({F }) = fi (x) x∈F
for i ∈ {1, . . . , n}, δ an environment over RL , δ˜ := δ[U˙ 1 /r1 , . . . , U˙ n /rn ] and ˜ ˜ ˙ δ˜ (r) := δ(r)(x). Notice, that VR L (ti )(δ) for i = 1, 2 is crisp, i.e., of the form R for a suitable R, since all constants and all values for variable in ti are crisp. Therefore, this relation is an antimorphism. ˜ ˜ For x = 0 we immediately conclude that VRL (t1 )(δ)(x) = = VRL (t2 )(δ)(x) since a value of a term is an antimorphism. If x = 0 we get for i = 1, 2 ˜ ˜ VRL (ti )(δ)(x) = τ (VR L (ti )(δ))(x) ˜ = VR L (ti )(δ)(x) ˜ = VR (ti )(δ ) = VR (ti )(˜ σ )
Lemma 7(1) ˜ VR L (ti )(δ) is an antimorphism Lemma 7(2)
where the last equality holds since for all variable ri ∈ {r1 , . . . , rn } within ˜ i )(x) = U˙ i (x) = σ t1 and t2 we have δ˜ (ri ) = δ(r ˜ (ri ). Last but not least, ˙ ˙ ˙ VRL (t1 )(δ[U1 /r1 , . . . , Un /rn ]) = VRL (t2 )(δ[U1 /r1 , . . . , U˙ n /rn ]) follows. Since G and G ↑
Sc[G]
are isomorphic, we get the following corollary.
Corollary 1. Let G be a Goguen category with a proper underlying lattice Sc[G], S be a set of equations with variable within {r1 , . . . , rn } and R1 , . . . , Rn relations such that R1 , . . . , Rn |= S. Then there are crisp relations Q1 , . . . , Qn with Q1 , . . . , Qn |= S. Since products, sums and subobjects induced by crisp partial identities are defined by equations, we may require without loss of generality that the related relations are crisp. Unfortunately, we were just able to prove Theorem 4 for Goguen categories RL with a proper lattice L. The experiences we made during searching a counterexample to this theorem in general leads us to the following conjecture. Conjecture 1. Theorem 4 is true for all Goguen categories RL .
References 1. Birkhoff G.: Lattice Theory. American Mathematical Society Colloquium Publications Vol. XXV (1940). 2. Chin L.H., Tarski A.: Distributive and modular laws in the arithmetic of relation algebras. University of California Press, Berkley and Los Angeles (1951). 3. Freyd P., Scedrov A.: Categories, Allegories. North-Holland (1990). 4. Goguen J.A.: L-fuzzy sets. J. Math. Anal. Appl. 18 (1967), 145-157. 5. J´ onsson B., Tarski A.: Boolean algebras with operators, I, II, Amer. J. Math. 73 (1951) 891-939, 74 (1952) 127-162. 6. Kawahara, Y., Furusawa H.: Crispness and Representation Theorems in Dedekind Categories. DOI-TR 143, Kyushu University (1997). 7. Kawahara, Y., Furusawa H.: An Algebraic Formalisation of Fuzzy Relations. Fuzzy Sets and Systems 101 (1999), 125-135.
Relational Constructions in Goguen Categories
227
8. Olivier J.P., Serrato D.: Cat´egories de Dedekind. Morphismes dans les Cat´egories de Schr¨ oder. C.R. Acad. Sci. Paris 290 (1980) 939-941. 9. Olivier J.P., Serrato D.: Squares and Rectangles in Relational Categories - Three Cases: Semilattice, Distributive lattice and Boolean Non-unitary. Fuzzy sets and systems 72 (1995), 167-178. 10. Schmidt G., Str¨ ohlein T.: Relationen und Graphen. Springer (1989); English version: Relations and Graphs. Discrete Mathematics for Computer Scientists, EATCS Monographs on Theoret. Comput. Sci., Springer (1993). 11. Schmidt G., Hattensperger C., Winter M.: Heterogeneous Relation Algebras. In: Brink C., Kahl W., Schmidt G. (eds.), Relational Methods in Computer Science, Advances in Computer Science, Springer Vienna (1997). 12. Winter M.: Strukturtheorie heterogener Relationenalgebren mit Anwendung auf Nichtdetermismus in Programmiersprachen. Dissertationsverlag NG Kopierladen GmbH, M¨ unchen (1998). 13. Winter M.: A new Algebraic Approach to L-Fuzzy Relations Convenient to Study Crispness. INS Information Science 139/3-4 (2001), 233-252 14. Winter M.: Representation Theory of Goguen Categories. Submitted to Fuzzy Sets and Systems. 15. Zadeh L.A.: Fuzzy sets. Information and Control 8 (1965), 338-353.
A Subintuitionistic Logic and Some of Its Methods Ernst Zimmermann
[email protected] Abstract. We develop a predicate logical extension of a subintuitionistic propositional logic. Therefore a Hilbert type calculus and a Kripke type model are given. The propositional logic is formulated to axiomatize the idea of strategic weakening of Kripke’s semantic for intuitionistic logic: dropping the semantical condition of heredity or persistence leads to a nonmonotonic model. On the syntactic side this leads to a certain restriction imposed on the deduction theorem. By means of a Henkin argument strong completeness is proved making use of predicate logical principles, which are only classically acceptable. Semantic tableaux and an embedding into modal logic are defined straightforward.
1
Introduction
In 1994 Greg Restall developed a class of subintuitionistic propositional logics axiomatizing the model theoretical idea, that strategic weakening could be imposed in various ways on Kripke’s semantic for intuitionistic logics: the model theoretical conditions of heredity, transitivity, reflexivity and absurdity may be given up in different combinations finally reaching models for a subintuitionistic basic logic. Restall not only described models for these logics, he also gave completeness proofs for 11 subintuitionistic propositional logics. Closely related to the considerations of Restall are the considerations of Johan van Benthem going back to 1986: with classical modal logic as starting point van Benthem tried to introduce aspects of nonmonotonicity into the models for intuitionistic propositional logic. Nonmonotonicity was his motivation to give up the condition of heredity in Kripke’s semantic for intuitionistic propositional logic. Van Benthem was able to give an axiomatization for his nonmonotonic Kripke model too, thus ending in a subintuitionistic logic. But the considerations of Restall and van Benthem were restricted to the propositional case, and especially Restall has given a summary of difficulties and negative results for the extension of subintuitionistic propositional logic to predicate logic. In this summary Restall has shown at least the direction of a further research: he stated that the usual intuitionistic models in terms of growing domains are not applicable in the case of subintuitionistic logics; instead of growing domains one has to choose for a predicate logical extension of any subintuitionistic logic models with constant domains he pointed out. Exactly this is the direction I would like to continue the work of Restall and van Benthem: that subintuitionistic propositional logic is regarded as base logic, H. de Swart (Ed.): RelMiCS 2001, LNCS 2561, pp. 228–240, 2002. c Springer-Verlag Berlin Heidelberg 2002
A Subintuitionistic Logic and Some of Its Methods
229
which axiomatizes the transitive, reflexive, but not hereditary or not persistent Kripke model. With regard to the fact, that this logic has two negations, we can claim that we treat two of Restall’s subintuitionistic logics. Further, a predicate logical extension of this subintuitionistic propositional logic is formulated. Bearing in mind that only quantifiers over constant domains may give any completeness results, we finally reach at a complete predicate logical extension of the subintuitionistic propositional logic being acceptable only from a classical point of view.
2
Hilbert Type Calculus and Kripke Type Model
Definition 1. The sets V, K, P constitute the language of predicate logic: a set V of denumerably many individual variables: x1 , . . . , xk , . . . ∈ V , for k ∈ ω; a set K of denumerably many individual constants: a1 , . . . , ak , . . . ∈ K, for k ∈ ω; a set P of predicates, which contains for every n ∈ ω a set Pn ⊂ P of denumerably many n-ary predicates: Pn1 , . . . , Pnk , . . . for k ∈ ω. The set P0 ⊆ P of predicates P0k for k ∈ ω is the set of propositional variables; the set of terms is the set T : t ∈ T , if t ∈ V or t ∈ K; r, s, t, t1 , t2 , . . . , tk , . . . with k ∈ ω are the terms, x, y, z are the variables, and a, b, c are the constants, with or without numerical indices. Further the following logical symbols are used: → , ∧, ∨ are dyadic connectives; ∀, ∃ are monadic quantifiers; (, ) are parentheses. The following well formed formulae wff are precisely the well formed formulae: Pjk (t1 , . . . , tj ) is a wff, if t1 , . . . , tj ∈ T and Pjk ∈ P ; (A → B), (A ∧ B), (A ∨ B) are wff, if A and B wff; ∀xA, ∃xA are wff, if A is a wff. Parentheses, especially external parentheses are often missing. Definition 2. X(s/t) has the meaning, that in the expression X – a term or a formula – the term s is substituted in a uniform manner at all places by the term t. A variable x is called free in (the formula) A if x is not bound in the formula A by a quantifier. A term t is called free for x in (the formula) A, if t is a constant, or if t is a variable and is not bound by a quantifier after substitution. Definition 3. A frame F is a triple W, R, D , where W = ∅, R ⊆ W × W , R is reflexive and transitive and D = ∅. A model M as usual relates a frame with a language by some mappings, i.e. a model M is a quadruple W, R, D, v , i.e. a frame F with a function v such that v(k) ∈ D for arbitrary constants k ∈ K and vα (Pnj ) ⊆ Dn for arbitrary n-ary predicates Pnj ∈ P and arbitrary α ∈ W . Finally a function h is defined mapping the set of individual variables V into D; i.e. h(x) ∈ D for every x ∈ V . Further on there holds: hxa (y) = h(y), if y = x, and hxa (y) = a, if y = x. Values of terms are defined as follows: tM,h := v(t), if t is a constant, and tM,h := h(t), if t is a variable. Then we extend the functions v and h to arbitrary wff and arbitrary α ∈ W : M,h n |=M , . . . , tM,h
∈ vα (Pnj ) for all Pnj ∈ P ; α Pj (t1 , . . . , tn )[h] iff t1 n M M |=α A → B[h] iff for all β ∈ W with αRβ holds: |=M β A[h] or |=β B[h]; M M |=M α A ∧ B[h] iff |=α A[h] and |=α B[h]; M M M |=α A ∨ B[h] iff |=α A[h] or |=α B[h];
230
E. Zimmermann
M x |=M α ∀xF [h] iff for all a ∈ D holds: |=α F [ha ]; M x |=M ∃xF [h] iff there is a a ∈ D such that |= α α F [ha ].
Definition 4. A wff A is valid iff |= A; |= A iff for all F , all v, all α and all h F,v |=α A[h]. A wff A is a consequence of a set of wff Γ iff Γ |= A; Γ |= A iff for F,v F,v all F , all v, all α and all h, if |=α Γ [h], then |=α A[h]. |=M α Γ [h] has the M meaning, that |=α C[h] for all C ∈ Γ . As a matter of fact the following wffs are not valid: A → (B → A), A → ((A → B) → B), (A → (B → C)) → (B → (A → C)), A → (B → (A ∧ B)). Lemma 1 (Coincidence of Substitution and Valuation of Terms). M x |=M α A(x/t)[h] iff |=α A[htM,h ]. Proved by an induction on the degree of A. Definition 5. The following axiom schemata and rules define the Hilbert type calculus of positive subintuitionistic predicate logic. A1. (A → (B → C)) → ((A → B) → (A → C)); A2. A → A; A3. (A → B) → (C → (A → B)); A4./A5. (A ∧ B) → A; (A ∧ B) → B; A6. (A → B) → ((A → C) → (A → (B ∧ C))); A7./A8. A → (A ∨ B); B → (A ∨ B); A9. ((A ∧ C) → B) → (((A ∧ D) → B) → ((A ∧ (C ∨ D)) → B)); PL1. ∀xF → F (x/t), if t is free for x in F ; PL2. ∀x(F → G) → (F → ∀xG), if x is not free in F or not free in G; PL3. ∀x(F ∨ G) → (F ∨ ∀xG), if x is not free in F ; PL4. F (x/t) → ∃xF , if t is free for x in F ; PL5. ∀x((F ∧ H) → G) → ((∃xF ∧ H) → G), if x is not free in F or not free in H, G; MP. if A and A → B, then B; RA. if A and B, then A ∧ B; GR. if F , then ∀xF . Finally we define the relation Γ A for pairs of sets of wffs Γ and wffs A, i.e. we define deducibility from assumptions. Γ A holds iff there is a sequence of wffs A1 , . . . , Ak for k ∈ ω, such that A = Ak and for every j ≤ k it holds that either Aj is an instance of an axiom, or Aj ∈ Γ , or there are g, h ≤ j such that Aj is a result of rule MP. applied on wffs Ag and Ah , or there are g, h ≤ j such that Aj is a result of rule RA. applied on wffs Ag and Ah , or there is h ≤ j such that Aj is a result of rule GR. applied on wff Ah , and either the quantified variable rule GR. contains is not free in Ah or not free in any Ag with g ≤ h and Ag ∈ Γ . Theorem 1 (On Deducibility I). If A ∈ Γ , then Γ A; if Γ A and Γ A → B, then Γ B; if Γ A and Γ B, then Γ A ∧ B; if Γ A, then Γ ∪ ∆ A; if Γ A, then there is a finite ∆ ⊆ Γ such that ∆ A;
A Subintuitionistic Logic and Some of Its Methods
231
if Γ A and ∆ C for all C ∈ Γ , then ∆ A; if Γ A, then Γ ∀xA, if x is not free in A or not free in Γ . Without Proof. Definition 6. The usual definition of consistency is assumed, where a set of wff Γ is consistent iff there is an A such that Γ A. A special set of formulae Γ∀→ is to be defined: Γ∀→ = {∀x1 , . . . , ∀xn (A → B) | ∀x1 , . . . , ∀xn (A → B) ∈ Γ }, where quantifier prefixes ∀x1 , . . . , ∀xn of implications A → B may consist of any number n of universal quantifiers or none of them. From time to time Γ ∪ {A1 , . . . , Ak } B is abbreviated as Γ, A1 , . . . , Ak B. Lemma 2. Γ, A, B C iff Γ, A ∧ B C. For the proof use the at once with A4., A5. and RA. provable, usual properties of conjunction: if Γ A and Γ B, then Γ A ∧ B; if Γ A ∧ B, then Γ A and Γ B. Then prove Γ, A, B A ∧ B and Γ, A ∧ B A and Γ, A ∧ B B. Finally use the structural properties of deducibility. Theorem 2 (On Deducibility II). If Γ∀→ ∪ {A} B, then Γ∀→ A → B; if Γ A ∧ B, then Γ A and Γ B; if Γ A or if Γ B, then Γ A ∨ B; if Γ ∪ {C} B and Γ ∪ {D} B and Γ C ∨ D, then Γ B; if Γ ∀xA, then Γ A(x/t); if t is free for x in A; if Γ A ∨ B, then Γ A ∨ ∀xB, if x is not free in Γ, A or not free in B; if Γ A(x/t), then Γ ∃xA; if t is free for x in A; if Γ ∪ {F } B and Γ ∃xF , then Γ B, if x is not free in Γ, B or not free in F. At first the modified deduction theorem is proved: if Γ∀→ ∪ {A} B, then Γ∀→ A → B; i.e. it is to be proved: if there is a sequence of wffs C1 , . . . , Ck for k ∈ ω, such that B = Ck and for every j ≤ k it holds that either Cj is an instance of an axiom, or Cj ∈ Γ∀→ ∪ {A}, or there are g, h ≤ j such that Cj is a result of rule MP. or RA. or GR. applied on wffs Ag and Ah , with respective variable conditions in case of rule GR.; then there is a sequence of wffs D1 , . . . , Dm for m ∈ ω, such that A → B = Dm and for every j ≤ m it holds that either Dj is an instance of an axiom, or Dj ∈ Γ∀→ , or Dj is a result of rule MP., or of rule RA., or of rule GR. with respective conditions. It can be shown that for a given sequence of wffs C1 , . . . , Ck a sequence of wffs D1 , . . . , Dm can be constructed effectively, and that in D1 , . . . , Dm at most assumptions from C1 , . . . , Ck are occuring. If B is an instance of an axiom schema, B is an implication, say B = B1 → B2 , and A3. together with MP. gives the desired result, i.e. C1 = B1 → B2
and D1 , . . . , D3 = B1 → B2 , (B1 → B2 ) → (A → (B1 → B2 )), A → (B1 → B2 ) . If B ∈ Γ∀→ , and B is an implication, then again A3. together with rule MP. leads to the desired result; the respective D1 , . . . , Dm can be constructed as in the former case. If B ∈ Γ∀→ and B = ∀x1 . . . ∀xn (B1 → B2 ), i.e. B is a multiple universal closure of an implication, we conclude with various instances of PL1. and several applications of MP. Γ∀→ (B1 → B2 )(x1 /y1 , . . . , xn /yn )
232
E. Zimmermann
and Γ∀→ B1 (x1 /y1 , . . . , xn /yn ) → B2 (x1 /y1 , . . . , xn /yn ), for y1 , . . . , yn not in Γ∀→ , B, A; now we apply once more A3. and MP. to get the desired result with several applications of rule GR. and instances of PL2. If B = A apply A2.; i.e. C1 = A and D1 = A → A . The last case ends the begin of the induction on the length of the deduction. Observe that assumptions occuring in D1 , . . . , Dm are at most assumptions occuring in C1 , . . . , Ck . Now consider the induction steps consisting of applications of rules MP., RA. and GR.. At first consider the case B being a result of MP.: apply A1. and make use of the induction base. If B is a result of RA., apply A6. and make use of the induction base. If B is a result of GR., i.e. B = ∀xC for some C, we have Γ∀→ ∪ {A} ∀xC; this is a result of Γ∀→ ∪ {A} C. We treat this case very carefully. The last claim has with respect to free variables the meaning, that there is a sequence C1 , . . . , Ck with Ck = C such that variable x is not free in any Ci with 1 ≤ i ≤ k, if Ci ∈ Γ∀→ ∪ {A}. Now induction base tells that there exists a sequence D1 , . . . , Dm such that Dm = A → C and such that assumptions occuring in D1 , . . . , Dm are at most assumptions occuring in C1 , . . . , Ck ; i.e. even free variables of assumptions in D1 , . . . , Dm are at most free variables of assumptions in C1 , . . . , Ck . Due to this reason we are allowed to apply GR., to reach D1 , . . . , A → C, ∀x(A → C) . If A = Ci for some Ci in C1 , . . . , Ck with 1 ≤ i ≤ k, x is not free in A, so we reach the desired result with an instance of PL2. and an application of MP.: D1 , . . . , A → C, ∀x(A → C), ∀x(A → C) → (A → ∀xC), (A → ∀xC) . This last claim has the meaning that Γ∀→ A → ∀xC. If A = Ci for all Ci in C1 , . . . , Ck with 1 ≤ i ≤ k, x may be free in A. But since x is not free in any assumption Ci ∈ Γ∀→ for Ci in C1 , . . . , Ck with 1 ≤ i ≤ k, we choose a variable y not in Γ∀→ , A, B, and substitute y for any free variable x in the sequence to reach C1 (x/y), . . . , Ck (x/y) , which does not affect the shape of Γ∀→ , A. Induction assumption tells that there is a sequence D1 , . . . , Dm such that Dm = A → Ck (x/y) and such that the free variables of this last sequence are at most the free variables of C1 (x/y), . . . , Ck (x/y) . The next step is an application of GR. to reach D1 , . . . , A → Ck (x/y), ∀y(A → Ck (x/y)) , where the variable condition for the application of GR. is given. The next wffs in the sequence are an instance of PL2. and a result of MP. D1 , . . . , A → Ck (x/y), ∀y(A → Ck (x/y)), ∀y(A → Ck (x/y)) → (A → ∀yCk (x/y)), A → ∀yCk (x/y) . It remains to change the quantifier in the last formula, and deduction theorem is proved. Next or elimination is proved: if Γ, C B and Γ, D B and Γ C ∨ D, then Γ B. At first we prove with A9. and the deduction theorem if A ∧ C B and A ∧ D B, then A ∧ (C ∨ D) B; now assume Γ, C B and Γ, D B; the finiteness of the deducibility relation guarantees that there is a finite ∆ ⊆ Γ such that ∆, C Band ∆, D B; we write this having Lemma 2 in mind as ∆ ∧ C B and ∆ ∧ D B, where ∆ is the multiple conjunction of all formulae in ∆; then we conclude with the just proved fact ∆ ∧ (C ∨ D) B; and then we conclude ∆, C ∨ D B and Γ, C ∨ D B; therefore we have that, if Γ, C B and Γ, D B, then Γ, C ∨ D B; this last claim is trivially equivalent to or elimination on the background of the defined deducibility relation.
A Subintuitionistic Logic and Some of Its Methods
233
At last existence elimination is proved: if Γ ∪ {F } B and Γ ∃xF , then Γ B, if x is not free in Γ, B or not free in F . Clearly existence elimination is equivalent to: if Γ ∪ {F } B, then Γ ∪ {∃xF } B, if x is not free in F or not free in Γ, B. To prove this last claim we first deduce the rule if (A ∧ B) → C and x is not free in A or not free in B, C, then (∃xA ∧ B) → C. Therefore assume (A ∧ B) → C and x not to be free in A or not in B, C; with GR. we conclude ∀x((A ∧ B) → C), and with the mentioned variable condition, PL7. and an application of MP. (∃xA ∧ B) → C. Now we assume Γ, A C and x not to be free in A or in Γ, C; then we have a finite ∆ ⊆ Γ with ∆, A C; write this finite ∆ with Lemma 2 as ∆, the multiple conjunction of all formulae in ∆, then we have ∆, A
C, A ∧ ∆ C and with deduction theorem
(A ∧ ∆) → C; with the deduced rule and the mentioned variable conditions
(∃xA∧ ∆) → C; this is equivalent with deduction theorem to ∃xA, ∆ C; at last conclude ∆, ∃xA C and Γ, ∃xA C. Theorem 3 (Strong Correctness). If Γ A, then Γ |= A. The reader is asked to convince himself that all axiom schemata are valid, and all rules of the defined deducibility relation lead to consequences; then the proof of strong correctness goes by induction on the length of a deduction. Definition 7. Now assume a language L with a set of constants K and a set of predicates P , i.e. L = K, P . Assume the constants from Γ, A, B to be in K and the predicates in P . Γ is deductively closed iff if Γ A, then A ∈ Γ ; Γ has the or property iff if A ∨ B ∈ Γ , then A ∈ Γ or B ∈ Γ ; Γ is prime iff Γ is consistent, deductively closed and has the or property; Γ is existence complete in the language L iff if Γ ∃xB, then there is a c ∈ K such that Γ B(x/c); Γ is all complete in L iff if Γ B(x/c) for all c ∈ K, then Γ ∀xB; Γ is saturated in L iff Γ is prime, all complete in L and existence complete in L. Lemma 3 (On saturated sets). 1. If Γ B, then there exists a saturated ∆ such that Γ ⊆ ∆ and B ∈ ∆; 2. if A → B ∈ Γ and Γ is saturated in L, then there exists a saturated ∆ in L such that Γ∀→ ⊆ ∆, A ∈ ∆ and B ∈ ∆. These two lemmata show the difference to the proof of completeness in the intuitionistic case: there Lemma 3.2 is a corollary to Lemma 3.1, but in the absence of the full deduction theorem it is not; the proof of Lemma 3.2 is the hard case due to the additional fact, that we are not allowed to extend languages. Proof of Lemma 3.1. Assume Γ B and assume an enumeration of all disjunctions, an enumeration of all existence formulae, an enumeration of all universal formulae and assume constants, new relative to languages associated with certain given sets of formulae; define a sequence of sets of formulae ∆k and a sequence of formulae Ak , where ∆0 = Γ and A0 = B. If k = 0 (mod 3) in ∆k , take the first disjunction C ∨ D in the enumeration, which has not been treated, such that ∆k C ∨D: if ∆k ∪{C} Ak , then ∆k+1 = ∆k ∪{D} and Ak+1 = Ak ; if ∆k ∪ {C} Ak , then ∆k+1 = ∆k ∪ {C} and Ak+1 = Ak . Or elimination guarantees either ∆k ∪ {D} Ak or ∆k ∪ {C} Ak . If k = 1 (mod 3) in ∆k , take the first universal formula ∀xC in the enumeration, which has not been
234
E. Zimmermann
treated: if ∆k Ak ∨ ∀xC, then ∆k ∪ {∀xC} Ak with or elimination, and ∆k+1 = ∆k ∪ {∀xC} and Ak+1 = Ak ; but if ∆k Ak ∨ ∀xC, then it holds with Theorem 2 that ∆k Ak ∨ C(x/y) for a y ∈ ∆k , Ak , C, and for a new constant c, which is not in the language of ∆k , Ak , C we have ∆k Ak ∨ C(x/c); then we define ∆k+1 = ∆k and Ak+1 = Ak ∨ C(x/c). If k = 2 (mod 3) in ∆k , take the first existence formula ∃xC not yet been treated such that ∆k ∃xC; with existence elimination it holds that ∆k ∪ {C(x/y)} Ak for y not in ∆k , Ak , C; further it holds for a new constant c, which is not in the language of ∆k , Ak , C ∆k ∪ {C(x/c)} A k , and ∆k+1 is defined as ∆k ∪ {C(x/c)} and Ak+1 = Ak . Finally define ∆ = 0≤k ∆k . Now prove that ∆ has all the desired properties: Γ ⊆ ∆, B ∈ ∆ because B ∈ ∆k for all k, and ∆ is saturated, i.e. ∆ is prime, ∆ is all complete and ∆ is existence complete. This proof is similar to the proof of a related Henkin lemma in Thomason, 1968, p. 3, with appropriate modifications. Lemma 4. If Γ is saturated in L, then Γ∀→ is all complete in L. Assume L = K, P , Γ∀→ B(x/c) for all c ∈ K, and Γ to be saturated in L, show that Γ∀→ ∀xB; prove this by an induction on the length of the deduction of one B(x/ck ). If B(x/ck ) is an instance of an axiom schema or a generalisation of some, then B as well, and with GR. we conclude ∀xB. If B(x/ck ) ∈ Γ∀→ , then B(x/ck ) is an implication or a (multiple) universal implication; further on Γ B(x/c) for all c ∈ K, and ∀xB ∈ Γ since Γ is all complete in L; further on ∀xB ∈ Γ∀→ , because of the definition of Γ∀→ and the fact that ∀xB has to be a (multiple) universal implication. These cases constitute the begin of the induction, now we change to the induction step consisting of applications of rules MP., RA. or GR. Assume Γ∀→ B(x/ck ) to be a result of MP., i.e. there is a C such that Γ∀→ C → B(x/ck ) and Γ∀→ C; for all other cn Γ∀→ B(x/cn ). Therefore Γ∀→ ∪ {C} B(x/cn ); with deduction theorem we have Γ∀→ C → B(x/cn ) for all cn , i.e. this holds for really all c ∈ K. We treat the most complicated case, where C may contain x as free variable. Deduce for all c Γ∀→ C → B(x/y)(y/c) for a variable y not free in B, C, and further Γ∀→ (C → B(x/y))(y/c). Then apply induction assumption on variable y, since induction begin was proved for arbitrary variables and get Γ∀→ ∀y(C → B(x/y)). Predicate logical reasoning gives Γ∀→ C → ∀yB(x/y); further Γ∀→ ∀yB(x/y), and the desired Γ∀→ ∀xB. The proof of the lemma is completed by considering Γ∀→ B(x/ck ) to be a result of rules RA. and GR. Proof of Lemma 3.2. Assume L = K, P , A → B ∈ Γ and Γ to be saturated in L; show that there exists a saturated ∆ in L such that Γ∀→ ⊆ ∆, A ∈ ∆ and B ∈ ∆. With the assumption we have Γ A → B and Γ∀→ A → B and with the deduction theorem Γ∀→ ∪ {A} B. According to Lemma 4 Γ∀→ is all complete in L. Now construct a sequence of sets of formulae ∆k and a sequence of formulae Bk with ∆k = Γ∀→ ∪{A, D1 , . . . , Dk−1 }, i.e. set ∆0 = Γ∀→ and ∆1 = Γ∀→ ∪{A}; further B0 = B. Assume an enumeration of all disjunctions, an enumeration of all
A Subintuitionistic Logic and Some of Its Methods
235
universal formulae and an enumeration of all existence formulae and continue to define ∆k as follows. If k = 0 (mod 3) in ∆k , take the first disjunction C ∨ D in the enumeration, which has not been treated, such that ∆k C ∨ D: if ∆k ∪ {C} Bk , then ∆k+1 = ∆k ∪ {D} and Bk+1 = Bk ; if ∆k ∪ {C} B, then ∆k+1 = ∆k ∪ {C} and again Bk+1 = Bk . Or elimination guarantees either ∆k ∪{D} B or ∆k ∪{C} B. If k = 1 (mod 3) in ∆k , take the first universal formula ∀xC in the enumeration, which has not been treated: if ∆k Bk ∨ ∀xC, then ∆k ∪{∀xC} Bk with or elimination, and ∆k+1 = ∆k ∪{∀xC} and Bk+1 = Bk ; but if ∆k Bk ∨ ∀xC, then ∆k ∀x(Bk ∨ C) with PL3. and eventually an appropriate renaming of bound variable x. This can be written as Γ∀→ ∪{A∧D1 ∧ . . .∧Dk−1 } ∀x(Bk ∨C), and Γ∀→ (A∧D1 ∧. . .∧Dk−1 ) → ∀x(Bk ∨C). This is with PL2. Γ∀→ ∀x((A ∧ D1 ∧ . . . ∧ Dk−1 ) → (Bk ∨ C)); with all completeness of Γ∀→ conclude Γ∀→ ((A∧D1 ∧. . .∧Dk−1 ) → (Bk ∨C))(x/c) for some c ∈ K. This amounts finally to ∆k Bk ∨ C(x/c) for a c ∈ K. Now define ∆k+1 = ∆k and Bk+1 = Bk ∨ C(x/c). If k = 2 (mod 3) in ∆k , take the first existence formula ∃xC in the enumeration not yet been treated, such that ∆k ∃xC: since ∆k Bk , ∆k ∪ {∃xC} Bk . Now each ∆k can be written as Γ∀→ ∪ {A, D1 , . . . , Dk−1 }, Γ∀→ being allcomplete; i.e. we have Γ∀→ ∪{A∧D1 ∧. . .∧Dk−1 ∧∃xC} Bk ; now it holds that Γ∀→ (A ∧ D1 ∧ . . . ∧ Dk−1 ∧ ∃xC) → Bk ; and with an appropriate renamed bound variable x and an application of axiom schema PL5. we reach Γ∀→ ∀x((A ∧ D1 ∧ . . . ∧ Dk−1 ∧ C) → Bk ); again with all completeness of Γ∀→ this gives Γ∀→ ((A ∧ D1 ∧ . . . ∧ Dk−1 ∧ C) → Bk )(x/c) for some c ∈ K; this amounts after reversing the processes the desired ∆k ∪ {C(x/c)} B for one c ∈ K; define here ∆k+1= ∆k ∪ {C(x/c)} and Bk+1 = Bk . Finally define ∆ = 0≤k ∆k , then ∆ has all desired properties; deductive closure, or property, nonelementhood of B, existence completeness and all completeness, the proofs are as shown in Lemma 3.1. Theorem 4 (On Saturated Sets). For all in L = K, P saturated sets of formulae Γ it holds: A → B ∈ Γ iff for all in L saturated ∆ with Γ∀→ ⊆ ∆: A ∈ ∆ or B ∈ ∆; A ∧ B ∈ Γ iff A ∈ Γ and B ∈ Γ ; A ∨ B ∈ Γ iff A ∈ Γ or B ∈ Γ ; ∀xF ∈ Γ iff for all c ∈ K F (x/c) ∈ Γ ; ∃xF ∈ Γ iff there is an c ∈ K such that F (x/c) ∈ Γ . The proofs of these theorems are well done with Lemma 3.2 and the theorems on deducibility 2. Definition 8. The canonical model M for a in L = K, P saturated set of formulae Γ is a quadruple W, R, D, v defined as follows: W = {α|Γ∀→ ⊆ α and α is saturated in L}, αRβ iff α∀→ ⊆ β, D = K, v(c) = c for arbitrary c ∈ K, vα (Pnj ) = { c1 , . . . , cn |c1 , . . . , cn ∈ D, α ∈ W, Pnj (c1 , . . . , cn ) ∈ α} for all α ∈ W and all Pnj ∈ P . Lemma 5 (Canonical Truth). For all α of the canonical model M , all A and all h it holds: |=M α A[h] iff A ∈ α. Proof by induction on the degree of A with the theorems on saturated sets 4.
236
E. Zimmermann
Theorem 5 (Strong Completeness). If Γ |= C, then Γ C. For strong completeness assume Γ C, show Γ |= C; Lemma 3.1 gives a saturated ∆ ⊇ Γ such that C ∈ ∆. With the truth Lemma 5 it holds in the canonical M model for ∆ for any h: |=M ∆ C[h] and |=∆ Γ [h]. The developed logic can be extended easily with negation: add ⊥ as a propositional variable to the underlying language and add the definition ¬A ↔df (A → ⊥). Now, as well as in intuitionistic logic, there are two negations: a strong negation, according to the additional axiom schema ⊥ → A, and the additional semantical definition |=M α ⊥ for any α; and a weak negation without any additional axiom schema and semantical definition. In the case of the strong negation ⊥ ∈ Γ can be proved for any saturated set Γ , so that the truth lemma may be extended to the new language and the new axiom schema to get further results of correctness and completeness. The following wff is valid with weak and therefore with strong negation: (A → B) → (¬B → ¬A). The following two wffs are not valid with strong and therefore not with weak subintuitionistic negation – but they are valid with weak and therefore with strong intuitionistic negation: A → ¬¬A, (A → ¬B) → (B → ¬A).
3 3.1
Semantic Tableaux The Method of Semantic Tableaux
The structure underlying the method of semantic tableaux is an ordered class of objects called tableaux. Every tableau α consists of a tree of signed wff, i.e. wff being true (T ) or false (F ), the root of the tree at the top. As language for subintuitionistic tableaux presuppose the previously defined language of subintuitionistic logic (may be with ⊥). The formula trees within the tableaux are at most binary branching, and branches may be infinitely long. A tableau is called to be closed, if every branch of the tree of signed wff is closed; and a branch is closed, if it contains a pair T A and F A for some wff A. Now a wff A is provable with the tableaux method, if it is true in every tableaux under any assignment of constants to variables. The tableaux method proceeds by searching a counterexample: to prove a wff A assume F A under an arbitrary assignment in a tableau, say α, and look for closure of tableau α. Clearly the assignment of truth values (T, F ) to wff A in dependence on tableaux α proceeds inductively: basically, as already told, one defines closure of a branch of any tableaux α, if the branch contains a pair T A and F A for some wff A. For the inductive steps one assumes a definition by the following rules, holding for all tableaux α: αT
T T
A∧B α A F B
A
F
A∧B / F
B
A Subintuitionistic Logic and Some of Its Methods
α
T
A
(α F α
F
αF β
T
A∨B / T
B
αF
F F
237
A∨B A B
⊥)
A
T
A→B / T
B
A→B T F
A B
The rule F A → B needs some explanation. If a signed wff F A → B occurs in the formula tree of a tableau α one has to create a new tableau, say β. β has as entries all T C → D for any C, D, with T C → D ∈ α – and additionally T A and F B are entries of β. The following description should be a clear working device: copy the subtree of the signed wff F A → B from α to β, and delete in this copy all entries not being signed wffs T C → D for any wff C, D, and at the place of F A → B put T A and F B. α T ∀xA α F ∀xA where a is any constant, and b is a new constant T Axa F Axb assigned to variable x. α T ∃xA α F ∃xA where b is a new constant, and a is any constant F Axa T Axb assigned to variable x. These rules are nearly self explanatory. All rules except the one and only rule for F A → B are classically boolean. Given a signed wff T A or F A in a tableau α, with the exception of F A → B, one has to add the respectively listed signed wffs to the tree of wffs in the tableau α, i.e. the respectively listed signed wffs branching or not branching – have to be added to the end of the related branch of the original wff. To give an example: if T A → B occurs in the tree of wffs of a tableau α, the signed wffs F A and T B have to be added as branching entries to the end of that branch, in which T A → B occurs in the formula tree of α. Only the rule F A → B requires the creation of a new tableau with respect to the described conditions, and hence is not a classical rule. 3.2
Equivalence Theorem
A wff is provable by tableaux iff it is semantically valid. Prove this from left to right by connecting a closed branch of a signed wff in a tableau with a semantically unsatisfiable set of wff. For the proof from right to left a slight modification of Fitting’s idea of a consistency property is sufficient (Fitting, 1983, p. 481 ff.).
238
E. Zimmermann
3.3
Examples
Two examples of tableaux conclude the section. One disproves A → (B → A), the other proves (A → B) → (C → (A → B)). α
F
(A → B) → (C → (A → B))
β F
F
A
γ
A→B T C F A→B /
A
δ
T F
αF
γ
4 4.1
T
B
T
B
T
B
T
F
β
T A→B C → (A → B) /
A
A→B T A F B /
A → (B → A) F
T A B→A T F
B A
An Embedding into Modal Logic S4 A Language of Modal Predicate Logic
Take the language of predicate logic as previously defined, ⊥ included, and add the logical sign 2, a necessity operator. Add as wffs all expressions 2A, where A is a wff. This is the language of modal predicate logic. 4.2
A Model for Modal Predicate Logic S4 with Barcan’s Formulas
In the sequel we are interested in a special, well known system of modal predicate logic, classical modal predicate logic S4 with the formulas of Barcan. Since we are only interested in a semantical description of this logic it suffices to define a respective model. The fact that this model is exactly the model for the described logic can be found in many books on modal logic. For that reason suppose a model N as a quadruple W, R, D, v , and define the components of this quadruple
A Subintuitionistic Logic and Some of Its Methods
239
exactly similar to the components of the model M of subintuitionistic logic. Make only one difference: after having defined the functions v and h on terms, extend these definitions of v and h as follows to arbitrary wff and arbitrary α ∈ W: N,h n N,h n n |=N α Pj (t1 , . . . , tn )[h] iff t1 , . . . , tn ∈ vα (Pj ) for all Pj ∈ P ; N not |=α ⊥[h]; N N |=N α A → B[h] iff not |=α A[h] or |=α B[h]; N N N |=α A ∧ B[h] iff |=α A[h] and |=α B[h]; N N |=N α A ∨ B[h] iff |=α A[h] or |=α B[h]; N x |=α ∀xF [h] iff for all a ∈ D holds: |=N α F [ha ]; N x |=α ∃xF [h] iff there is a a ∈ D such that |=N α F [ha ]; N N |=α 2A[h] iff for all β ∈ W with αRβ holds: |=β A[h]. 4.3
A Translation T
Define recursively a translation T of the language of predicate logic into the language of modal predicate logic: T (Pjk ) = Pjk , T (⊥) = ⊥, T (A → B) = 2(T (A) → T (B)), T (A∧B) = T (A)∧T (B), T (A∨B) = T (A)∨T (B), T (∀xA) = ∀xT (A), T (∃xA) = ∃xT (A). 4.4
An Embedding Lemma
Remember M to be the model for subintuitionistic logic, and N for classical modal predicate logic S4 with Barcan’s formulas, then it holds – if the valuations of the models M and N are defined to agree on atoms: N |=M α A[h] iff |=α T (A)[h] .
Prove this by induction on the degree of A. From this embedding lemma derive that subintuitionistic validity of a wff A and modal validity of the translation T (A) of this formula agree – for this result presuppose a straightforward definition of modal validity.
References van Benthem, J.: Partiality and Nonmonotonicity in Classical Logic, in: Logique et Analyse, 29 (1986) p. 226-246 Fitting, M.C.: Intuitionistic Logic, Model Theory and Forcing, North-Holland, Amsterdam (1969) Fitting, M.C.: Proof Methods for Modal and Intuitionistic Logic, Kluwer, Dordrecht (1983) Gabbay, D.: Semantical Investigations in Heyting’s Intuitionistic Logic, Reidel, Dordrecht (1981) Kripke, S.A.: Semantical Analysis of Intuitionistic Logic (1965), in: Crossley, J.N., Dummett, M.A.E. (Eds.): Formal Systems and Recursive Functions, NorthHolland, Amsterdam (1965), p. 93-130
240
E. Zimmermann
Restall, G.: Subintuitionistic Logics, in: Notre Dame Journal of Formal Logic, 35 (1994) p. 116-129 Thomason, R.H.: Strong Completeness of Intuitionistic Predicate Logic, in: Journal of Symbol Logic, 33 (1968) p. 1-7
Remark. The author would like to thank two anonymous referees for critical and competent comments, which certainly helped to improve this paper.
Implementation of Relational Algebra Using Binary Decision Diagrams Rudolf Berghammer, Barbara Leoniuk, and Ulf Milanese Institut f¨ ur Informatik und Praktische Mathematik Christian-Albrechts-Universit¨ at Kiel Olshausenstraße 40, D-24098 Kiel, Germany
Abstract. We show how relations and their operations can efficiently be implemented by means of Binary Decision Diagrams. This implementation is used in the computer system RelView. To demonstrate the power of the approach, we show how it can be applied to attack computationally hard problems
1
Introduction
In the last decades, relational algebra has widely been used as a very convenient means of dealing with fundamental concepts like graphs, orders, lattices, games, data bases, Petri nets, data types, and semantics in mathematics and computer science. For many examples and references to relevant literature, see [11,6]. One main reason for using relational algebra is that it has a small but surprisingly very powerful set of operations all of which can easily and efficiently be implemented. Thus, a supporting computer system can easily be implemented, too. RelView (see [1,3] for an overview) is such a tool for the manipulation and visualization of relations and relational programming. It has been developed at Kiel University since 1993. Written in the C programming language, it runs on Sun SPARC workstations and INTEL-based Linux systems1 . The benefits of RelView are manifold and have been described in [5,1,2,4] for example. But in several applications its use was restricted to small input relations only due to the array-like implementation of relations in which every entry consumes 1 Bit of memory. As an example, the membership relation ∈ between a set X and its powerset 2X for |X| = 25 allocates more than 100 MByte of storage. In this particular case, the memory consumption is exponential in the size of X. Therefore, in the last three years we have exchanged RelView’s implementation of relations and their operations to overcome these problems. We have carried out a number of experiments which showed that a promising alternative is given by Binary Decision Diagrams (shortly: BDDs). The main purpose of this paper is to describe this implementation and to illustrate its advantages and RelView’s extended possibilities by means of some examples. Its rest is organized as follows. Sections 2 and 3 present background material on relational algebra and BDDs. The implementation of relational algebra using BDDs 1
RelView is available free of charge via the World-WideWeb. For more details, see URL http://www.informatik.uni-kiel.de/progsys/relview.shtml
H. de Swart (Ed.): RelMiCS 2001, LNCS 2561, pp. 241–257, 2002. c Springer-Verlag Berlin Heidelberg 2002
242
R. Berghammer, B. Leoniuk, and U. Milanese
is shown in Section 4. In Section 5 we demonstrate how a relation-algebraic problem formulation combined with this implementation technique can be used to solve certain computationally hard problems for medium-sized inputs and present some experimental results. Section 6 contains some concluding remarks.
2
Relations and Relational Description of Sets
We denote the set (also: type) of all relations with domain X and range Y by [X ↔ Y ] and write R : X ↔ Y instead of R ∈ [X ↔ Y ]. If the sets X and Y are finite and of size m respectively n, then one may consider R as a Boolean matrix with m rows and n columns. Since this interpretation is well suited for many purposes and also used by RelView as one possibility to depict a relation on the screen, in the following we often use matrix terminology and matrix notation. The latter means that we write Rxy instead of x, y ∈ R. We assume the reader to be familiar with the basic operations on relations, viz. RT (transposition), R (complement), R ∪ S (join), R ∩ S (meet), R · S (composition; in the sequel abbreviated as RS), R ⊆ S (inclusion), and the special relations O (empty relation), L (universal relation), and I (identity relation). A relation R is called symmetric if R = RT and irreflexive if R ⊆ I . If X is a set, then a subset of it can be described/modeled by a vector v : X ↔ Y , a relation satisfying v = vL. Having the Boolean matrix model of relations in mind, v = vL means that v is row-constant. Consequently, for a vector the range is irrelevant. Therefore, in the following we only consider vectors v : X ↔ 1 with a specific singleton set 1 = {⊥} as range and omit the second subscript, i.e., write vx instead of vx⊥ . In the matrix model then v is a Boolean column vector and describes the subset {x ∈ X : vx } of its domain. A relation R is called univalent if RT R ⊆ I, injective if RRT ⊆ I, and total if RL = L. As usual, a univalent and total relation is a mapping. We can use injective mappings as a second way of describing subsets. Given an injective mapping R : Y ↔ X, we call Y a subset of X given by R. In this case, the vector RT L : X ↔ 1 describes Y in the above sense. Clearly, the construction of an injective mapping inj(v) : Y ↔ X from a given vector v : X ↔ 1 describing a subset Y of X is also possible. In combination with the membership relation ∈ : X ↔ 2X , such injective mappings generated by vectors can be used to enumerate sets of sets. More specifically, if v : 2X ↔ 1 describes a subset S of 2X , then inj(v) is of type [S ↔ 2X ] and – using matrix terminology – we obviously T obtain the elements of S as the columns of the composition ∈ inj(v) : X ↔ S.
3
Binary Decision Diagrams
In recent years, BDDs have emerged as an efficient data structure to manipulate very large Boolean functions. Their introduction is relatively old, but only after the work of Bryant they transformed into a useful tool. For an overview, see [7]. We will present BDDs by means of a small example. Given the ternary Boolean function f : B3 → B in disjunctive normal form f (x1 , x2 , x3 ) = (x2 ∧ x3 ) ∨ (x1 ∧ x2 ∧ x3 ) ,
(1)
Implementation of Relational Algebra Using Binary Decision Diagrams
243
its BDD is shown in Figure (a) below. A BDD is a directed acyclic graph with one root and two leaf nodes, the latter labeled with 0 and 1. Each node not being a leaf has two outgoing arcs, labeled 1 and 0 and shown in the drawings as continuous and dotted lines, respectively, oriented from top to bottom. To evaluate f for the assignment of variables x1 = 1, x2 = 0, and x3 = 1, we only have to follow the corresponding directed arcs from the root node. The first node we encounter is labeled with variable x1 , whose value is 1. Given this assignment, the 1-arc must be taken. Next, a node labeled with variable x2 is found. Since x2 = 0, the 0-arc must be taken now. Finally, the 1-arc for variable x3 reaches the 0 leaf node, which is the value of f for that assignment, i.e., of f (1, 0, 1). x1
x1
x2
x2
x3
x3
1
x3
x2
x3
0
(a) Ordered BDD
x3
1
x2
x3
0
(b) Ordered, reduced BDD
The representation of a Boolean function by means of a BDD is not unique. Figures (a) and (b) depict different BDDs representing the Boolean function f of (1). The BDD in (b) can be obtained from that in (a) by successively applying reduction rules that share isomorphic subgraphs within and eliminate superfluous nodes from the representation. Both BDDs are ordered. In an ordered BDD, all variables appear in the same order along all paths from the root to the leaves. If a BDD is ordered and reduced (no further reductions can be applied), then we have a Reduced Ordered BDD (shortly: ROBDD). Given a total ordering of variables, a ROBDD is a so-called canonical form, i.e., the representation is unique. See [7]. Figure (b) is a ROBDD with variable ordering x1 < x2 < x3 . As shown in [9], the size of a BDD can be exponential in the number of variables; however, ROBDDs are a compact representation of many Boolean functions. Furthermore, Boolean binary operations on them (like disjunction and conjunction) can be calculated in polynomial time in the size of the ROBDDs. Finally, representing Boolean functions as ROBDDs allows to implement many important tests (like implication, equivalence, and satisfiability) in constant time. Henceforth, we will implicitly assume that BDDs are reduced and ordered. Note that each ROBDD node represents, at the same time, a Boolean function whose root is the node itself. This allows to implement BDD packages which manage all BDDs using the same set of variables in only one multi-rooted graph; details can be found in [7]. For the RelView system we used the CUDD package [12] from the University of Colorado, Boulder (USA).
244
4
R. Berghammer, B. Leoniuk, and U. Milanese
Implementation of Relations by ROBDDs
r s
ROBDDs provide a very efficient way to manipulate very large Boolean functions. Therefore, in the following we will present a description of a relation (on, of course, finite sets) as such a function, too. This description is based upon binary encodings of the domain and range of the relation. We will briefly illustrate our approach by a small example. Assume two sets X = {a, b, c} and Y = {r, s} and a relation R : X ↔ Y which is depicted in RelView as the following labeled Boolean 3 × 2 matrix: a b c
Using binary encodings cX : X → B2 and cY : Y → B of X and Y , defined by cX (a) = 00, cX (b) = 01, cX (c) = 10 respectively cY (r) = 0, cY (s) = 1, the relation R can be interpreted as a ternary partial Boolean function f : B3 → B. The left of the following two tables shows this interpretation of R, where the variables x1 and x2 come from the encoding of X and the variable y1 comes from the encoding of Y . To describe R by a ROBDD, in the next step we use the truth value 0 as a dummy result and pass from the partial function f over to the total function fR : B3 → B as indicated in the right of the following tables. x1 0 0 0 0 1 1 1 1
x2 0 0 1 1 0 0 1 1
y1 f (x1 , x2 , y1 ) 1 0 0 1 0 0 1 0 0 1 1 1 0 – 1 –
x1 0 0 0 0 1 1 1 1
x2 0 0 1 1 0 0 1 1
y1 fR (x1 , x2 , y1 ) 1 0 0 1 0 0 1 0 0 1 1 1 0 0 1 0
Based on a fixed variable ordering, for example x1 < x2 < y1 , the ROBDD of fR together with a description of the sets X and Y represents the relation R completely. Omitting this additional information, we get the following ROBDD: x1 x2
x2
y1 1
0
Generally, using binary encodings cX : X → Bm and cY : Y → Bn (where m equals log |X| and n equals log |Y | ) a relation R : X ↔ Y is implemented by the two sizes |X| and |Y | and the ROBDD of the totalized Boolean function fR : Bm+n → B, where fR (x1 , . . . , xm , y1 , . . . , yn ) = 1 if and only if the decodings
Implementation of Relational Algebra Using Binary Decision Diagrams
245
−1 c−1 X (x1 , . . . , xm ) and cY (y1 , . . . , yn ) are related via R and the variable ordering is x1 < . . . < xm < y1 < . . . < yn . To get an impression of possible advantages of this ROBDD representation of relations, one has to consider at least three important aspects.
– First one has to compare the sizes of ROBDDs and of Boolean matrices representing relations typically used in practical applications of the relational framework. – In a second step, the simplicity of implementing the relational operations based on ROBDDs has to be investigated. – Finally, the efficiency of these implementations should be analysed by a number of tests and compared with other implementations. It should be emphasized that the original RelView system with the array-based implementation of relations in combination with some features of the CUDD package provided excellent support for the investigation of all of these aspects. The first aspect – the size of ROBDDs in comparison with Boolean array implementations – is demonstrated by the following examples. They deal with relations which are helpful if one attacks a computationally hard problem by first generating all candidates and then testing each to see whether it is a solution. The picture below shows the RelView representation of the membership relation ∈ : X ↔ 2X for the set X = {a, b, c, d} as a Boolean 4 × 16 matrix: a b c d
After encoding the elements of X and 2X we get the following ROBDD; in it the variables x1 and x2 encode the 4 elements of the domain X of ∈ and the variables ya , yb , yc and yd encode the 16 elements of the range 2X of ∈. x1 x2
x2
ya yb yc yd
0
1
At this point it is essential to note that in the general case each variable used to encode the range of a membership relation corresponds to exactly one row of its Boolean matrix representation as indicated in the above example by their
246
R. Berghammer, B. Leoniuk, and U. Milanese
indices. As a consequence, the number of ROBDD nodes of an arbitrary finite membership relation ∈ : X ↔ 2X is linear in the number of elements of its domain X – in contrast to the exponential size of its Boolean matrix representation. In the previously mentioned case of a membership relation on a set with 25 elements we obtain an ROBDD with about 1 KByte only. As we will see in Section 5, membership relations can help to generate all subsets of a given universum X which are candidates for the solution to a present problem. If this problem is an optimization problem and the task is to compute a subset of X with minimum or maximum size, then a relation C : 2X ↔ 2X may be very helpful which compares the sizes of sets from 2X , i.e., fulfils CU V if and only if |U | ≤ |V | for all U, V from 2X . The left of the next two pictures shows this size comparison relation C for the universum X = {a, b, c, d} used already above as a Boolean 16 × 16 matrix (produced by RelView) and the right-hand picture shows the ROBDD implementation of C. Here the ordering of the sets coincides with the ordering used in the membership example above and is determined by the columns of the Boolean matrix representation of the membership relation ∈ : X ↔ 2X . For instance, the first column of the matrix for C corresponds to the empty set, column 6 corresponds to the set {b, d}, and the last column corresponds to the entire set X. X1
x1
X2
x2
x2
X3
x3
x3
x3
X4
x4
x4
x4
x4
X1’
y1
y1
y1
X2’
y2
y2
y2
X3’
y3
y3
X4’
y4
1
y1
0
It is noteworthy, that in the general case the size of the Boolean matrix for a size comparison relation C : 2X ↔ 2X is exponential in |X|, but the number of ROBDD nodes is polynomial in the size of X. In the following, we present algorithms which compute the ROBDDs representing the membership relation ∈ : X ↔ 2X and the size comparison relation C : 2X ↔ 2X , respectively. The input n of both algorithms is the size of the universum X. We start with the algorithm MemberBdd for computing the ROBDD of the membership relation. In its subsequent description, the call IthVar(x) returns a 3-node ROBDD representing the Boolean variable x and the call Ite(p,t1 ,t2 ) implements the conditional expression if p then t1 else t2 . MemberBdd(n) 1 index := log n ; 2 for i := 0 to n − 1 do
Implementation of Relational Algebra Using Binary Decision Diagrams
247
3 res[i] := IthVar(index + i) od; 4 top nodes := n; 5 while index > 0 do 6 node := IthVar(index − 1); 7 i := 0; j := 0; 8 while i < top nodes − 1 do 9 res[j] := Ite(node, res[i + 1], res[i]); 10 i := i + 2; j := j + 1 od; 11 if Odd(top nodes) then res[j] := Ite(node, zero, res[i]); 12 j := j + 1; top nodes := j fi; 13 index := index − 1 od; 14 return res The algorithm MemberBdd creates a ROBDD, which can be subdivided into two parts as seen in the picture above for X = {a, b, c, d}. By the for-loop (l. 2-3) all nodes of the lower part are created. These nodes are used in the while-loop for building the upper part in a bottom-up manner. In every run through this loop, top nodes contains the number of nodes which have to be put together, and index is the index of the level of nodes which are generated. During the inner while-loop (l. 8-10) in each case a new node is produced, the children of which are two nodes from the level below. If top nodes is odd, then the last node from the level below gets combined with the Boolean constant zero (l. 11-12). The ROBDD for the size comparison relation C : 2X ↔ 2X with n = |X| is also built from the bottom to the top; here is the corresponding algorithm: SizecompBdd(n) 1 index := 2 ∗ n − 1; 2 high[0] := IthVar(index); 3 for index := 2 ∗ n − 2 downto n do 4 node := IthVar(index); 5 low [0] := Ite(node, high[0], zero); 6 for i = 1 to 2 ∗ n − index − 2 do 7 low [i] := Ite(node, high[i], high[i − 1]) od; 8 i := 2 ∗ n − index − 1; 9 low [i] := Ite(node, one, high[i − 1]); 10 high := low od; 11 for index = n − 1 downto 0 do 12 node := IthVar(index); 13 for i = 0 to index do 14 if i = n − 1 then low [i] := Ite(node, high[i], high[i + 1]) 15 else low [i] := Ite(node, high[i], one) fi od; 16 high := low od; 17 return low[0] As this program shows, in a first step the single node in the last row of variables is produced (l. 2); in the picture above this node carries the label xd . The variable index always contains the index of the nodes which are generated in the following
248
R. Berghammer, B. Leoniuk, and U. Milanese
steps. In lines 3-10 the lower triangle of the ROBDD is created. Line 5 shows the construction of the node on the very left in each row and in line 9 the node on the row’s very right is built. The code between these lines shows the creation of the nodes in between. The upper triangle of the ROBDD is constructed in lines 11-16. Only the node on the very left in the lowest row needs a special treatment (l. 14-15), all other nodes are simply built using the Ite operator. Having described the construction of two important relational constants represented by ROBDDs, now we deal with the implementation of relational operations. We consider again two examples. Often it is possible to lead the ROBDD-based implementation of a relational operation back to the operation’s component-wise definition. An example of this is the composition of two relations. Let X, Y and Z be finite sets with |X| = n, |Y | = m and |Z| = k, and let R : X ↔ Y and S : Y ↔ Z be relations. Then, the composition RS : X ↔ Z of R and S is component-wise defined by (RS)xz ⇐⇒ ∃ y (Rxy ∧ Syz ) for all x from X and z from Z. In the following, we use the three abbreviations ln = log n , lm = log m , and lk = log k . Using, furthermore, binary encodings cX : X → Bln , cY : Y → Blm and cZ : Z → Blk for X, Y , and Z, we obtain, for the representation of R respectively S, the Boolean functions fR : Bln +lm → B and fS : Blm +lk → B such that fR (r1 , . . . , rln , rln +1 , . . . , rln +lm ) = 1 if and only if c−1 X (r1 , . . . , rln ) and −1 cY (rln +1 , . . . , rln +lm ) are related via R and fS (s1 , . . . , slm , slm +1 , . . . , slm +lk ) = −1 1 if and only if c−1 Y (s1 , . . . , slm ) and cZ (slm +1 , . . . , slm +lk ) are related via S. If we can identify the variables rln +1 , . . . , rln +lm with s1 , . . . , slm , hence, the Boolean function for the composition of RS could be computed by fRS = ∃s1 . . . ∃ slm (fR ∧ fS ) using conjunction and repeated existential quantification of Boolean functions, where the latter operation is defined by (∃ xi g)(x1 , . . . , xn ) = 1 if and only if g(x1 , . . . , xi−1 , 1, xi+1 , . . . , xn ) = 1 or g(x1 , . . . , xi−1 , 0, xi+1 , . . . , xn ) = 1. The operations ∧ and ∃ are available in all ROBDD packages we know, so we only have to take care of the identity of the variables. Now, one could assume that for the representation of the relations the variables could be chosen in such a way, that the indices of these shared variables are always convenient for composition. Unfortunately, this is not always possible. For example, for a relation R of specific type [X ↔ X], a so-called homogeneous relation, we want to be able to calculate RR. Here we have the simple case that the variables for the representation of the range of R in no way could be the same as those used for its domain, except for the special case that X contains only one element. Therefore we have decided to encode relations by starting with index 0 and number the following variables consecutively. This leads to a fixed sequence x of variables, x0 , x1 , x2 , . . . say, and allows the handling of any arbitrary relation over finite sets using a certain prefix of x. Furthermore, the construction of
Implementation of Relational Algebra Using Binary Decision Diagrams
249
the relational constants described above concurs with this fixed sequence x of variables. For the relations R and S mentioned above – respectively the Boolean functions fR and fS representing them – this approach means, that variables r1 , . . . , rln are equal to x0 , . . . , xln −1 , variables rln +1 , . . . , rln +lm are identical with xln , . . . , xln +lm −1 , variables s1 , . . . , slm are the same as x0 , . . . , xlm −1 , and variables slm +1 , . . . , slm +lk coincide with xlm , . . . , xlm +lk −1 . Unfortunately, this fixed encoding causes a problem for the composition RS, because now we have to ensure that the variables representing the range of R are the same as those used for the representation of the domain of S. Let N denote ln + lm + lk . Then, this problem can be solved by introducing new Boolean functions fR and fS , which are defined as follows: fR (x0 , . . . , xln +lm −1 , . . . , xN −1 ) fS (xlm +lk , . . . , xN −1 , x0 , . . . , xlm +lk −1 )
= 1 ⇐⇒ fR (x0 , . . . , xln +lm −1 ) = 1 = 1 ⇐⇒ fS (x0 , . . . , xlm +lk −1 ) = 1 .
The computation of these Boolean functions is mainly a renumbering of the indices of the variables. But, because ROBDDs are strongly connected structures with shared nodes, we cannot simply change the variable’s indices directly within the ROBDD nodes. Indeed we have to build up a new ROBDD for the Boolean functions fR and fS . If we take a look onto the function fR , it is obviously, that it is independent of the last lk parameters. Because such variables do not require any nodes in an ROBDD, it is clear that the ROBDDs for fR and fR are equal. This is not the case with fS : This Boolean function also is independent of some parameters (the first ln ones), but the needed variables of fS have got other indices as the ones of fS . In fact, every index is incremented by exactly ln . For the computation of fS we have implemented a new ROBDD algorithm, which we called Shift. Here is its formulation in the program notation used so far: Shift(F, border, n, m) 1 if F = one or F = zero 2 then res := F 3 else index := F → index; 4 E := ElseBdd(F); 5 T := ThenBdd(F); 6 SE := Shift(E, border, n, m); 7 ST := Shift(T, border, n, m); 8 if index > border 9 then node := IthVar(index+n) 10 else node := IthVar(index+m) fi; 11 res := Ite(node, ST, SE) fi; 12 return res The first argument F of this algorithm denotes the ROBDD, whose variable indices we want to renumber, the second argument border is an index which divides the indices of the variables of F into two sets, the next argument n is the increasement value for the first set of variables, whose indices are less than border, and the last argument m is the increasement value for the second set
250
R. Berghammer, B. Leoniuk, and U. Milanese
of variables. If F is a constant, then F has got no index, which can be shifted. Otherwise the index of the root of F is stored into index. The call ElseBdd(F) returns the ROBDD for F ∧ index and the call ThenBdd(F) returns F ∧ index. We store these results into new variables T and E and start this algorithm recursively on these ROBDDs. In the next step (l. 8-11) we test whether index is situated in the first set of variables or in the second, and construct the result by using the Ite operator. If one stores additionally already computed results in a table, then it is obviously possible to call algorithm Shift exactly once for every node of F . After these preparations, we are able to compute the ROBDD FS representing fS by a call Shift(FS , lm , ln , ln ), where FS is the ROBDD of fS . We use lm as the border, because all variables, whose index is less than lm , belong to the encoding of the domain of S, and the rest are used for the range of S. This call creates a new ROBDD, which has the same structure as FS , but all indices are incremented by ln . As one can see further, fR and fS have suitable variable indices for the conjunction. Hence, we are allowed to define = ∃ xln . . . ∃ xln +lm −1 (fR ∧ fS ) fRS
and, based on this, finally the Boolean function representation fRS for the composition RS by (x0 , . . . , xN −1 ) = 1 . fRS (x0 , . . . , xln −1 , xln +lm , . . . , xN −1 ) = 1 ⇐⇒ fRS is independent of the variables in It is obvious, that the Boolean function fRS the range from xln to xln +lm −1 . Because of the convention, that all variables are numbered consecutively as x0 , x1 , x2 , . . ., we must shift the indices of the variables representing the range of the result into the correct values. This can easily be done by storing the result of the call Shift(FRS , ln , 0, −lm ) into FRS with FRS and FRS representing fRS and fRS . Altogether, this leads to the following algorithm Compose for the composition of two relations R and S which are implemented as ROBDDs FR and FS :
Compose(FR , FS , ln , lm ) 1 res := Shift(FS , ln , lm , lm ); 2 res := And(FR , res); 3 res := Exist(ln , . . . , ln + lm − 1, res); 4 res := Shift(res, ln , 0, −lm ); 5 return res The above algorithm Shift is also useful for the computation of the transposition of a relation R : X ↔ Y which is implemented as a ROBDD. Assume n = |X| and m = |Y |. We start with the component-wise definition T ⇐⇒ Ryx Rxy
for all x from X and y from Y . If we use the abbreviations ln = log n and lm = log m , then it will immediately lead to the equivalence fRT (xln , . . . , xln +lm −1 , x0 , . . . , xln −1 ) = 1 ⇐⇒ fR (x0 , . . . , xln +lm −1 ) = 1 .
Implementation of Relational Algebra Using Binary Decision Diagrams
251
For the ROBDD representation of the Boolean function fRT this means that we only have to exchange the variables used for the encoding of the domain of R with those used for the representation of for its range. This can easily be obtained by a call Shift(FR , ln , lm , −ln ). Within the RelView system, however, we use the following refined version: Transpose(FR , ln , lm ) 1 if ln = 0 and lm = 0 2 then res := Shift(FR , ln , lm , −ln ) 3 else res := FR fi; 4 return res It takes advantage of the fact that ROBDD representations of relations with domain or range 1, especially vectors v : X ↔ 1 which frequently occur in relationalgebraic specifications and programs, do not use variables for the singleton set 1, so that a shifting of variables is not necessary in these cases. Besides the examples presented in this section, we have developed ROBDDbased algorithms for all the constants and basic operations on relations mentioned in Section 2 and many further constants, operations, and tests typically used in relation-algebraic specifications and programs. Examples of the latter are residuals, symmetric quotients, and constants and operations appearing in relational descriptions of domains (like the two projections π1 : X × Y ↔ X and π2 : X × Y ↔ Y in the case of direct products), which have already been available in the former, array-based version of RelView (see [1] for details). But we have also extended the system by some new operations like the filtering of subsets with a certain size from a vector v : 2X ↔ 1, the efficiency of which mainly bases on the new ROBDD implementation. Finally, we have developed an algorithm for the random generation of relations which are implemented as ROBDDs. Among other things, randomly generated relations are very profitable in specification and program testing. Detailed descriptions of all these algorithms, including the correctness proofs, are presented resp. will be presented in the Ph.D theses [8] and [10].
5
Some Applications
To analyze the efficiency of the ROBDD implementation of relations, we have carried out a lot of experiments for different kinds of problems and with different kinds of randomly generated relations. We have tested relation-algebraic specifications and programs using the new and the former version of the RelView system and have compared both performances. The new ROBDD-implementation proved to be by far superior to the older one, especially if the generate-and-test approach is used to solve computationally hard problems. This is mainly due to the fact that for many such problems the very efficient ROBDD-representation of certain important relations on powersets (membership, inclusion, size comparison, filtering) allows to handle the combinatorial explosion even in the case of medium-sized input relations. In the following, we present three examples
252
R. Berghammer, B. Leoniuk, and U. Milanese
from graph-theory and an application for 0/1 matrices. Further examples and the evaluation of the corresponding experiments can be found in [8]. First, we assume a directed graph g = (X, R) with the adjacency relation R : X ↔ X on the node set X containing the arcs of g. A set A of nodes is said to be absorbant in g if from every node outside of A there is at least one arc leading into A , i.e., if the formula ∀ x (x ∈ / A → ∃ y (y ∈ A ∧ Rxy )) holds. Furthermore, a set S of nodes is called stable in g if no two nodes of S are related via R. This situation is characterized by the formula ∀ x (x ∈ S → ∀ y (y ∈ S → R xy ). Finally, a kernel of g is a set of nodes which is at the same time absorbant and stable. It is an easy exercise to express absorption and stability in terms of relational operations, the universal vector L : X ↔ 1, and the membership relation ∈ : X ↔ 2X ; see e.g., [1]. This leads to the relation-algebraic specifications L∈ ∪ R ∈
T
: 2X ↔ 1 and L(∈ ∩ R ∈)
T
: 2X ↔ 1
of the vectors describing the set of absorbant sets respectively the set of stable sets of g. As a consequence, a relation-algebraic specification of the vector kernel (R) : 2X ↔ 1 describing the set of kernels of g is kernel (R) = L ∈ ∪ R ∈
T
T
∩ L(∈ ∩ R ∈) .
(2)
Now, assume g = (X, E) to be an undirected graph, i.e., the edge set E to consist of sets {x, y} of distinct nodes x and y. A set C of nodes is a clique of g if each set {x, y} of distinct nodes x and y from C forms an edge. If we define the adjacency relation R : X ↔ X of g by Rxy if and only if {x, y} is an edge, then C is a clique of g if and only if it is stable in the so-called directed variant g = (X, R ∩ I ) of the complement graph of g. The relation R ∩ I of g is symmetric and irreflexive. Hence, the inclusion-maximal stable sets of g are precisely its kernels; see e.g., [11]. Using (2), we consequently obtain the vector maximalclique(R) : 2X ↔ 1 describing the inclusion-maximal cliques of g as maximalclique(R) = kernel ( R ∩ I ) .
(3)
In our third example we assume g = (X, E) again to be an undirected graph. Then we can associate with g its incidence relation R : E ↔ X, defined by Rex if and only if node x is contained in edge e. The relation RRT ∩ I : E ↔ E is called the edge adjacency relation of g and coincides with the adjacency relation of the line graph L(g) of g, in which the edges of g are the nodes and two nodes e, f of L(g) form an edge in L(g) if and only if e = f and e ∩ f = ∅. A matching of g is a set M of edges such that no two distinct edges have a node in common. Using the directed variant g = (E, RRT ∩ I ) of the line graph L(g), it is easy to enumerate all inclusion-maximal and maximum-size matchings of g. From the
Implementation of Relational Algebra Using Binary Decision Diagrams
253
definition we immediately obtain that M is a matching of g if and only if it is a stable set in g . Since RRT ∩ I is again a symmetric and irreflexive relation, we can use specification (2) to compute the vector maximalmatch(R) : 2E ↔ 1 describing the set of inclusion-maximal matchings of g as maximalmatch(R) = kernel (RRT ∩ I ) .
(4)
Applying the relation-algebraic specification greatest(S, v) = v ∩ S T v of the greatest elements of the set described by the vector v wrt. the quasi order S (see [11] for example) in combination with specification (4), we get maximummatch(R) = greatest(C, maximalmatch(R))
(5)
for the vector maximummatch(R) : 2E ↔ 1 describing the set of maximum-size matchings of g. In this relation-algebraic specification C : 2E ↔ 2E is the size comparison relation as introduced in Section 4. As already mentioned, RelView allows to generate random relations, where it is additionally possible to specify the percentage of their entries. Using this feature, a lot of experiments have been carried out with random graphs of various sizes and with various percentages of arcs respectively edges on a Sun Ultra 80 workstation running Solaris 7 at 450 MHz and with 4 Gbyte of main memory. The following table shows some experimental results of the enumeration of cliques for randomly generated symmetric and irreflexive adjacency relations on n elements, which contain approx. 10% respectively 20% and 30% of all entries. These adjacency relations correspond to undirected graphs with n nodes and the same percentages of edges. nodes 50 100 incl-m. 10% 135 751 m.-size 10% 9/4 5/5 time (sec.) 0.02 1.2 incl-m. 20% 150 791 m.-size 20% 10/4 2/5 time (sec.) 0.15 1.4 incl-m. 30% 233 1665 m.-size 30% 10/5 5/6 time (sec.) 0.28 4.1
150 200 250 300 350 400 450 2253 4833 8966 15701 25196 38858 56898 36/5 1/6 7/6 14/6 48/6 2/7 124/6 8 31 87 233 628 2437 6145 2243 4820 8884 15631 25255 38985 – 26/5 2/6 5/6 7/6 2/7 107/6 – 9 32 90 248 622 2361 – 5581 14392 32181 62311 – – – 1/7 246/6 28/7 89/7 182/7 – – 26 171 784 3056 36282 – –
Lines 2, 5, 8 show the number of inclusion-maximal cliques, lines 3, 6, 9 the number of maximum-size cliques and their sizes (which can be obtained from the column-wise representation of sets described in Section 2), and lines 4, 7, 10 the time to compute the latter results. The former array-based version of RelView was able to handle only instances with at most 25 nodes, because for |X| = 25 the computation of the membership relation ∈: X ↔ 2X required more than 100 MByte of memory space. Therefore, it is not possible to compare the results of the above table with the former version. Of course, we have compared both versions also in the
254
R. Berghammer, B. Leoniuk, and U. Milanese
case of computations not using membership relations and similar constructions. Again the new version proved to be superior. See [8] fore details. We close this section with two concrete applications of the RelView system. First, we demonstrate how to compute the permanent n per(A) = π i=1 ai,π(i) of a n × n 0/1-matrix, where the sum is over all permutations π on the set {1, . . . , n}. The computation of permanents bases on the fact that per(A) equals the number of perfect matchings of the bipartite undirected graph gA = (X, E), where X = {1, . . . , 2 ∗ n} and {i, j} ∈ E if and only if 1 ≤ i ≤ n < j ≤ 2 ∗ n and ai,j−n = 1. (A matching M is said to be perfect if every node is incident with at least one edge from M .) Now, we interpret the following 15 × 15 Boolean RelView-matrix as 0/1-matrix A:
Then the undirected bipartite graph gA = (X, E) has 30 nodes and 79 edges. On the screen of RelView this graph is depicted as in the following picture: 1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
The line graph L(gA ) of gA has even 79 nodes and 396 edges. This graph is too large to be reproduced here. Using the relation-algebraic specification (5), on the Sun workstation mentioned above RelView needs 541 seconds to compute a vector v : 2E ↔ 1 which describes the 32844 greatest stable sets of L(gA ), i.e., the 32844 maximum-size matchings of gA . To test that one of these matchings consists of 15 edges can be done very fast. One only has to select a so-called point p : 2E ↔ 1 contained in v (see e.g., [2] for details), which describes a single maximum-size matching of gA as an element of the powerset 2E , and then to transform it into the vector ∈ p : E ↔ 1 which describes the same matching.
Implementation of Relational Algebra Using Binary Decision Diagrams
255
Hence, we get per(A) = 32844. Here it should be mentioned that even 89 seconds suffice to compute this result if one directly works with the relation-algebraic specification of stable sets and the filter operation mentioned at the end of Section 4. More details on the relation-algebraic computation of permanents can be found in [8]. As the second application we consider two queens problems on a chess board. Consider the following RelView picture:
If we describe the possible moves of a queen on a chess board as a relation R : X ↔ X on a set X representing the 64 squares of the chess board, then this Boolean matrix exactly represents the relation R. Using the RelView system, it is very easy to check that the directed graph g = (X, R) has exactly 10118 kernels, of which 92 have the maximal size 8. The latter kernels correspond exactly to the 92 solutions of the well-known 8-queens problem. I.e., each kernel represents a set of 8 queens which are placed on the chess board in such a way that no two of them are enemies. Kernels of size 8 are greatest stable sets of the directed graph g = (X, R). We have also considered the dual problem dealing with smallest absorbant sets of g. It corresponds to the following question: How can the entire chess board be checked by as few queens as possible and how many solutions exist? It is wellknown that 5 queens suffice; therefore the problem is also called the 5-queens problem. RelView needs some seconds to compute the 4860 solutions to this problem. One of them is shown in the following picture:
256
R. Berghammer, B. Leoniuk, and U. Milanese
In [8] one finds also further solutions to these two problems on chess boards of other sizes than 8 × 8 computed by RelView. For example, 5 queens also suffice to check the entire 11 × 11 chess board. But for this board size there exist only 2 solutions.
6
Conclusion
We have shown how relational algebra can be implemented by using ROBDDs. This implementation is used in the new version of RelView and we have demonstrated the power of this tool in combination with relational formulations by solving some computationally hard graph-theoretic problems for medium-sized graphs. We have applied the RelView system to further hard problems from graph-theory and other domains (like lattice theory, games, Condition/Event Petri nets, automata theory), too, and in all these cases we have had very positive experience. Concerning the system, presently we work on the fine-tuning of the ROBDDimplementation of relations since some operations (e.g., the computation of inj(v) and some operations of relational domains) are not always as efficient as they should be. We also plan to extend the language of the system, for instance, by an assert-command to allow flexible invariance tests. Another future extension of the system concerns the improvement of RelView’s debugging and visualization features. But, of course, we are also interested in further applications. Among other things, our current investigations concern the use of RelView in software architecture and re-engineering.
References 1. Behnke R., Berghammer R., Schneider P.: Machine support of relational computations: The Kiel RelView system. Bericht Nr. 9711, Inst. f¨ ur Inf. und Prak. Math., Univ. Kiel (1997) 2. Behnke R., Berghammer R., Hoffmann T., Leoniuk B., Schneider P.: Applications of the RelView system. In: Berghammer R., Lakhnech Y. (eds.): Tool support for system specification, development and verification. Advances in Computing, Springer, 33-47 (1999)
Implementation of Relational Algebra Using Binary Decision Diagrams
257
3. Behnke R., Berghammer R., Meyer E., Schneider P.: RelView – A system for calculating with relations and relational programming. In: Astesiano E. (ed.): Proc. Conf. on Fundamental Approaches to Software Engineering (FASE ’98), LNCS 1382, Springer, 318-321 (1998) 4. Berghammer R., Hoffmann T.: Modelling sequences within the RelView system. J. of Univ. Comp. Sci. 7, 107-123 (2001) 5. Berghammer R., Karger B. von, Ulke C.: Relation-algebraic analysis of Petri nets with RelView. In: Margaria T., Steffen B. (eds.): Proc. Workshop on Tools and Applications for the Construction and Analysis of Systems (TACAS ’96), LNCS 1055, Springer, 49-69 (1996) 6. Brink C., Kahl W., Schmidt G. (eds.): Relational methods in computer science. Advances in Comp. Sci., Springer (1997) 7. Bryant R.E.: Symbolic Boolean manipulation with ordered binary decision diagrams. ACM Comp. Surv. 24, 293-318 (1992) 8. Leoniuk B.: ROBDD-based implementation of relational algebra with applications (in German). Ph.D. thesis, Inst. f¨ ur Inf. und Prak. Math., Univ. Kiel (2001) 9. Liaw H.-T., Lin C.-S.: On the OBDD representation of general Boolean functions. IEEE Transactions on Comp. 41, 661-664 (1992) 10. Milanese U.: On the implementation of a ROBDD-based tool for the manipulation and visualization of relations (in German). Ph.D. thesis, Inst. f¨ ur Inf. und Prak. Math., Univ. Kiel (to appear) 11. Schmidt G., Str¨ ohlein T.: Relations and graphs. Discrete Mathematics for Computer Scientists, EATCS Monographs on Theor. Comp. Sci., Springer (1993) 12. Somenzi, F.: CUDD: CU decision diagram package; Release 2.3.0. http://vlsi. colorado.edu/˜fabio/CUDD/cuddIntro.html (1998)
Calculating a Relational Program for Transitive Reductions of Strongly Connected Graphs Rudolf Berghammer and Thorsten Hoffmann Institut f¨ ur Informatik und Praktische Mathematik Christian-Albrechts-Universit¨ at Kiel Olshausenstraße 40, D-24098 Kiel
Abstract. Based on a generic program for computing minimal sets, we present a simple relational program for computing transitive reductions of strongly connected directed graphs. It uses a precomputation phase and can easily be implemented in quadratic running time. The presentation is done as an exercise in generic programming and for combining the Dijkstra-Gries method and relational algebra to derive graph algorithms.
1
Introduction
Since many years it is widely accepted that the traditional ad-hoc approach to program construction does not lead to reliable software and, therefore, a more systematic and formal approach is required. One such approach is program derivation which means that a program is calculated from a formal specification. In the case of imperative programs, specifications consist of preconditions and postconditions and program derivations are mainly based on loop invariants. The latter are usually hypothesized as modifications of postconditions and programs are constructed in such a way that invariants are established and maintained. Based on work of Floyd and Hoare, the use of such techniques started with Dijkstra; see his textbook [6]. Later, especially Gries elaborated Dijkstra’s ideas and formulated the fundamental principle that “a program and its correctness proof should be developed hand-in-hand with the proof usually leading the way” (cf. [7], p. 164). Therefore, nowadays one often speaks of the Dijkstra-Gries program development method. To support formal program derivation, adequate algebraic frameworks are very helpful. For problems on graphs – which constitute a very important domain in algorithmics, see [1,5] just to cite two well-known textbooks on this topic – relational algebra plays a prominent role. Concerning the modeling side, this is due to the fact that relations and graphs are essentially the same and there are many simple and elegant ways to describe fundamental graph-theoretic objects and properties with relation-algebraic means. With regard to proofs, the use of relational algebra has the advantage to reduce the danger of doing wrong proof steps drastically, to clarify the proof structure frequently, and to open the possibility for mechanical support. Finally, the relation-algebraic framework also supports prototyping and validation tasks in a significant manner. In this paper we want to illustrate the formal derivation of a relational program by combining ideas from generic programming, the Dijkstra-Gries program H. de Swart (Ed.): RelMiCS 2001, LNCS 2561, pp. 258–275, 2002. c Springer-Verlag Berlin Heidelberg 2002
Calculating a Relational Program for Transitive Reductions
259
development method, graph searching techniques, and relational algebra. We consider the computation of a transitive reduction of a strongly connected directed graph, a problem which is, for instance, discussed in [11,14]. In the second paper a quite complicated linear algorithm is presented, which, however, in the meantime has shown to be not correct; see [9]. The relational program we will develop is very simple. It can be performed with the RelView tool [2] almost as it stands and its implementation in quadratic running time in a conventional programming language is trivial. The rest of the paper is organized as follows: Section 2 contains some basic notions of relational algebra and graph theory and shows how to model certain graph-theoretic objects resp. properties with relation-algebraic means. Similar to the approach of Ravelo in [12], then in Section 3 we derive a generic imperative program that computes for a finite set a minimal subset satisfying a certain predicate. We instantiate this program in Section 4 to compute a transitive reduction of a strongly connected directed graph. In doing so, we use a combination of relational depth-first and breadth-first graph searching programs in a precomputation phase to improve the running time of the instantiation as well as the quality of the computed result in view of the number of its arcs. Section 5 contains a short conclusion.
2
Relations and Graphs
Given sets X, Y = ∅, the set of all relations with domain X and range Y is denoted by [X ↔ Y ] and we write R : X ↔ Y instead of R ∈ [X ↔ Y ]. We assume the reader is familiar with the basic operations on relations, viz. RT (transposition), R (negation), R ∪ S (union), R ∩ S (intersection), RS (composition), R ⊆ S (inclusion, subrelation test), and the special relations O (empty relation), L (universal relation), and I (identity relation). The set-theoretic operations , ∪, ∩, ⊆ and the constants O, L form a complete Boolean lattice. T Further well-known laws of relations are e.g., RT = R, Q(R ∩ S) ⊆ QR ∩ QS, T and (RS) = S T RT . The theoretical framework for such laws holding is that of a relational algebra. As constants and operations of this algebraic structure we have those of the set-theoretic relations; its axioms are those of a complete Boolean lattice for , ∪, ∩, ⊆, O, L, those of a monoid for composition and I, the equivalence of QT S ⊆ R , QR ⊆ S and S RT ⊆ Q (in [13] called Schr¨ oder equivalences), and the equivalence of R = O and LRL = L (in [13] called Tarski rule 1 ). In the proofs we shall mention only the latter two axioms and their “non-obvious” consequences. Well-known laws like those presented above or in Sections 2.1 until 2.3 of [13] remain unmentioned. A relation R, for which domain and range coincide, is called reflexive if I ⊆ R and transitive if RR ⊆ R. The least transitive relation containing R is its transitive closure R+ = i≥1 Ri and the least reflexive and transitive relation 1
Since we include the Tarski rule we deal with relational algebras which nowadays normally are refered to as “simple”.
260
R. Berghammer and T. Hoffmann
containing R is its reflexive-transitive closure R∗ = I ∪ R+ = i≥0 Ri , where R0 = I and Ri+1 = RRi for all i ∈ N. Later on, we will apply the implication SLS ⊆ S
=⇒ (R ∪ S)∗ = R∗ ∪ R∗ SR∗ ,
(1)
which is an immediate consequence of the well-known star-decomposition rule (Q ∪ R)∗ = (Q∗ R)∗ Q∗ , and the equation R∗ R∗ = R∗ . A (now arbitrary) relation R is said to be univalent (or a partial function) if RT R ⊆ I and total if RL = L, which again is equivalent to I ⊆ RRT . R is called injective if RT is univalent and surjective if RT is total. Since we want to solve a concrete graph-theoretic problem and use relational algebra only as an adequate framework for this task, we work within its standard model of set-theoretic relations. Now let’s make the transition to graphs, where we likewise assume the reader is familiar with the basic notations. Directed graphs and relations are essentially the same since the relation R : V ↔ V of a directed graph g = (V, R) with vertex set V = ∅ describes its arc set. This implies immediately that each set of arcs of g is a subrelation of R. If we identify a singleton set with its only element, we can also model a single arc of g by a specific subrelation of R. An exact characterization is given below. To reason relation-algebraically also about subsets and elements of the vertex set of g, we have to encode these objects as relations, too. Subsets of V can be modeled by vectors. A vector is a relation v satisfying v = vL (one conventionally uses lowercase letters in this context). This definition implies that for describing vectors the range is irrelevant. Therefore, one usually uses a singleton set 1 such that v : V ↔ 1 describes the set {x ∈ V : x, ⊥ ∈ v}, where ⊥ is the element of 1. To give an example, if v : V ↔ 1 describes the subset X of V , then RT v : V ↔ 1 describes the set of all successors of vertices of X and hence RT L : V ↔ 1 describes the set of all endpoints of arcs of g. A vector is said to be a point if it is injective and surjective. For p : V ↔ 1 these properties mean that it describes a singleton set, i.e., a vertex of g if we identify {x} with x. Relational algebra is very helpful for describing many graph-theoretic properties. Let again g = (V, R) be a directed graph. In this paper we shall use the following relation-algebraic characterizations: The graph g is strongly connected if and only if R∗ = L and circuit-free if and only if R+ ⊆ I . A point r : V ↔ 1 is said to be a root of g if and only if rL ⊆ R∗ . This property means that from the vertex described by r each vertex of g can be reached via a path. Hence, g is a (directed) tree (also: arborescence) with root r if and only if R+ ⊆ I , RRT ⊆ I, and rL ⊆ R∗ . Finally, g is a subgraph of a directed graph h = (W, S) if and only if V = W and R ⊆ S. In the standard model of relational algebra the so-called point axiom of [13] holds. It says that for every R = O there exist points p, q such that pq T ⊆ R. As a consequence, for each vector v = O there exists a point p fulfilling p ⊆ v. The choice of an atom atom(R) (i.e., a relation of the form pq T ) contained in R = O and a point p contained in a vector v = O are fundamental for a relational approach to graph algorithms since they correspond to the choice of an arc from a set of arcs resp. a vertex from a set of vertices. Our demands on atom(R) are: atom(R) ⊆ R
T
atom(R)L and atom(R) L are points.
(2)
Calculating a Relational Program for Transitive Reductions
261
From the axioms (2) we get the inclusion (a proof can be found in [4]) atom(R)Latom(R) ⊆ atom(R) .
(3)
As relation-algebraic axioms for point(v) we demand the following properties: point(v) ⊆ v
point(v) is a point.
(4)
Note that atom : [X ↔ X] → [X ↔ X] and point : [X ↔ 1] → [X ↔ 1] are (deterministic, partial) functions in the usual mathematical sense. Each call yields the same object so that e.g., atom(R) = atom(R) holds. Of course, the above axiomatizations (2) and (4) allow different realizations. For instance, the specific implementation of atom in RelView uses that the system deals only with finite sets which are totally ordered by an internal enumeration. A call atom(R) then yields the relation { x, y} which consists of the lexicographically smallest pair x, y of R.
3
A Generic Program for Computing Minimal Subsets
For the following we assume a finite set M , a predicate P on the powerset 2M of M , and an element R from 2M . Furthermore we suppose that P is upwardsclosed , which means that for all X, Y from 2M if X is a subset of Y and P(X) is true, then P(Y ) is also true. Note that this property implies the negation ¬P of P to be downwards-closed , i.e., if Y ⊇ X and P(Y ) is false, then P(X) is false. In this section we shall systematically derive a generic imperative program that computes – for R as its input – a minimal subset of R that satisfies the predicate P. If we use the variable A as the program’s output, then a formal specification of the requirements on A is given by the postcondition ∧
post(R, A) = P(A) ∧ A ⊆ R ∧ ∀ X ∈ 2A : P(X) → X = A . For the derivation of the program we combine the Dijkstra-Gries program development method with set theory and logic. In doing so, for a set X from 2M and an element x from M we write X \ x instead of X \ {x} and X ∪ x instead of X ∪{x} to enhance readability. Heading towards a program with an initialization followed by a while-loop, the derivation is carried out in three parts. First we have to develop a loop invariant. Here we follow the most commonly used technique of generalizing the postcondition (for details see [7]). The corresponding calculation introduces a new variable B and looks as follows: ⇐⇒ ⇐⇒ ⇐⇒ ⇐=
post(R, A) P(A) ∧ A ⊆ R P(A) ∧ A ⊆ R P(A) ∧ A ⊆ R P(A) ∧ A ⊆ R
∧ ∧ ∧ ∧
∀ X ∈ 2A : P(X) → X = A ∀ X ∈ 2A : X = A → ¬P(X) ∀ x ∈ A : ¬P(A \ x) B = ∅ ∧ B ⊆ A ∧ ∀ x ∈ A \ B : ¬P(A \ x) .
Only the direction “⇐=” of the third step is non-trivial and needs an additional explanation: If X ⊂ A, then there exists an element x from A such that X ⊆ A\x
262
R. Berghammer and T. Hoffmann
and hence ¬P(A \ x) implies ¬P(X) since ¬P is downwards-closed. Guided by the above implication, now we choose B = ∅ as exit condition of the while-loop and the following loop invariant: ∧
inv (R, A, B) = P(A) ∧ A ⊆ R ∧ B ⊆ A ∧ ∀ x ∈ A \ B : ¬P(A \ x) . In the second part of the derivation, we now have to consider the initialization of the variables A and B. We do this in combination with the choice of a suitable precondition, i.e., a property on R which is sufficiently general and implies additionally that the initialization establishes the invariant. Since we want to compute a minimal subset of R satisfying P, it seems to be reasonable to require that there exists a subset X of R such that P(X) is true. Due to the upwards-closedness of P, however, the existence of X is equivalent to the fact that the program’s input R satisfies P. Hence, we choose ∧
pre(R) = P(R) as precondition. This works and yields an initialization which assigns R to both variables A and B. Here is the formal justification: pre(R) ⇐⇒ P(R) ∧ R ⊆ R ∧ ∀ x ∈ ∅ : ¬P(R \ x) ⇐⇒ P(R) ∧ R ⊆ R ∧ R ⊆ R ∧ ∀ x ∈ R \ R : ¬P(R \ x) ⇐⇒ inv (R, R, R) . Having achieved an initialization, in the last part of the program derivation we have to elaborate the body of the loop in such a way that its execution maintains the invariant and ensures termination of the loop. To this end, we suppose that B is non-empty and b is an arbitrary element contained in it. We consider two cases. First we assume P(A \ b) to be true. Then we have: ⇐⇒ =⇒ =⇒ =⇒ ⇐⇒ ⇐⇒
inv (R, A, B) P(A) ∧ A ⊆ R ∧ B ⊆ A ∧ ∀ x ∈ A \ B : ¬P(A \ x) P(A) ∧ A ⊆ R ∧ B \ b ⊆ A \ b ∧ ∀ x ∈ A \ B : ¬P(A \ x) P(A \ b) ∧ A \ b ⊆ R ∧ B \ b ⊆ A \ b ∧ ∀ x ∈ A \ B : ¬P((A \ b) \ x) P(A \ b) ∧ A \ b ⊆ R ∧ B \ b ⊆ A \ b ∧ ∀ x ∈ (A \ B) \ b : ¬P((A \ b) \ x) P(A \ b) ∧ A \ b ⊆ R ∧ B \ b ⊆ A \ b ∧ ∀ x ∈ (A \ b) \ (B \ b) : ¬P((A \ b) \ x) inv (R, A \ b, B \ b) .
This calculation uses in the third step that P(A \ b) is true and the predicate ¬P is downwards-closed. In the remaining steps only basic laws of set theory and logic and the definition of the invariant are applied. Now we deal with the remaining case of P(A \ b) being false. Here we obtain
Calculating a Relational Program for Transitive Reductions
⇐⇒ =⇒ ⇐⇒ ⇐⇒ ⇐⇒
inv (R, A, B) P(A) ∧ A ⊆ R ∧ P(A) ∧ A ⊆ R ∧ P(A) ∧ A ⊆ R ∧ P(A) ∧ A ⊆ R ∧ inv (R, A, B \ b) ,
263
B ⊆ A ∧ ∀ x ∈ A \ B : ¬P(A \ x) B \ b ⊆ A ∧ ∀ x ∈ A \ B : ¬P(A \ x) B \ b ⊆ A ∧ ∀ x ∈ (A \ B) ∪ b : ¬P(A \ x) B \ b ⊆ A ∧ ∀ x ∈ A \ (B \ b) : ¬P(A \ x)
where again the assumption of the case, i.e., that P(A \ b) is false, is used in the derivation’s third step and basic laws of set theory and logic in combination with the definition of the invariant are applied otherwise. In view of the above implications we have to change in each run of the whileloop the value of B into B \ b and, provided P(A \ b) is true, the value of A into A\b. This can be achieved by a conditional and leads – in a syntax which is quite similar to that of the RelView system – to the following program Minimum. In this program it is assumed that the call elem(B) of the pre-defined operation elem yields an element of the non-empty set B. Termination of the program is obvious since M is finite, hence B is initially finite, too, and the cardinality of this set is strictly decreased by every run through the while-loop. Minimum(R) DECL A, B, b BEG A, B := R, R; WHILE B = ∅ DO b := elem(B); IF P(A \ b) THEN A, B := A \ b, B \ b ELSE B := B \ b FI OD RETURN A END . The running time of the program Minimum mainly depends on the costs of evaluating the predicate P and the number of runs of the while-loop. In many cases the evaluation of P can be improved if P is decremental in the sense that P(X \ x) ⇐⇒ P(X) ∧ Q(X, x)
(5)
for all non-empty X ∈ 2M and x ∈ X, where Q is an additional predicate on 2M × M . Since P(A) is part of the invariant inv (R, A, B), in this case the condition P(A \ b) of the conditional of Minimum can be replaced by Q(A, b) and this reduces the costs if the evaluation of Q is less expensive than that of P. The use of (5) is motivated by [12], where the dual property “to be incremental” is applied for computing maximal sets. Exceeding the work of [12], in the following we also consider the number of runs of the while-loop of Minimum. It is obvious that this number equals the cardinality of the input set R. Therefore, it seems to be a good idea to refine the initialization by a precomputation phase which yields, of course only if the precondition pre(R) holds, a set S such that P(S) is true, S ⊆ R, and the cardinality of S is much smaller than that of R.
264
R. Berghammer and T. Hoffmann
Since the conjunction of P(S) and S ⊆ R is equivalent to inv (R, S, S), this leads – in combination with requirement (5) – to the following correct refinement: Minimum (R) DECL S, A, B, b BEG precomputation phase ; { P(S) ∧ S ⊆ R } A, B := S, S; WHILE B = ∅ DO b := elem(B); IF Q(A, b) THEN A, B := A \ b, B \ b ELSE B := B \ b FI OD RETURN A END . In the next section we will use a relation-algebraic instantiation of this refined version to compute transitive reductions of strongly connected directed graphs.
4
Computing a Transitive Reduction
During this section we suppose that g = (V, R) is a fixed directed graph with finite vertex set V = ∅ and relation R : V ↔ V . Then computing a transitive reduction of g means to determine a directed graph r∗ = (V, A) such that A is a minimal subrelation of R satisfying A∗ = R∗ , i.e., r∗ is an arc-minimal subgraph of g with the same information as g concerning reachability. This task is a fundamental optimization problem with many applications (some of them are sketched in [11] for example) since a transitive reduction of a directed graph can be seen as a minimal storage representation of its reachability information. A transitive reduction of g with a minimum number of arcs is called a minimum equivalent digraph of g. In contrast to the problem of computing a transitive reduction this problem is NP-hard. Approximation algorithms for minimum equivalent digraphs have been developed in [10]. Of course, each algorithm for computing a transitive reduction can be seen as an approximation algorithm for minimum equivalent digraphs, too. We will come back to this later. Now let the supposed graph g in addition be strongly connected, i.e., R∗ = L be true. If we choose the universe M of Section 3 as the set [V ↔ V ] and the predicate P(X) of the generic program Minimum as P(X) :⇐⇒ X ∗ = L ,
(6)
then this instantiation solves for the input relation R the problem of computing the relation of a transitive reduction of the directed graph g since its precondition exactly describes that g is strongly connected, its postcondition exactly describes that A is the relation of a transitive reduction of g, and X ∗ = L is upwards-closed. Of course, we have to replace the empty set ∅, the choice elem(B), and the
Calculating a Relational Program for Transitive Reductions
265
difference A \ b resp. B \ b by their relation-algebraic counter-parts O, atom(B), and A ∩ b resp. B ∩ b . The predicate P of (6) is also decremental in the sense of (5) since for all non-empty relations X : V ↔ V and all x = atom(X) the equation (X ∩ x )∗ = L ⇐⇒ X ∗ = L ∧ x ⊆ (X ∩ x )∗
(7)
holds. A proof of “=⇒” is trivial and the remaining direction “⇐=” is shown by the calculation L = = = ⊆ =
X∗ ((X ∩ x ) ∪ x)∗ (X ∩ x )∗ ∪ (X ∩ x )∗ x(X ∩ x )∗ (X ∩ x )∗ ∪ (X ∩ x )∗ (X ∩ x )∗ (X ∩ x )∗ (X ∩ x )∗ .
assumption (1) and (3) assumption
Guided by the right-hand side of equivalence (7) we now choose the implication x ⊆ (X ∩ x )∗ as Q(X, x). To complete the instantiation of the generic program Minimum we still have to find a precomputation phase which establishes S ∗ = L and S ⊆ R. Here we follow an idea mentioned in [11] and seek for subgraphs h1 = (V, T1 ) of g and h2 = (V, T2 ) of g T = (V, RT ) with a common root r : V ↔ 1. Then the precomputation phase can consist of the assignment of T1 ∪ T2 T to S since L = LrT rL = (rL) rL ⊆ (T2∗ ) T1∗ = (T2 T )∗ T1∗ ⊆ (T1 ∪ T2 T )∗ T
T
follows with the help of rT r = O and the Tarski-rule, the root properties of r wrt. h1 and h2 , and two simple laws of reflexive-transitive closures, and T
T1 ∪ T2 T ⊆ R ∪ R T = R is trivial. For reasons of efficiency, the result S of the precomputation phase should also contain as few arcs as possible. Therefore, we seek in addition for subgraphs h1 and h2 which are trees. If we postpone this task to two relational programs Tree 1 and Tree 2 which both have Q and r as inputs, T as output, and are specified by the precondition ∧
pre(Q, r) = Q∗ = L ∧ r = rL ∧ rrT ⊆ I ∧ rT L = L (the latter three conjuncts say that r is a point) and the postcondition
266
R. Berghammer and T. Hoffmann ∧
post(Q, T, r) = T ⊆ Q ∧ rL ⊆ T ∗ ∧ T T T ⊆ I ∧ T + ⊆ I (here the latter three conjuncts say that T is the relation of a tree with root r), then in combination with the point property of the choice point(L), i.e., axioms (4), we immediately obtain correctness of the following relational program TransRed for computing the relation of a transitive reduction of g: TransRed (R) DECL S, A, B, b, r BEG r := point(L); T
S := Tree 1 (R, r) ∪ Tree 2 (RT , r) ; A, B := S, S; WHILE B = O DO b := atom(B); IF b ⊆ (A ∩ b )∗ THEN A, B := A ∩ b , B ∩ b ELSE B := B ∩ b FI OD RETURN A END . The union of the two trees in this program contains at most 2 ∗ |V | − 2 arcs. A transitive reduction of g has at least |V | arcs what in turn implies that the while-loop of TransRed is executed not more than |V | − 2 times. Hence, we obtain an overall running time O(|V |2 ) if we use two programs Tree 1 and Tree 2 with running times O(|V |2 ) and if we can test the condition b ⊆ (A ∩ b )∗ of TransRed in O(|V |) steps. The latter task, however, is trivial since the inclusion b ⊆ (A ∩ b )∗ describes that the sink of b can be reached from its source via arcs from (A ∩ b )∗ and this reachability problem can be solved in running time O(|A|) if A is implemented by successor lists. One possibility to test b ⊆ (A ∩ b )∗ in time O(|A|) is depth-first-search (usually abbreviated with DFS). But this well-known graph searching technique is also a possibility to solve the first of the just mentioned tasks. From [3] we immediately obtain that the following relational DFS program is correct wrt. the above precondition pre(Q, r) and postcondition post(Q, T, r): Dfs(Q, r) DECL T, b, g, p, w BEG T, b, g, p := O, O, r, r; WHILE QT r ∩ b ∪ r = O DO w := QT p ∩ b ∪ g ; IF w = O THEN b, g, p := b ∪ p, g ∩ p , T p ELSE q := point(w); g, T, p := g ∪ q, T ∪ pq T , q FI OD RETURN T END .
Calculating a Relational Program for Transitive Reductions
267
Of course, we are particularly interested in transitive reductions with few arcs, i.e., in approximations of minimum equivalent digraphs. To get an impression how the relational program TransRed with Dfs as instantiation of both Tree 1 and Tree 2 behaves wrt. this expectation, we have implemented it in RelView and performed numerous experiments. In each experiment we fixed the number N of vertices and generated random graphs – more exactly: their relations – using a specific basic operation of RelView, where we varied the number of arcs from 10% to 100% of all N 2 possible arcs. The following Figure 1 shows one of the results. Here we have N = 200, the density of the input relation of TransRed is listed at the x-axis, and the cardinality of the computed transitive reduction is listed at the y-axis:
Fig. 1. Density/cardinality-diagram for N = 200 and DFS precomputation
The curves of all other experiments are rather similar to that curve. They start at 10% a little bit above N , are nearly linear, and end at 100% a little bit below 2 ∗ N . Hence, the behaviour of the program does not correspond to our expectation because the number of arcs of a transitive reduction of g = (V, R) obviously should decrease if the density of R increases. Altogether, DFS seems not to be the right choice for the precomputation phase of TransRed . A second well-known approach for constructing a spanning subtree of a strongly connected directed graph from a given root is breadth-first-search, usually abbreviated with BFS. To get a relational formulation of this strategy, we start with the following relational program: Reach(Q, r) DECL x, y BEG x, y := r, QT r ∩ r ; WHILE y = O DO x := x ∪ y; y := QT x ∩ x OD RETURN x END .
268
R. Berghammer and T. Hoffmann
Given a directed graph and a vertex, this program computes the vertices which are reachable from the given vertex. In terms of relational algebra this means that it is correct wrt. the precondition ∧
pre(Q, r) = r = rL ∧ rrT ⊆ I ∧ rT L = L and the postcondition ∧
post(Q, r, x) = x = (Q∗ ) r . T
A formal derivation of the program Reach can be found in [4]. It combines relational algebra and the Dijkstra-Gries program development method and uses the following loop invariant: ∧
inv (Q, r, x, y) = r ⊆ x ∧ x ⊆ (Q∗ ) r ∧ y = x ∩ QT x . T
If one takes a closer look at the above reachabililty program, then one can see that it successively computes the vertices of the different layers of a BFS tree with root r. This is the reason why we use it as starting point of the relational BFS program we are interested in. Our idea is to introduce a new variable T , to compute a BFS tree with its help by adding some lines of code to the initialization and the body of the loop, and to return, finally, T as result instead of x. The initialization of T by xy T is obvious. To enlarge T by the vertices of the next layer by every run through the while-loop, we insert a second while-loop between the assignments x := x ∪ y and y := QT x ∩ x . This while-loop scans all vertices from the last layer (which is described by the vector y) and inserts arcs to their successors in the next layer (which is described by the vector QT (x ∪ y) ∩ x ∪ y ) in such a way that each vertex of the next layer is an endpoint of exactly one inserted arc. Thus, we arrive at the following relational BFS program: Bfs(Q, r) DECL a, p, w, x, y, z, T BEG x, y := r, QT r ∩ r ; T := xy T ; WHILE y = O DO x, w, z := x ∪ y, QT (x ∪ y) ∩ x ∪ y , O; WHILE z = w DO p := point(y); a := QT p ∩ x ∪ z ; y, T, z := y ∩ p , T ∪ paT , z ∪ a OD; y = QT x ∩ x OD RETURN T END . This program is correct wrt. the precondition pre(Q, r) and the postcondition post(Q, T, r); a formal proof is given in the appendix of this paper.
Calculating a Relational Program for Transitive Reductions
269
We have also implemented Bfs in RelView and used as instantiation of both Tree 1 and Tree 2 in the precomputation phase of TransRed . The experiments we have performed show that this version is even worser than the first version using the DFS program. As an example, in Figure 2 we present again the density/cardinality-diagram for N = 200 vertices; all other curves are rather similar to the curve of the figure.
Fig. 2. Density/cardinality-diagram for N = 200 and BFS precomputation
Both versions of the relational program TransRed we have investigated until now show the same tendency, viz. the denser the input relation gets the more arcs are contained in the result. This is in contrast to our expectation that the number of arcs of a transitive reduction should decrease if the density of the underlying directed graph increases. The reason for the bad behaviour of our hitherto relational programs gets clear if one compares them with the approximation algorithms for minimum equivalent digraphs presented in [10]. These algorithms compute approximations which are composed of a few large cycles only. In contrast to this, when using the BFS strategy for computing spanning subtrees h1 = (V, T1 ) of g and h2 = (V, T2 ) of g T with a common root r, then h1 contains shortest paths leading from r to every other vertex of g and hT 2 contains shortest path leading from every vertex of g except r to r. Thus, the result of precomputation phase of TransRed , which consists of the union of the relations T1 and T2 T , contains a lot of short cycles which are getting shorter if the graph g gets denser. The same phenomenon appears when the DFS strategy is applied twice in the precomputation phase of TransRed . The computation of h1 following the DFS strategy leads to long paths from r to every other vertex of g, and in the same manner hT 2 contains long paths back to r. Therefore, a lot of vertices lie on both paths which in turn implies that the union of T1 and T2T contains many short cycles. As in the case of DFS, these are getting shorter if the graph gets denser. Now, to obtain long cycles the idea is to use the DFS strategy to compute h1 , which then contains long paths from r to every other vertex of g. For the
270
R. Berghammer and T. Hoffmann
computation of h2 , however, the use of the BFS strategy seems to be profitable, because then hT 2 contains shortest paths from every vertex of g except r back to r and, hence, the value T1 ∪ T2T of S in TransRed leads to long cycles and better results. Again we have tested this approach with the RelView tool and arrived, for example, in the case N = 200 at the diagram of Figure 3; the diagrams of all other experiments with the two relational programs Dfs and Bfs for computing T1 resp. T2 look rather similar.
Fig. 3. Density/cardinality-diagram for N = 200 and DFS/BFS precomputation
Summing up, the experiments suggest that the combination of DFS and BFS is the right choice for the computation of a transitive reduction with as few arcs as possible using the program TransRed . In contrast to the other two approaches here the resultant transitive reduction has much fewer arcs. Furthermore we obtain the expected tendency that the result is getting better when the graph gets denser.
5
Conclusion
In this paper we have developed a relational program for computing the transitive reduction of strongly connected directed graphs as a specific instantiation of a generic program for computing minimal subsets. It uses a precomputation phase which combines DFS and BFS. When regarding it as an approximation algorithm for minimum equivalent digraphs, then the approximation factor is 2. This means that the computed transitive reduction contains at most twice as much arcs as a minimum equivalent digraph2 . But our numerous practical tests with the RelView system have shown that for randomly generated relations 2
In the case of minimalization problems an approximation algorithm has an approximation factor f if for all inputs the size c of the computed result is bounded by c ≤ f ∗ opt, where opt is the size of an optimal result.
Calculating a Relational Program for Transitive Reductions
271
all results are very close to a minimum equivalent digraph, i.e., the “average approximation factor” is very close to 1. In the course of the diploma thesis [9] we have compared our approach with the approximation algorithms of [10]. We have implemented the best algorithm of [10], which computes an approximation of a minimum equivalent digraph in 2 2 nearly linear time with approximation factor f between π6 ≈ 1.64 and π 6+1 ≈ 1.81. But in practical tests this algorithm leads to results which contain in the average 3∗n 2 arcs, where n is the cardinality of the minimum equivalent digraphs. In the concrete case of N = 200 vertices, which we have considered in detail in Section 4, this means that the result contains in the average 300 arcs. This is far away from the optimum and also worse than the results our final program yields. Furthermore these results are no transitive reductions. To overcome the disadvantages of both approaches, in [9] they have been combined by replacing the DFS/BFS precomputation phase in TransRed with a call of the algorithm of [10]. This, finally, leads to an algorithm with the theoretical approximation factor of [10] and our good practial results.
References 1. Aho A.V., Hopcroft J.E., Ullman J.D.: The design and analysis of computer algorithms. Addison-Wesley (1974). 2. Behnke R. et al.: Applications of the RelView system. In: Berghammer R., Lakhnech Y. (eds.): Tool support for system specification, development and verification. Advances in Computing, Springer, 33-47 (1999). 3. Berghammer R., Hoffmann T.: Relational depth-first-search with applications. In: Desharnais J. (ed.): Proc. 5th Int. Seminar on Relational Methods in Computer Science, Laval Univ., Dept. d’ Informatique, 11-20 (2000). Extended version in: Information Sciences 139, 167-186 (2001). 4. Berghammer R.: Combining relational calculus and the Dijkstra-Gries method for deriving relational programs. Information Sciences 119, 155-171 (1999). 5. Cormen T.H., Leiserson C.E., Rivest R.L.: Introduction to algorithms. The MIT Press (1990). 6. Dijkstra E.W.: A discipline of programming. Prentice-Hall (1976). 7. Gries D.: The science of computer programming. Springer (1981). 8. Hoffmann T.: Case studies in relational program development by selected strategies for graph searching (in German). Ph.D. thesis, Institut f¨ ur Informatik und Praktische Mathematik, Universit¨ at Kiel (2002). 9. Kasper C.: Investigating algorithms for transitive reductions and minimum equivalent digraphs (in German). Diploma thesis, Institut f¨ ur Informatik und Praktische Mathematik, Universit¨ at Kiel (2001). 10. Khuller S., Raghavachari B., Young N.: Approximating the minimum equivalent digraph. SIAM Journal on Computing 24, 859-972 (1995). 11. Noltemeier H.: Reduktion von Pr¨ azedenzstrukturen. Zeitschrift f¨ ur Operations Res. 20, 151-159 (1976). 12. Ravelo J.: Two graph algorithms derived. Acta Informatica 36, 489-510 (1999). 13. Schmidt G., Str¨ ohlein T.: Relations and graphs. Springer (1993). 14. Simon K.: Finding a minimal transitive reduction in a strongly connected digraph within linear time. In: Nagl M. (ed.): Proc. 15th Workshop on Graph-Theoretic Concepts in Computer Science, LNCS 411, Springer, 245-259 (1990).
272
R. Berghammer and T. Hoffmann
Appendix In this appendix we focus on the formal verification of the relational BFS program from Section 4. First we start with some preparatory steps. A relation T is called a singleton forest if the following three formulae hold: TTT ⊆ I
T+ ⊆ I
∗
(T ∪ T T )L(T ∪ T T ) ⊆ (T ∪ T T ) .
Graph-theoretically this means that t = (V, T ) is a forest in which every pair of non-isolated vertices is connected via an undirected path. If additionally r is a point fulfilling r = T L∩ T T L , we call T a singleton forest with root r. Obviously, this is a generalization of the definition of rooted trees from Section 2. In this paper we only deal with relations on finite sets. We can describe finiteness of a set X in relational algebra by demanding that the membership relation ∈ : X ↔ 2X is Noetherian. But this characterization is often unwieldy. Therefore, we have decided to use the postulate (T ∗ ) (T L ∩ T T L ) = (T ∪ T T )L T
(8)
for all singleton forests T since this is exactly what we need in the subsequent verification. Using graph-theoretic terminology, the equality means: If there exist a root of t = (V, T ), then exactly all non-isolated vertices are reachable from it. Otherwise, all vertices are isolated. Note that (8) does not hold for arbitrary relations. For example take the successor-relation on the integers, which obviously is a singleton forest. In this case the left-hand side of (8) is the empty vector, but its right-hand side is the universal vector. Before we start the formal program verification, we collect two properties of singleton forests. Their proofs can be found in [8]. Let p and r be points, v be a vector, and T be a singleton forest with root r. (i) If p ⊆ v , then the relation pv T is a singleton forest with root p. (ii) If p ⊆ (T ∪ T T )L and v ⊆ (T ∪ T T )L , then the following properties hold: ∗
∗
(a) (T ∪ pv T ) = T ∗ (pv T )
(b) T ∪ pv T is a singleton forest with root r.
After these preparatory steps we begin with the verification of the relational program Bfs, where, see Section 4, ∧
pre(Q, r) = Q∗ = L ∧ r = rL ∧ rrT ⊆ I ∧ RT L = L is the precondition and ∧
post(Q, T, r) = T ⊆ Q ∧ rL ⊆ T ∗ ∧ T T T ⊆ I ∧ T + ⊆ I is the postcondition. The program uses two nested while-loops. Consequently, we need two invariants, an invariant inv outer (Q, T, r, x, y) for the outer whileloop and a second invariant inv inner (Q, T, r, w, x, y, z) for the inner while-loop.
Calculating a Relational Program for Transitive Reductions
273
Here inv outer (Q, T, r, x, y) is an extension of the invariant inv (Q, r, x, y) of the program Reach and consists of the conjunction of the upcoming properties: r⊆x
(9) ∗ T
x ⊆ (Q ) r y = QT x ∩ x T ⊆Q
(10) (11) (12)
x ∪ y = (T ∗ ) r T is a singleton forest with root r.
(13) (14)
T
The invariant inv inner (Q, T, r, w, x, y, z) of the inner while-loop consist of the conjunction of the next five properties: T ⊆Q
(15) ∗ T
x ∪ z = (T ) r T is a singleton forest with root r w = QT x ∩ x
(16) (17) (18)
y ⊆ (T ∗ ) r.
(19)
T
The verification of the relational BFS program is done in five steps. First we focus on the inner invariant and show its establishment and maintenance. Afterwards we do the same with the invariant of the outer while-loop. Finally, we prove that y = O and the validity of the outer invariant imply the postcondition. During all these steps we are allowed to use the precondition pre(Q, r). We start the correctness proof with the establishment of the inner invariant by the assignment to x, w, and z in front of the inner while-loop. Here we are allowed to use the properties of inv outer (Q, T, r, x, y). The establishment of the formulae (15), (17), and (19) follow directly from the validity of (12), (14), and (13). With the aid of equation (13) and the initialization of x by x ∪ y resp. z by O, the validity of formula (16) for the latter values is shown by x ∪ y ∪ O = x ∪ y = (T ∗ ) r . T
Finally, the establishment of formula (18) is a direct consequence of the assignment w := QT x ∩ x . In the second step we show the maintenance of the inner invariant starting with inclusion (15). Using that p is a point, we get paT ⊆ ppT Q ⊆ Q from the assignment to a which in turn yields T ∪ paT ⊆ Q due to T ⊆ Q. The maintenance of formulae (16) and (17) can be shown with property (ii). Therefore we have to check whether p and a (instead of v) fulfill the assumptions. Obviously, both inclusions p ⊆ (T ∪T T )L and a ⊆ (T ∪ T T )L follow with postulate (8) from p ⊆ y, (19), and (17) resp. a ⊆ x ∪ z and (16). Consequently the assumptions of (ii) are fulfilled and part (a) of (ii) helps us to prove equation (16) for the new values of the variables z and T as follows:
274
R. Berghammer and T. Hoffmann
x∪z∪a = = = =
(T ∗ ) r ∪ a T (T ∗ ) r ∪ aL T ∗ T (T ) r ∪ apT (T ∗ ) r T (T ∗ ∪ T ∗ paT ) r
x ∪ z = (T ∗ ) r a vector T T (19) ⇒ p ⊆ (T ∗ ) r ⇒ L ⊆ pT (T ∗ ) r
T
T
∗ T
∗
= (T ∗ (paT ) ) r
(paT ) = I ∪ paT
∗ T
= ((T ∪ paT ) ) r
(a) of (ii)
The maintenance of property (17) follows directly from part (b) of (ii). Since the values of x and w remain unchanged by every run through the inner while-loop, maintenance of equation (18) is trivial. The maintenance of the last formula (19) of the inner invariant, finally, follows from ∗ T
y ∩ p ⊆ y ⊆ (T ∗ ) r ⊆ ((T ∪ paT ) ) r , T
where the validity of y ⊆ (T ∗ ) r is applied in the second step. This completes the maintenance proof of the invariant of the inner while-loop. In the next two steps we concentrate on the invariant of the outer while-loop. Here we only have to look at the formulae (12)-(14), because the other properties (9)-(11) are exactly the conjuncts of the invariant of the relational reachability program Reach. The corresponding correctness proofs can be found in [4]. We start with the proof that the formulae (12)-(14) are established by the assignments to x, y, and T in front of the outer while-loop. The assignments to T , x, and y imply T = xy T ⊆ rrT Q, which, in combination with the point property of r, establishes inclusion (12). With aid of the assignments to x and ∗ T , equation (xy T ) = I ∪ xy T , and xT x = rT r = L (which is again a consequence of the point property of r) we get (13) from T
∗ T
(T ∗ ) r = (T ∗ ) x = ((xy T ) ) x = (I ∪ yxT )x = x ∪ yxT x = x ∪ yL = x ∪ y . T
T
The establishment of formula (14) is an immediate consequence of the point property of r and property (i). Step four of the verification is the maintenance proof of the outer invariant. Here we are allowed to use the properties of inv inner (Q, T, r, w, x, y, z) and the exit condition of the inner while-loop. The maintenance of formula (12) follows directly from the validity of (15). In combining the exit condition z = w with equation (18) we get z = x ∩ QT x. Now, with the help of this formula, equation y = x ∩ QT x, and formula (16), the maintenance of property (13) can be shown with by following calculation: x ∪ y = x ∪ ( x ∩ QT x) = x ∪ z = (T ∗ ) r . T
The last property (14) for the new value of T ensues from property (17). As the fifth step of the verification it remains to show that the postcondition post(Q, T, r) follows from the invariant of the outer while-loop and this loop’s exit condition y = O. We only have to show the second conjunct rL ⊆ T ∗ of post(Q, T, r) because the other formulae are immediate consequences of (12) and
Calculating a Relational Program for Transitive Reductions
275
(14). From equation (13) and y = O we get x = (T ∗ ) r and x = (Q∗ ) r also holds due to the correctness of the reachability program. Putting these equations together and using the strong connectivity of Q and the point property of r we obtain the desired inclusion by T
T
rL = rrT L = rrT Q∗ = rrT T ∗ ⊆ T ∗ . Thereby the correctness proof of the relational BFS program is complete.
Calculating Church-Rosser Proofs in Kleene Algebra Georg Struth Institut f¨ ur Informatik, Universit¨ at Augsburg Universit¨ atsstr. 14, D-86135 Augsburg, Germany Tel:+49-821-598-3109, Fax:+49-821-598-2274,
[email protected] Abstract. We present simple calculational proofs of Church-Rosser theorems for equational theories, quasiorderings and non-symmetric transitive relations in Kleene algebra. We also calculate the abstract part of two standard proofs of Church-Rosser theorems in the λ-calculus and further central statements of rewriting. Since proofs avoid deduction, in particular induction, and large parts are amenable to automata, the approach is suited for mechanization. Since proofs algebraically reconstruct precisely the usual diagrams, they are also very natural for a human. In all considerations, Kleene algebra is an excellent means of abstraction. Keywords: Kleene algebra, rewriting, λ-calculus, Church-Rosser theorem, formal mathematics.
1
Introduction
A main concern of formal mathematics and formal methods is the combination of simple readable specifications with powerful proof search. This combination depends on several principles. First, the combination of formal and informal rigor. Specifications and proofs should be precise, concise, natural and revealing both for machines and for humans. Second, the balance of expressive and deductive power. These properties are mutually exclusive. Third, the replacement of deduction by calculation. Routine work, decision procedures, algorithms should be regarded prior to creativity, like for instance induction. Algebraic abstraction is a key to these principles, especially to the second and third one. Abstracting, one may divide and conquer, hide proof complexity in bridge lemmata between different levels, gain structural insight and obtaining simple, natural, revealing specifications and proofs at all levels. Here, we demonstrate these principles by calculating proofs of Church-Rosser theorems in Kleene algebra. These include transitive relations, quasi-orderings, equational theories and abstract parts of the two standard techniques for the λ-calculus, namely that of Tait-Martin-L¨ of (c.f. [2]) and Takahashi [26] and Barendregt’s indexing technique [2]. We also calculate proofs of the commutation lemma and commutative union lemma of Hindley and Rosen. Church-Rosser H. de Swart (Ed.): RelMiCS 2001, LNCS 2561, pp. 276–290, 2002. c Springer-Verlag Berlin Heidelberg 2002
Calculating Church-Rosser Proofs in Kleene Algebra
277
theorems are central to term rewriting and the λ-calculus with important applications in functional programming, universal algebra, computer algebra and mechanized reasoning. Proofs usually treat properties of (first-order or λ-) terms and those of the associated rewrite relation separately. Bridge lemmata describe their coupling. Reasoning at the relational level is supported by diagrams; their semantics can be given in terms of logic and binary relations. Textbook proofs are usually only informally rigorous. They use diagrams without any safe rules for their transformation or combination. The literature contains several errors even with rather simple statements. Formal Church-Rosser proofs with interactive proof checkers are considered challenging. To obtain at least formal rigor, one typically proves ad hoc structural lemmata about diagrams in a bottom-up way, using induction and essentially higher-order logic. Proof complexity and therefore our second and third principle is usually disregarded. Unlike previous approaches, we proceed top-down via two abstraction steps. We first translate the logical specifications of Church-Rosser theorems into pointfree relational ones. We then extract the relevant algebraic properties of settheoretic relations. Since only the regular operations (union, composition, reflexive transitive closure) occur in Church-Rosser theorems, Kleene algebra is a more focused target theory than, for instance, relation algebra. Syntactically, a Kleene algebra is a semiring with an additional Kleene star operation modeling iteration or fixed point computation. The Kleene star is defined roughly as the least solution of a fixed point equation, but neither as a limit of an approximation sequence nor as a supremum of an infinite sum of powers. Semantically, Kleene algebra is precisely the algebra of regular events. Since also the regular operations on set-theoretic relations is among its models, properties of Kleene algebra can immediately be transfered to rewrite relations in a top down way. Kleene algebra yields simple formally and informally rigorous proofs of Church-Rosser theorems. The first claim follows in particular from our implementation of Kleene algebra and proofs with the Isabelle proof checker [25]. The second claim holds, since our proofs formally reconstruct precisely the diagrammatic proofs and thereby give an algebraic semantics to diagrams. Moreover they use only natural and general properties of regular operations. Expressive and deductive power are well-balanced. Deduction is completely replaced by calculation. There is no induction and higher-order reasoning involved, fixed points are computed instead. By the strong connection between Kleene algebra and regular languages, large parts of proofs are even amenable to automata. Beyond Church-Rosser theorems, the replacement of “regular” induction (the fragment of induction captured by automata) by first-order fixed point computation seems of general interest for formal methods, since first-order proof search methods and even decision procedures can be used then. Relational specifications and proofs are at the heart of methods like Z and B [22,1]. We have tried to include the most important proofs in the paper. A full formal treatment can be found in an extended version [24]. The remainder is organized as follows. Section 2 and section 3 introduce some Church-Rosser and Kleene algebra basics. Section 4 recalls further properties
278
G. Struth
of Kleene algebra for the Church-Rosser proofs for rewriting with quasiorderings and non-symmetric transitive relations in section 5. Section 6 introduces a conversion operation and functions, to specialize the theorems of section 5 to equational rewriting and to prepare for two classical proofs of the Church-Rosser theorem in the λ-calculus in section 7 and section 8, with particular emphasis on mechanized reasoning. Section 9 contains a concluding discussion.
2
Church-Rosser Basics
Let (A, →) be a rewrite system of a set A and a binary set-theoretic relation → on A. We usually write A for (A, →) and a → b. The relation ← denotes the converse of →, the relations →+ , →∗ and ↔∗ denote the smallest transitive relation, quasiordering and equivalence containing →. Juxtaposition of arrows denotes relational composition. The diamond property 3(x1 , x2 , x3 ) and the ChurchRosser property (x1 , x2 ) are defined by 3(x1 , x2 , x3 ) (x1 , x2 )
⇔
(x1 →∗ x2 ∧ x1 →∗ x3 ⇒ ∃x4 .x2 →∗ x4 ∧ x3 →∗ x4 ), (1)
⇔
(x1 ↔∗ x2 ⇒ ∃x3 .x1 →∗ x3 ∧ x2 →∗ x3 ).
(2)
The rewrite system A has the diamond property, if ∀x1 , x2 , x3 ∈ A.3(x1 , x2 , x3 ) and the Church-Rosser property, iff ∀x1 , x2 ∈ A. (x1 , x2 ). The properties are usually visualized by the diagrams ∗ x2
x1
∗ x3
∗ x ∗ 4
x1
∗
x2
∗ x ∗ 3
which are often also decorated with universal or existential quantifiers or different kinds of arrows to point out the relation with (1) and (2). Theorem 1 (Church-Rosser). A rewrite system has the diamond property, iff it has the Church-Rosser property. Proof. We only sketch the standard proof that the diamond property implies the Church-Rosser property. It is by induction on the number of rewrite steps in ↔∗ . In the base case, the Church-Rosser property trivially holds for the zero-step sequence t ↔ t. The induction step is visualized by the diagrams n−1 · · · n−1 ∗ ∗ · · · · ∗ ∗ ∗ ∗ · · Only the right-hand diagram depends on an assumption, which can be shown to follow from the diamond property. 2 The Church-Rosser and the diamond property can be expressed purely relationally, hence in a more abstract point-free way.
Calculating Church-Rosser Proofs in Kleene Algebra
279
Lemma 1. Let A be a rewrite system. (i) ∀x1 , x2 , x3 ∈ A.3(x1 , x2 , x3 ) ⇔ ←∗ →∗ ⊆ →∗ ←∗ , (ii) ∀x1 , x2 ∈ A. (x1 , x2 ) ⇔ (→ ∪ ←)∗ ⊆ →∗ ←∗ . Lemma 1 is proven using the definitions of set-theoretic union, relational composition and the different closures of →. In section 3 we introduce an algebra to reason rigorously about the relevant algebraic properties of these relations.
3
Kleene Algebra
A Kleene algebra is a structure for modeling sequential composition, non-deterministic choice and iteration or fixed point computation. The field has been pioneered by Conway [5] in the context of the algebra of regular events. Besides formal languages and automata, Kleene algebras also arise, for instance, in the context of relation algebra (c.f. [20]) and logics, analysis and construction of programs (c.f. [11]). We follow Kozen’s definition [13]. These structures are called Kozen semirings in [3]. For concise proofs of all statements in this text see [24]. Some formal proofs with the Isabelle proof checker can be found in [25]. A semiring is an algebra (A, ·, +, 0, 1) of a set A, two binary operations of multiplication and addition and two constants 0 and 1. Thereby (A, +, 0) is a commutative monoid, (A, ·, 1) a monoid, multiplication distributes over addition from the left and right and 0 is both a left and a right annihilator (0a = a0 = 0). We often write A for (A, ·, +, 0, 1) and ab for a · b. The relation ≤ defined on a semiring A by a ≤ b ⇔ (∃c ∈ A).a + c = b is a quasiordering. It is a partial ordering, if A is idempotent, that is a + a = a holds for every a ∈ A and a ≤ b ⇔ a + b = b. Addition and (left and right) multiplication are monotonic operations with respect to ≤ on A; 0 is the minimal element. If A is idempotent, it is a semilattice with respect to addition and least upper bounds. Thus for all a, b, c ∈ A, a + b ≤ c ⇔ a ≤ c ∧ b ≤ c.
(3)
A Kleene algebra (A, ·, +, ∗ , 0, 1) is an idempotent semiring (A, ·, +, 0, 1) endowed with an additional unary operation ∗ (Kleene star ) that satisfies 1 + aa∗ ≤ a∗ , ∗
∗
1+a a≤a ,
(4) (5)
ac + b ≤ c ⇒ a∗ b ≤ c, ca + b ≤ c ⇒ ba∗ ≤ c,
(6) (7)
for all a, b, c ∈ A. Thus a∗ b and ba∗ are the least pre-fixed points of the monotonic mappings λx.b + ax and λx.b + xa. a∗ is also a least fixed point of λx.1+ax, λx.1+xa; the inequalities (4) and (5) can be strengthened to equations. Moreover, this fixed point is uniquely defined. (6) and (7) are algebraic variants of fixed point induction. Computationally we use them as star elimination rules. Kleene algebra characterizes the Kleene star algebraically (or combinatorially) and not analytically as a limit of an approximation sequence or a supremum
280
G. Struth
of an infinite sum of powers. Our proofs are therefore simple induction-free calculations (often finite combinatorics) that are well-suited for automation. Since relations satisfy infinite distributivity laws and continuity axioms, proofs in Kleene algebras are more general than relational ones. We define the Kleene plus by a+ = aa∗ . Obviously, a∗ = 1 + a+ = a+ + a∗ . There are pre-fixed point laws like (4) to (7). Lemma 2. let A be a Kleene algebra. Then for all a, b, c ∈ A a(a+ + 1) ≤ a+ , (a + 1)a ≤ a , +
+
(8)
a(b + c) ≤ c ⇒ a+ b ≤ c,
(10)
(9)
(b + c)a ≤ c ⇒ ba ≤ c.
(11)
+
The Kleene plus and star are monotonic operations on a Kleene algebra.
4
Some Properties of Kleene Algebra
We now review some well-known properties of Kleene algebra that are important for the Church-Rosser proofs. The first statement is a completeness result. Theorem 2 ([13]). There is a deductive system for the universal Horn theory of Kleene algebra such that an equation between Kleene algebra terms follows from the axioms of Kleene algebra iff the two terms denote the same regular set. Completeness results for related structures have been given, for instance, in [5, 3]. Remind that there is no finite equational axiomatization of the algebra of regular events [19]. By theorem 2, every identity in Kleene algebra can be decided by automata. In contrast, Kleene algebra is incomplete for the Horn theory of regular events [13] and the Horn theory of Kleene algebra is undecidable (by unsolvability of the uniform word problem for semigroups). The second statement guarantees that our Church-Rosser proofs in Kleene algebra are adequate for set-theoretic relations. Theorem 3. The set-theoretic relations on a given set A are a Kleene algebra. Thereby, + is interpreted as set-theoretic union, · as relational composition, ∗ as the reflexive transitive closure operation, 0 as the empty, 1 as the identity relation and ≤ as set inclusion. The third and fourth statement collect some standard properties of the Kleene star and plus. Most of them are well-known also for relation algebra, regular languages or automata (c.f. [5,9,13,20]). Lemma 3. Let A be a Kleene algebra. For all a, b, c ∈ A, 1 = 1∗ , ∗
1≤a , a a = a∗ , a ≤ a∗ , ∗ ∗
(12) (13) (14) (15)
a∗∗ = a∗ , ac ≤ cb ⇒ a∗ c ≤ cb∗ , ∗
∗
cb ≤ ac ⇒ cb ≤ a c, (a + b)∗ = a∗ (ba∗ )∗ .
(16) (17) (18) (19)
In particular, (ab)∗ a = a(ba)∗ and a∗ a = aa∗ are simple consequences of (17).
Calculating Church-Rosser Proofs in Kleene Algebra
281
Lemma 4. Let A be a Kleene algebra. For all a, b, c ∈ A, a+ a+ ≤ a+ , a≤a , +
∗
a ≤a , +
aa = a a ≤ a , +
5
+
+
(20)
a+ a∗ = a∗ a+ = a+ ,
(24)
(21)
ac ≤ cb ⇒ a c ≤ cb ,
(25)
(22)
cb ≤ ac ⇒ cb ≤ a c,
(26)
(23)
(a + b) = (a b) a + a .
+
+
+
∗
+ ∗
+
+
+
(27)
Church-Rosser Theorems
We now present the Church-Rosser theorems. They are further discussed in the remaining sections. In all Kleene algebra computations, inequational reasoning pays. (6) and (7)—as star elimination rules—are the working horses. Also monotonicity and (3) (splitting into subgoals) are abundantly used. Proposition 1. Let A be a Kleene algebra. For all a, b ∈ A, (i) ba∗ ≤ a∗ b∗ ⇒ (a + b)∗ ≤ a∗ b∗ . (ii) b∗ a∗ ≤ a∗ b∗ ⇒ (a + b)∗ ≤ a∗ b∗ . (iii) ba ≤ ab ⇒ (a + b)∗ ≤ a∗ b∗ . Proof. (ad i) By (6) it suffices to show (a + b)a∗ b∗ + 1 ≤ a∗ b∗ ; thus, by (3), 1 ≤ a∗ b∗ , ba∗ b∗ ≤ a∗ b∗ .
(28)
aa∗ b∗ ≤ a∗ b∗ ,
(29)
(30)
(28) follows immediately from (13): 1 ≤ a∗ = a∗ 1 ≤ a∗ b∗ . (29) follows immediately from (15), (14) and monotonicity of multiplication: a ≤ a∗ implies aa∗ b∗ ≤ a∗ a∗ b∗ = a∗ b∗ . (30) is the only inequality that depends on the assumptions and therefore is beyond automata theory. ba∗ b∗ ≤ a∗ b∗ b∗ = a∗ b∗ , by the assumption, (14) and monotonicity of multiplication. (ad ii) It suffices to show b∗ a∗ ≤ a∗ b∗ ⇒ ba∗ ≤ a∗ b∗1 . This is straightforward from (15) and monotonicity of multiplication. (ad iii) It suffices to show ba ≤ ab ⇒ ba∗ ≤ a∗ b∗ . By(17), (15) and the 2 assumption, ba∗ ≤ a∗ b ≤ a∗ b∗ . The proof of proposition 1 (i) is simple and natural. The first steps are obvious. Fixed-point induction (axiom 6) eliminates the Kleene star in (a + b)∗ and (3) splits into three subgoals. Comparing with the proof of theorem 1, we see the precise correspondence between (28), (29) and (30) with the base case and the two cases of the induction step. In particular we are almost automatically lead to the right assumption in (30). (28) and (29) are trivial; they can be decided 1
In fact, this statement is a logical equivalence.
282
G. Struth
by automata. The properties used for proving (28), (29) and (30) are again simple and natural. They formalize exactly the arguments that are implicitly used in the diagrammatic proof of theorem 1. Besides the semiring arithmetics and monotonicity we use (13), (15),(14), which encode reflexivity, extensivity and transitivity of a∗ . Thus our calculations yield precisely formally rigorous algebraic reconstructions of the informally rigorous diagrammatic proofs. Corollary 1. In proposition 1, the inequality (a+b)∗ ≤ a∗ b∗ can be strengthened to an equality, since a∗ b∗ ≤ (a + b)∗ . Proof. By (19), it suffices to show that a∗ b∗ ≤ a∗ (ba∗ )∗ . By (13), 1 ≤ a∗ . By monotonicity of multiplication and the Kleene star, therefore b ≤ ba∗ , b∗ ≤ (ba∗ )∗ and a∗ b∗ ≤ a∗ (ba∗ )∗ . 2 Theorem 4 (Church-Rosser). Let A be a Kleene algebra. For all a, b ∈ A, b∗ a∗ ≤ a∗ b∗ ⇔ (a + b)∗ ≤ a∗ b∗ . Proof. By proposition 1 (ii), it remains to prove that (a + b)∗ ≤ a∗ b∗ implies b∗ a∗ ≤ a∗ b∗ . By corollary 1 and the hypothesis, b∗ a∗ ≤ (a + b)∗ ≤ a∗ b∗ . 2 All calculations for the Church-Rosser theorems—starting from scratch—are straightforward and induction-free. They often use fixed-point induction (axioms (6), (7)) instead. This simplicity also appears in our formalization with the proof checker Isabelle [25]. We generally reach a high degree nof automation. Proposition 1 and theorem 4 can be generalized to sums i=0 ai and products n a in more than two variables. The proofs (c.f. [24]) use induction on the i i=0 number of generators and proposition 1 as base cases. The induction is not difficult, but proofs become longer and require some bookkeeping. Proposition 2. Let A be a Kleene algebra. For all a0 , . . . , an ∈ A. n n (i) a∗j a∗i ≤ a∗i a∗j for all 0 ≤ i < j ≤ n implies ( i=0 ai )∗ ≤ i=1 a∗i . n n (ii) aj ai ≤ ai aj for all 0 ≤ i < j ≤ n implies ( i=0 ai )∗ ≤ i=1 a∗i . Proposition 2 (i) is related to Church-Rosser theorems modulo an equivalence relation [12]. Proposition 2 (ii) is an algebraic completeness proof of bubble sort. We now calculate Church-Rosser proofs for the Kleene plus. Proposition 3. Let A be a Kleene algebra. For all a, b ∈ A (i) ba∗ ≤ a∗ b+ + a+ b∗ ⇒ (a + b)+ ≤ a∗ b+ + a+ b∗ . (ii) b+ a∗ ≤ a∗ b+ + a+ b∗ ⇒ (a + b)+ ≤ a∗ b+ + a+ b∗ . (iii) ba ≤ ab ⇒ (a + b)+ ≤ a∗ b+ + a+ b∗ . Proof. (ad i) By (10) it suffices to show (a + b)(a∗ b+ + a+ b∗ ) ≤ a∗ b+ + a+ b∗ , thus by distributivity and (3), aa∗ b+ ≤ a∗ b+ + a+ b∗ , + ∗
∗ +
+ ∗
aa b ≤ a b + a b ,
(31) (32)
ba∗ b+ ≤ a∗ b+ + a+ b∗ , + ∗
∗ +
+ ∗
ba b ≤ a b + a b .
(33) (34)
Calculating Church-Rosser Proofs in Kleene Algebra
283
(31) holds, since aa∗ b+ ≤ a∗ b+ by (15) and (14) and a∗ b+ ≤ a∗ b+ + a+ b∗ by definition of ≤. (32) holds, since aa+ b∗ ≤ a+ b∗ by (23) and monotonicity of multiplication. Moreover, a+ b∗ ≤ a∗ b+ + a+ b∗ by definition of ≤. (33) depends on the assumption and therefore is beyond automata theory. ba∗ b+ ≤ a∗ b+ b+ + a+ b∗ b+
by assumption, distributivity
∗ +
+ ∗ ∗
by (20), (22), monotonicity of + and ·
∗ +
+ ∗
by (14), monotonicity of + and ·
≤a b +a b b =a b +a b
(34) also depends on the assumption. ba+ b∗ ≤ ba∗ b∗ ∗ + ∗
by (22) + ∗ ∗
≤a b b +a b b ∗ +
+ ∗
≤a b +a b
by assumption, distributivity by (24), (14), monotonicity of + and ·
(ad ii) It suffices to show b+ a∗ ≤ a∗ b∗ ⇒ ba∗ ≤ a∗ b∗ . This is straightforward from (21) and monotonicity of multiplication. (ad iii) It suffices to show ba ≤ ab ⇒ ba+ ≤ a∗ b+ + a+ b∗ . By(25), (21), the assumption, monotonicity of addition and multiplication and the definition of ≤, we obtain ba+ ≤ a+ b ≤ a+ b∗ ≤ a+ b∗ + a∗ b+ . 2 Theorem 5 (Church-Rosser). Let A be a Kleene algebra. For all a, b ∈ A, b+ a∗ ≤ a∗ b+ + a+ b∗ ⇔ (a + b)+ ≤ a∗ b+ + a+ b∗ . Proof. By proposition 3 (ii), it remains to show that (a + b)+ ≤ a∗ b+ + a+ b∗ implies b+ a∗ ≤ a∗ b+ +a+ b∗ . We show b+ a∗ ≤ (a+b)∗ . The rest follows from transitivity and the hypothesis. By (27) it suffices to show that b+ a∗ ≤ (a∗ b)+ a∗ +a+ and therefore b+ a∗ ≤ (a∗ b)+ a∗ by definition of ≤. By (15), 1 ≤ a∗ . By monotonicity of multiplication and the Kleene plus, therefore b ≤ a∗ b, b+ ≤ (a∗ b)+ and b+ a∗ ≤ (a∗ b)+ a∗ . 2 Obviously, the proofs for the Kleene plus are somewhat harder than those with the Kleene star. Like in the proof of proposition 2, proposition 3 yields base cases for inductive proofs of Church-Rosser statements with more than two generators. Theorem 4 and theorem 5 are the basis for non-symmetric rewriting with quasiorderings and non-symmetric transitive relations [23]. They can be connected with the term structure by bridge lemmata that are usually called critical pair lemmata. The inequalities b∗ a∗ ≤ a∗ b∗ are examples of semicommutation properties (as opposed to commutation properties like ab = ba). As we will see in the next section, their interpretations in the relational model generalize the diamond property. Non-symmetric Church-Rosser theorems generalize their equational counterparts.
284
6
G. Struth
Conversion and Functions
To obtain precisely an algebraic variant of the Church-Rosser theorem (theorem 1), we add a converse operator. A Kleene algebra with converse [6] is a tuple (A, +, ·, ∗ , ◦ , 0, 1), where (A, +, ·, ∗ , 0, 1) is a Kleene algebra and the unary operation ◦ is a contravariant involution that distributes with + and ∗ : a◦◦ = a, (35) ◦ ◦ ◦ (a + b) = a + b , (36) (ab)◦ = b◦ a◦ , (37)
a∗◦ = a◦∗ , a + aa◦ a = aa◦ a.
(38) (39)
It is easy to show that the converse is monotonic and that 1 = 1◦ and 0 = 0◦ . Kleene algebras with converse allow the definition of partial and total functions, injections or surjections. We will need functions in section 7. An element a of a Kleene algebra A is total or entire, if 1 ≤ aa◦ , functional or simple, if a◦ a ≤ 1 and a function, if it is total and functional. From these definitions one can derive many of the usual properties of functions. We need only two of them. Lemma 5. The properties simple, entire, map are preserved by composition. Lemma 6. Let A be a Kleene algebra with converse. For all a, b, f ∈ A and f a function, af ≤ b ⇔ a ≤ bf ◦ ,
(40)
f ◦ a ≤ b ⇔ a ≤ f b.
(41)
The following corollary, interpreted in the relational model according to theorem 3, yields the Church-Rosser statement of theorem 1, using the relational abstraction of lemma 1. Corollary 2. Let A be a Kleene algebra with converse. For all a ∈ A, a◦∗ a∗ ≤ a∗ a◦∗ ⇔ (a + a◦ )∗ ≤ a∗ a◦∗ . Proof. Let b = a◦ in theorem 4 (i).
2
A point-free induction-based proof of this theorem in the relational model has been given in [20]. Similarly, the Church-Rosser theorems for three generators specialize to a variant of a Church-Rosser theorem modulo a congruence. Corollary 3. Let A be a Kleene algebra with converse. For all a, e ∈ A, ea∗ ≤ a∗ e∗ , a◦∗ a∗ ≤ a∗ a◦∗ , a◦∗ e∗ ≤ e∗ a◦∗ ⇒ (a + e + a◦ )∗ ≤ a∗ e∗ a◦∗ . Proof. Let n = 2, a0 = a, a1 = e and a2 = a∗ in theorem 2 (i).
2
The transition from Church-Rosser theorems without to those with converse did not use any algebraic properties of the converse. It worked by mere substitution. Thus corollary 2 and its extension to Church-Rosser modulo are simple instances of proposition 1 and 2. However, the presented properties of converse are used in the following section.
Calculating Church-Rosser Proofs in Kleene Algebra
7
285
The Church-Rosser Theorem of the λ-Calculus
In the Church-Rosser theorem of the λ-calculus, the hypothesis of our ChurchRosser theorem is shown. In this section, we reconstruct the proof in Barendregt [2] (pp.279-283). There some bridge lemmata deal with the term structure. We abstract them such that large parts of the proofs are again Kleene algebra. Let Λ be the set of λ-terms on a set of variables X and let →β be the associated reduction relation. s[t/x] denotes substitution of a term t for all occurrences of the variable x in a term s. The proof in [2] is based on redex-indexing to trace their behavior in reduction sequences. Intuitively, an indexed λ-term is obtained from a λ-term by indexing some of its redices (λx.s)t as (λi x.s)t in N. Λ denotes the set of indexed λ-terms. Obviously Λ ⊆ Λ . We extend →β to →β =→β0 + →β1 on indexed lambda terms such that →β0 acts on the indexed and →β1 on the non-indexed part. We (recursively) define the mappings π : Λ −→ Λ that forgets all indices and σ : Λ −→ Λ that β-reduces all indexed redices in a term from inside out: σ(x) = x σ(λx.s) = λx.σ(s) σ(st) = σ(s)σ(t), if s = λi x.s , σ((λi x.s)t = σ(s)[σ(t)/x]. We need types for modeling →β , →β , π and σ in Kleene algebra. These can be borrowed, for instance from allegories [10], see also [24]. Here, we leave the type-reasoning implicit. In [2] there are three relevant properties of λ-terms. 1. There is no difference between performing a β -reduction on an indexed term and forgetting the indices then or forgetting the indices and then performing a β-reduction (Lemma 11.1.6 (i) in [2]). This is expressed algebraically as π →β = →β π.
(42)
2. Every β -reduction of an indexed term followed by a σ-application can be simulated by a σ-application followed by a β-reduction. This is expressed as →β σ ≤ σ →β .
(43)
By (41) this is equivalent to σ ◦ →β σ ≤ →β (Lemma 11.1.7 (i), (ii) in [2]). 3. All σ-applications to indexed terms can be simulated by forgetting the indices followed by β-reductions. This is expressed as σ ≤ π →∗β
(44)
By (41) this is equivalent to π ◦ σ ≤ →∗β (Lemma 11.1.8 in [2]). It is easy to check that these bridge lemmata are well-typed. The remainder of the Church-Rosser theorem is then simple Kleene algebra, which again precisely reconstructs the usual diagrammatic reasoning.
286
G. Struth
Lemma 7. Let A be a (typed) Kleene algebra with converse and assume that (42), (43) and (44) holds for →β , →β , π, σ ∈ A. π →∗β = →∗β π,
(45)
σ ◦ →∗β σ ≤ →β ,
(47)
→∗β
(46)
π ≤ σ ←β .
(48)
σ≤σ
→∗β ,
Proof. (45) and (46) follow immediately from (42) and (43) by (17). (47) immediately follows from (46) by (41). For (48), note that π ◦ ≤ →∗β σ ◦ is equivalent to (44) by (40). Thus π = π ◦◦ ≤ (→∗β σ ◦ )◦ = σ ←∗β 2
by properties of the converse.
The following is an algebraic version of the strip lemma (lemma 11.1.9 in [2]). Proposition 4 (Strip Lemma). Under the assumptions of lemma 7, ←β →∗β ≤ →∗β ←∗β . Proof. Index the redex where the one-step expansion occurs. At this point, therefore, →β = π ◦ σ (which is consistent with (44)). ←β →∗β = (π ◦ σ)◦ →∗β = σ ◦ π →∗β = σ ◦ →∗β π ≤ σ ◦ →∗β σ ←∗β ≤ →∗β ←∗β , by the assumption, properties of the converse, (45), (48) and (47).
2
Corollary 4 (Church-Rosser). Under the assumptions of lemma 7. (→β + ←β )∗ ≤ →∗β ←∗β . Proof. By the strip lemma (lemma 4) and the abstract Church-Rosser result of corollary 2, which itself is a corollary of proposition 1. 2
8
More Church-Rosser Calculations
In this section we calculate the abstract part of the Church-Rosser theorem of the λ-calculus along the lines of the Tait-Martin-L¨ of (c.f. [2]) and Takahashi [26] method. This method is nowadays considered standard, in particular for the various approaches to mechanized proof checking, notice only [21,17,18]. Our previous proof uses more diagrammatics, the standard one does more work at the term level. The standard proof is shorter, whereas that of the last section— according to Barendregt—is more perspicuous. A key distinction at the abstract level is that all previous proofs use induction and therefore are essentially higherorder, whereas ours is strictly first-order and often even at the level of finite automata. Nipkow’s Isabelle implementation [18] is most concerned with the distinction between the abstract and the term level and therefore closest to
Calculating Church-Rosser Proofs in Kleene Algebra
287
ours. He uses a logical definition of a square diagram for proving some simple point-free lemmata for abstract reasoning. The properties of the Kleene star that he needs are probably derived from Isabelle theories about relations, although this is not explicitly mentioned in the text. We now show that all abstract statements needed for Church-Rosser proofs (c.f. [18]) with the Tait-Martin-L¨ of and Takahashi method are again simple calculations in Kleene algebra. We fist show two simple auxiliary facts. Lemma 8. Let A be a Kleene algebra. For all a, b ∈ A, b∗ = (b + 1)∗ ,
(a + b)∗ = (a∗ + b∗ )∗ .
(49)
(50)
Proof. For (49), (b + 1)∗ = b∗ (1b∗ )∗ = b∗ b∗∗ = b∗ b∗ = b∗ , by (19), (16) and (14). For (50), (a∗ + b∗ )∗ = a∗∗ (b∗ a∗∗ )∗ = a∗ (b∗ a∗ )∗ = (a + b∗ )∗ by (19) and (16). 2 Elimination of b∗ is similar. The next statement is essential for the Church-Rosser proof using the TaitMartin-L¨ of and Takahashi method (lemma 6 in [18]). Lemma 9. Let A be a Kleene algebra with converse. For all a, b ∈ A. a◦ a ≤ aa◦ , b ≤ a, a ≤ b∗ ⇒ b◦∗ b∗ ≤ b∗ b◦∗ . Proof. We first show that b ≤ a and a ≤ b∗ imply a∗ = b∗ . First b∗ ≤ a∗ follows from b ≤ a by monotonicity of the Kleene star. Moreover a∗ ≤ b∗∗ = b∗ follows from a ≤ b∗ by monotonicity of the Kleene star and (16). Applying (17) twice to a◦ a ≤ aa◦ yields a◦∗ a∗ ≤ a∗ a◦∗ . Moreover a∗ = b∗ implies a◦∗ = b◦∗ by monotonicity of the converse and (38). Therefore, replacing equals by equals, b◦∗ b∗ ≤ b∗ b◦∗ . 2 The next statement collects some alternatives to the diamond property. Lemma 10. Let A be a Kleene algebra. For all a, b, c ∈ A, (i) ab ≤ c(a + 1), b ≤ c ⇒ (a + 1)b ≤ c(a + 1), (ii) ba ≤ a∗ (b + 1) ⇒ b∗ a∗ ≤ a∗ b∗ , (iii) ba ≤ (a + 1)(b + 1) ⇒ b∗ a∗ ≤ a∗ b∗ . Proof. (ad i) (a+1)b = ab+b ≤ c(a+1)+b ≤ c(a+1)+c = c(a+1+1) = a(a+1), by distributivity, the assumption, monotonicity and idempotence. (ad ii) ba ≤ a∗ (b + 1) ⇒ (b + 1)a ≤ a∗ (b + 1) ∗
∗∗
by (i), (15)
⇒ (b + 1)a ≤ a (b + 1) ⇒ (b + 1)∗ a∗ ≤ a∗ (b + 1)∗
by (16), (17)
⇒ b∗ a∗ ≤ a∗ b∗
by(49)
by (17)
(ad iii) Since (a + 1) ≤ a∗ by (15) and (49) this follows immediately from (ii) and monotonicity of multiplication. 2
288
G. Struth
Lemma 10 (ii) (and (iii)) is known as the Hindely-Rosen commutation lemma (lemma 5 in [18]). Finally, we also calculate the Hindley-Rosen commutative union theorem (part (ii) of lemma 11, lemma 8 in [18]), which serves to show the Church-Rosser property in presence of β- and η-reduction. Lemma 11. Let A be a Kleene algebra with converse. For all a, b, c ∈ A, (i) c∗ a∗ ≤ a∗ c∗ , c∗ b∗ ≤ b∗ c∗ ⇒ c∗ (a + b)∗ ≤ (a + b)∗ c∗ , (ii) a◦∗ a∗ ≤ a∗ a◦∗ , b◦∗ b∗ ≤ b∗ b◦∗ , a◦∗ b∗ ≤ b∗ a◦∗ ⇒ ⇒ (a + b)◦∗ (a + b)∗ ≤ (a + b)∗ (a + b)◦∗ . Proof. (ad i) For c∗ (a + b)∗ ≤ (a + b)∗ c∗ , it suffices to show c∗ (a∗ + b∗ )∗ ≤ (a∗ + b∗ )∗ c∗ by (50) and c∗ a∗ + c∗ b∗ ≤ a∗ c∗ + b∗ c∗ by (17) and distributivity. This last inequality follows immediately from the hypotheses and monotonicity of addition. (ad ii) First note that (a + b)◦∗ = (a◦∗ + b◦∗ )∗ by (36) and (50). Now for (a + b)◦∗ (a + b)∗ ≤ (a + b)∗ (a + b)◦∗ it suffices to show (a◦∗ + b◦∗ )∗ (a + b)∗ ≤ (a + b)∗ (a◦∗ + b◦∗ )∗
(51)
by (17). Instantiating (ii) with c = a◦ and c = b◦ , respectively, yields a◦∗ a∗ ≤ a∗ a◦∗ , a◦∗ b∗ ≤ b∗ a◦∗ ⇒ a◦∗ (a + b)∗ ≤ (a + b)∗ a◦∗ , b◦∗ a∗ ≤ a∗ b◦∗ , b◦∗ b∗ ≤ b∗ b◦∗ ⇒ b◦∗ (a + b)∗ ≤ (a + b)∗ b◦∗ .
(52) (53)
Clearly, the hypotheses of (52) and (53) are hypotheses of (ii). In particular b◦∗ a∗ = b∗◦ a∗◦◦ = (a∗◦ b∗ )◦ = (a◦∗ b∗ )◦ ≤ (b∗ a◦∗ )◦ = (b∗ a∗◦ )◦ = a∗◦◦ b∗◦ = a∗ b◦∗ follows from a◦∗ b∗ ≤ b∗ a◦∗ by properties of the converse. (51) then follows from monotonicity of addition and distributivity. 2 Like in previous sections, all proofs are again simple, short and completely calculational. They cover precisely the abstract part of reasoning for the Tait-MartinL¨of and Takahashi proof and precisely reconstruct the usual diagrammatic reasoning. Compared to the proof in section 7, the arguments at the term level are more involved here. The key idea is as follows: Use lemma 9 to define a relation →l between →β and →∗β with the diamond property. Then →∗β has the Church-Rosser property. →l is a kind of parallel nested β-reduction. For proving its diamond property, the Tait-Martin-L¨ of method uses induction on →l and a case analysis on the relative positions of reductions. The Takahashi method uses a complete development to recursively contract all redices in a term. All this is beyond the applicability of Kleene algebra. We refer the reader to the literature cited above. For the Church-Rosser theorem for β and η reduction, it remains to show ∗ the diamond property of →∗η and semicommutation of →◦∗ β with →η , in order to apply the Hindley-Rosen commutative union lemma. This is again by induction on the term structure, using the Hindely-Rosen commutation lemma. The calculation of the Hindely-Rosen commutation lemma and commutative union lemma are interesting in their own right. By the results of this section
Calculating Church-Rosser Proofs in Kleene Algebra
289
we see that Kleene algebra is not restricted to one particular proof technique, but can handle the abstract part of the classical Church-Rosser proofs. It might even be the case that Kleene algebra suffices to prove all properties of abstract reduction or term rewrite systems that do not involve well-foundedness.
9
Discussion
The Church-Rosser theorems in section 5 are interesting for the foundations of rewriting. Theorem 4 is the Church-Rosser theorem for rewriting with quasiorderings [16], proposition 3 that for rewriting with non-symmetric transitive relations [23]. The commutation and semicommutation relations in the hypotheses are not only relevant to Church-Rosser theorems. They can be used to express independence or precedence in execution sequences or Mazurkiewicz traces, in imperative and concurrent programms [14,15,4,7]. Kleene algebra is related to certain allegories (c.f. [10]). Allegories can be understood either as categories with relations as arrows or as typed or heterogeneous relation algebras. In particular, semicommuting relational diagrams play the role of categorical functional diagrams. Also the Church-Rosser diagrams are nothing but semicommuting allegorial diagrams via their Kleene algebra semantics. See [24] for further discussion. This suggests to develop a generic machinery for compiling diagrammatic rewriting proofs to Kleene algebra or similar structures for proof checking or, conversely, visualizing (machine generated) proofs in these structures in terms of diagrams. Another natural question is how to calculate for instance Newman’s lemma or the Church-Rosser theorem modulo in Kleene algebra. In [24] we use allegorial tabulation techniques for a Kleene algebra extended by an ω -operation, to model infinite iteration and well-foundedness. A proof of Newman’s lemma in a different extension of Kleene algebra appears in [8]. We believe that Newman’s lemma is an excellent test example for extensions of Kleene algebra to infinite behavior. Finally, Church-Rosser proofs in Kleene algebra are also well-suited for automation. In [25] we have specified Kleene algebra in the Isabelle proof-checker and formalized all proofs up to theorem 4. It might be even more interesting to implement Kleene algebra in an automated first-order prover coupled with a decision procedure for automata.
References 1. J.-R. Abrial. The B-Book. Cambridge University Press, 1996. 2. H. P. Barendregt. The Lambda Calculus. Studies in Logic and the Foundations of Mathematics. North-Holland, revised edition, 1984. ´ 3. S. L. Bloom and Z. Esik. Iteration Theories. EATCS Monographs on Theoretical Computer Science. Springer-Verlag, 1991. 4. E. Cohen. Separation and reduction. In R. Backhouse and J. N. Oliveira, editors, Proc. of Mathematics of Program Construction, 5th International Conference, MPC 2000, volume 1837 of LNCS, pages 45–59. Springer-Verlag, 2000.
290
G. Struth
5. J. H. Conway. Regular Algebra and Finite State Machines. Chapman and Hall, 1971. ´ 6. S. Crvenkoviˇc, I. Dolinka, and Z. Esik. The variety of Kleene algebras with conversion in not finitely based. Theoretical Computer Science, 230:235–245, 2000. 7. V. Diekert and G. Rozenberg, editors. The Book of Traces. World Scientific, 1995. 8. H. Doornbos, R. C. Backhouse, and J. van der Woude. A calculation approach to mathematical induction. Theoretical Computer Science, 179:103–135, 1997. 9. S. Eilenberg. Automata, Languages and Machines, volume A. Academic Press, 1974. 10. P. Freyd and A. Scedrov. Categories, Allegories. North-Holland, 1990. 11. D. Harel, D. Kozen, and J. Tiuryn. Dynamic Logic. MIT Press, 2000. 12. J.-P. Jouannaud and H. Kirchner. Completion of a set of rules modulo a set of equations. SIAM J. Comput., 15:1155–1194, 1986. 13. D. Kozen. A completeness theorem for Kleene algebras and the algebra of regular events. Information and Computation, 110(2):366–390, 1994. 14. D. Kozen. Kleene algebra with tests and commutativity conditions. In T. Margaria and B. Steffen, editors, Proc. of TACAS’96, volume 1055 of LNCS, pages 14–33. Springer-Verlag, 1996. 15. D. Kozen and M.-C. Patron. Certification of compiler optimizations using Kleene algebra with tests. In J. Lloyd, V. Dahl, U. Furbach, M. Kerber, K.-K. Lau, C. Palmadessi, L.-M. Pereira, V. Sagiv, and P.-J. Stuckey, editors, 1st International Conference on Computational Logic (CL2000), volume 1861 of LNAI, pages 568– 582. Springer-Verlag, 2001. 16. J. Levy and J. Agust´ı. Bi-rewrite systems. J. Symbolic Comput., 22:279–314, 1996. 17. J. McKinna and R. Pollack. Some lambda calculus and type theory formalized. J. Automated Reasoning, 23(3–4):373–409, 99. 18. T. Nipkow. More Church-Rosser proofs (in Isabelle/HOL). J. Automated Reasoning, 26(1):51–66, 2001. 19. V. N. Redko. On defining relations for the algebra of regular events. Ukrainian Mathematical Journal, 16:120–126, 1964. 20. G. W. Schmidt and T. Str¨ ohlein. Relations and Graphs: Discrete Mathematics for Computer Scientists. EATCS Monographs on Theoretical Computer Science. Springer-Verlag, 1993. 21. N. Shankar. A mechanical proof of the Church-Rosser theorem. Journal of the ACM, 35(3):475–522, 1988. 22. J. M. Spivey. Understanding Z. Cambrigde University Press, 1988. 23. G. Struth. Canonical Transformations in Algebra, Universal Algebra and Logic. PhD thesis, Institut f¨ ur Informatik, Universit¨ at des Saarlandes, 1998. 24. G. Struth. Church-Rosser proofs in Kleene algebra and allegories. Technical Report 146, Institut f¨ ur Informatik, Albert-Ludiwigs-Universit¨ at Freiburg, 2001. http://www.informatik.uni-freiburg.de/∼struth/publications.html. 25. G. Struth. Isabelle-specification and proofs of Church-Rosser theorems. http://www.informatik.uni-freiburg.de/∼struth/publications.html, 2001. 26. M. Takahashi. parellel reductions in λ-calculi. Information and Computation, 118(1):120–127, 1995.
On the Definition and Representation of a Ranking Kim Cao-Van and Bernard De Baets Department of Applied Mathematics, Biometrics and Process Control Ghent University, Coupure links 653, B-9000 Gent, Belgium {Kim.CaoVan,Bernard.DeBaets}@rug.ac.be
Abstract. In this paper, we discuss how a proper definition of a ranking can be introduced in the framework of supervised learning. We elaborate on its practical representation, and show how we can deal in a sound way with reversed preferences by transforming them into uncertainties within the representation. Keywords: Ranking, Reversed preference, Supervised learning.
1
Introduction
Supervised learning has been studied by many research groups, largely coming from statistics, machine learning and information systems science. In these studies, the problems of classification (discrete) and regression (continuous) have received a lot of attention. More recently, the problem of ranking has made its appearance on the scene because of the wide variety of applications it can be used for. Ranking can be interpreted as monotone classification or monotone regression. The addition of the word “monotone” to the definition is, however, less trivial than it seems. And the problems this addition to the definition entails in the mathematical model used to deal with classification or regression are even more persistent. In this paper, we will focus on discrete models, in other words, we will discuss how we can deal with monotone classification, which is equivalent to monotone ordinal regression.
2
Problems with Earlier Proposals
The aim of supervised learning is to discover a function f : Ω → D based on a finite set of example pairs (a, f (a)) with a ∈ Ω. If D is finite, then f is referred to as a classification. Generally, the objects a ∈ Ω are described by means of a finite set Q = {q1 , . . . , qn } of attributes q : Ω → Xq . Therefore, to each a ∈ Ω corresponds a vector a = (q1 (a), . . . , qn (a)) ∈ X = q∈Q Xq (called the measurement space), and the problem is then restated as learning the function f based on examples (a, f (a)) with a ∈ Ω. Although this new definition is not less restrictive if handled with care, it does tend to encourage a H. de Swart (Ed.): RelMiCS 2001, LNCS 2561, pp. 291–299, 2002. c Springer-Verlag Berlin Heidelberg 2002
292
K. Cao-Van and B. De Baets
more narrow view, where f (a) is interpreted as f (a). This may lead to conflicting situations, since it is possible that a, b ∈ Ω, a = b but f (a) = f (b). We will use the term doubt to refer to such a situation. The problem of ranking is generally formulated as a classification problem in the narrow view, with the additional restriction that it has to be monotone, i.e. for all x, y ∈ X we must have that x ≤X y implies f (x) ≤D f (y), where (x1 , . . . , xn ) ≤X (y1 , . . . , yn ) if and only if xi ≤Xqi yi for i = 1, . . . , n, and the relations ≤Xq on Xq and ≤D on D are complete orders. Again, conflicting situations may arise, which we will refer to as reversed preference (see Section 6). Some authors [4,5] impose some additional restrictions, such as demanding the training data to fulfill the monotonicity requirement, to make sure these conflicts do not occur anymore. Others [1] propose a form of naive conflict resolution. However, a fundamental flaw in this definition is that it is formulated as a restriction not on the original definition f : Ω → D of a classification, but on its operationalization f : X → D, which was introduced in function of the description of the objects. Yet another problem is that ranking is not merely a restriction, but can also be seen as a generalization of classification, in which the equality relation is replaced by an order relation. For example, in the formulation of a ranking given above, we see that “x ≤X y implies f (x) ≤D f (y)” is an extension of “x = y implies f (x) = f (y)”. It is well known that different points of view on a basic definition may lead to completely different extensions, so, if possible, the most intrinsic definition should be chosen. In our case, this means f : Ω → D is preferred.
3
Classification and Ranking
Thus, following the preceding discussion, we define a classification in Ω as the assignment of the objects belonging to Ω, to some element, called a class label, in a finite universe D, which we will call the decision space. The class labels can be identified with their inverse image in the object space Ω, where they constitute a partition. We will call these inverse images (object) classes. For any class label d ∈ D, we denote the corresponding class by Cd := f −1 (d), and Cl := {Cd | d ∈ D}. So, the set of all classifications in Ω stands in one-to-one correspondence to the set of all partitions of Ω (which is equivalent to the set of all equivalence relations on Ω). If we want to define a ranking based on this definition, it becomes clear that there is no room for a concept such as monotonicity since Ω has no inherent structure such as X . Still, we have to plant the seeds for it, such that monotonicity will appear naturally when the measurement space X is introduced as a representation of Ω. This can be done, following the ideas from [3,6], by returning to the semantics behind ranking, which declares that the higher an object’s rank, the more it is preferred. We can model this preferential information by a complete preorder. A preorder R on Ω is a binary relation R on Ω that is reflexive and transitive. It is called complete if for all a, b ∈ Ω we have aRb or bRa. A preorder can be seen
On the Definition and Representation of a Ranking
293
as a special kind of weak (also called large) preference relation [7] S (which is only reflexive), and where the expression aSb stands for “a is at least as good as b” (it is also said that “a outranks b”). Given a complete preorder S, we can define a strict preference relation P by aP b if and only if aSb and not bSa, and an indifference relation I by aIb if and only if aSb and bSa. We clearly have that S = P ∪ I. Definition 1. A ranking in Ω is a classification f : Ω → D, together with an order ≥d on D. We denote this ranking by (f, ≥d ). Moreover, the order ≥d defines a weak preference relation S on Ω as follows: (∀s ∈ D)(∀r ∈ D)(∀a ∈ Cs )(∀b ∈ Cr )(aSb ⇐⇒ s ≥d r), where Cd = f −1 (d) is the class associated with the class label d ∈ D. We will only consider complete rankings, where ≥d is a complete order on D, i.e. (D, ≥d ) is a chain. This is in line with most of the current problems considered in supervised learning. In this way, we have a specific preference structure on Ω linked with the classes. For a, b ∈ Ω, assume a ∈ Cs and b ∈ Cr with s, r ∈ D. Then we have aP b ⇐⇒ s >d r
and
aIb ⇐⇒ s = r.
So, the set of all rankings in Ω stands in one-to-one correspondence to the set of all weak preference relations S on Ω that are a complete preorder. The classes are formed by the indifference relation I which is an equivalence relation (transitivity holds since S is a complete preorder). Hence, the indifference relation I determines the classification.
4
Representing a Classification
The previous definitions are not really useful in practice since they relate to a universe Ω that is in essence just an enumeration of all the objects. To access some of the interesting properties of the objects, we fall back on a set of attributes Q. In this way, we can represent each object a ∈ Ω by a vector ⊆ X , where Ω is the set of all measurement veca = (q1 (a), . . . , qn (a)) ∈ Ω tors corresponding to objects in Ω. This leads also to a representation fˆ of the classification f : Ω → D in the following way: → 2D , fˆ : Ω x → fˆ(x) = {f (a) | a ∈ Ω ∧ a = x}, where 2D is the power set of D, i.e. the set of all subsets of D. Thus, the representation of a classification is again a classification, but now in the space ⊆ X , with classes CD := fˆ−1 (D), where D ⊆ D (we may define a classiΩ fication fˆ∗ : Ω → 2D by setting fˆ∗ (a) = fˆ(a))1 . Moreover, if Ω ∼ = X , then fˆ ∼ = f. 1
Since fˆ and fˆ∗ both express the same idea, we will not restrain ourselves of mixing their usage.
294
K. Cao-Van and B. De Baets
This property of isomorphism states that the representation fˆ is a very natural one. We certainly want to keep this property in the case of rankings, therefore we will denote the representation of a ranking by (fˆ, Z) where Z is some relation on 2D , that should be isomorphic to ≥d if Ω ∼ = X. Moreover, the more general observation that representing a classification results back into a classification is also a very desirable property. We know that the real problem is one of classification, but we will work with a representation, so it would be against our intuition that this representation would become something different from a classification. In the same line of thinking, we would like this property, if possible, to hold for rankings as well.
5
Representing a Ranking
Let (f, ≥d ) be a (complete) ranking. A first remark concerns the range of fˆ when we are dealing with rankings. Because of the ordinal nature of the class labels, it is only meaningful to attach intervals to objects. Therefore, we define → D[2] = {[r, r ] | (r, r ) ∈ D2 ∧ r ≤d r }, fˆ : Ω x → [fˆ (x), fˆr (x)],
(1)
where fˆ (x) = inf{f (a) | a ∈ Ω ∧ a = x}, fˆr (x) = sup{f (a) | a ∈ Ω ∧ a = x}. This entails that the relation Z should be defined on D[2] . There are some other restrictions we should impose. The relation Z should be (i) reflexive, to ensure that equal intervals are also treated equally, (ii) an extension of ≥d , as discussed previously, and (iii) meaningful, just as ≥d has a meaning in terms of a preference relation. So, we would like to have an interpretation such as fˆ(a)Z fˆ(b) ⇐⇒ aSb,
(2)
means: based on the information derived from Q and fˆ, we conclude where aSb that a is at least as good as b. Lastly, closely related to the previous, we would like that (iv) Z does not depend on f , but only on ≥d . Remark that condition (i) has an additional advantage since it enables us to interpret Z as a weak preference relation. Starting from (ii) and (iii), we will try to extend the semantics behind ≥d . The two following equivalent expressions are the most straightforward way (that enables generalization) to capture these semantics: s ≥d r ⇐⇒ (∀(a, b) ∈ Ω 2 )((a, b) ∈ Cs × Cr ⇒ aSb) ⇐⇒ (∃(a, b) ∈ Ω 2 )((a, b) ∈ Cs × Cr ⇒ aSb).
(3) (4)
On the Definition and Representation of a Ranking
295
These expressions can be easily generalized by replacing s ≥d r with D Z D , where D, D ∈ D[2] , and Cs (resp. Cr ) with CD = ∅ (resp. with CD = ∅). If we impose reflexivity, and write D = [s1 , s2 ], D = [r1 , r2 ], this finally results in [s1 , s2 ] Z1 [r1 , r2 ] ⇐⇒ s1 ≥d r2 , as a generalization of (3), and [s1 , s2 ] Z2 [r1 , r2 ] ⇐⇒ s2 ≥d r1 , as a generalization of (4). Relation Z1 is an order (if we write ≥D1 instead of Z1 , we simply have [r1 , r2 ] ≤D1 [s1 , s2 ] ⇐⇒ r2 ≤d s1 ), and relation Z2 is an interval order2 . It is clear that these two relations fulfill conditions (i), (ii) and (iv). However, in both cases there are some problems with condition (iii). We have that [1, 3] and [1, 2] are uncomparable w.r.t. Z1 . However, we would prefer an object with label [1, 3] over another with label [1, 2] if that is all we know about these objects, so, following expression (2), we want [1, 3]Z[1, 2]. Whereas Z1 is too restrictive, the relation Z2 is too loose, being indifferent between [1, 2] and [1, 3]. This means we need to find something in between the generalizations of expressions (3) and (4). Yet another approach, guided by the previous observations, is to first state the desired semantics and to translate them afterwards into expressions of the right form. In that way we define D Z D if and only if D is an improvement over D or D is a deterioration compared to D. We will now translate this into mathematical expressions. Assuming that CD and CD are non-empty, we say that D is an improvement of D if and only if (∃a ∈ Ω)(∀b ∈ Ω)((a, b) ∈ CD × CD ⇒ aSb) (5) (∀a ∈ Ω)(∃b ∈ Ω)((a, b) ∈ CD × CD ⇒ aSb) . Likewise, D is a deterioration of D if and only if (∃b ∈ Ω)(∀a ∈ Ω)((a, b) ∈ CD × CD ⇒ aSb) (∀b ∈ Ω)(∃a ∈ Ω)((a, b) ∈ CD × CD ⇒ aSb) .
(6)
It immediately strikes that all these expressions can be seen as intermediate to the generalizations of expressions (3) and (4). If we write D = [s1 , s2 ] and D = [r1 , r2 ], it can be shown quite easily that (5) ⇐⇒ (6) ⇐⇒ ((r1 ≤d s1 ) ∧ (r2 ≤d s2 )) . Hence, we find that Z is an order, which we will denote by ≥D , defined as follows: Definition 2. Let D, D ∈ D[2] . If we write D = [r1 , r2 ] and D = [s1 , s2 ], we put [r1 , r2 ] ≤D [s1 , s2 ] ⇐⇒ ((r1 ≤d s1 ) ∧ (r2 ≤d s2 )) . 2
An interval order is a reflexive, complete and Ferrers relation. A relation R on X is called Ferrers if (aRb ∧ cRd) ⇒ (aRd ∨ cRb), for any a, b, c, d ∈ X. The Ferrers property can be seen as a relaxation of transitivity.
296
K. Cao-Van and B. De Baets
It is clear that now all four conditions are met. It should be noted that the order ≤D was derived from the premise that we only have access to intervals of values to reach a decision. If other information would be available, other orderings might prevail. For example, distributional information might lead to a stochastic ordering, or if risk aversion underlies the decision, then we would end up with a complete order based on leximin. Finally, remark that the order ≤D is well known and turns (D[2] , ≤D ) into a complete lattice3 .
6
The Monotonicity Constraint
Up to now, we have given a definition of a (complete) ranking (f, ≥d ) and have shown how it is represented by the (not necessarily complete) ranking (fˆ, ≥D ). Note that we did not need any form of monotonicity for the definition or the representation. Monotonicity will arise in a natural way when taking into account the attributes used to describe the properties of objects. Let us first turn back to classifications. Even if we assume a classification f to be deterministic in the sense that any object a ∈ Ω is assigned to exactly one class with label in D, we still cannot guarantee that we have fˆ∗ (a) ∈ D for all a ∈ Ω. This is a consequence of the possible occurrence of doubt. Definition 3. (i) We say there is doubt between the classification f and the set of attributes Q if (∃(a, b) ∈ Ω 2 )(a = b ∧ f (a) = f (b)). (ii) We say there is doubt inside the representation fˆ if fˆ(x)| > 1). (∃x ∈ Ω)(| It is clear that these two notions of doubt coincide. In the case of doubt, the function fˆ∗ can assign a set of labels to an object, denoting it is not possible to label the object with one specific class label based on the measurement vector. Remark that this does not result in a conflict, as the one discussed in Section 2, when using the representation fˆ. In the context of ranking, the attributes have a specific interpretation, and are usually referred to as criteria. A criterion is defined as a mapping c : Ω → (Xc , ≥c ), where (Xc , ≥c ) is a chain, such that it appears meaningful to compare two objects a and b, according to a particular point of view, on the sole basis of their evaluations c(a) and c(b). In this paper, we will only consider true criteria, where the induced weak preference relation is a complete preorder defined by aSc b ⇔ c(a) ≥c c(b). We assume to have a finite set C = {c1 , . . . , cn } at our disposal. 3
Because this order is the best known order on D[2] , it is the most evident one to consider. In that spirit, Potharst [5] also introduced this order in the setting of rankings, however, without being aware of its semantics.
On the Definition and Representation of a Ranking
297
The dominance relation4 on Ω w.r.t. C is defined by (∀c ∈ C)(aSc b) a b ⇐⇒ (∃c ∈ C)(aPc b) for any a, b ∈ Ω. It is said that a dominates b. We may also write b a, saying that b is dominated by a. We say that a weakly dominates b, a b, if only the first condition holds. Since we are working with true criteria we have that a b is equivalent with a ≥X b. A basic principle coming from Multicriteria Decision Aid (MCDA) [6] is that a b ⇒ aSb. On the other hand we have that aSb ⇐⇒ f (a) ≥d f (b). Merging all these expressions we find in a natural way the monotonicity constraint a ≥X b ⇒ f (a) ≥d f (b) . Remark that this constraint does not tolerate the presence of doubt since it advocates a = b ⇒ f (a) = f (b). Thus, it is too restrictive for applications in supervised learning. The reason for this lies in the fact that we have adopted a principle from MCDA without considering its context: build a ranking based on the set C of criteria. This is a different setting than for supervised learning where we try to reconstruct a ranking based on the set C. In the former, the set C is a framework, in the latter, this same set C is a restriction. We can solve this by applying the same principle but with the additional demand that we restrict our knowledge to the information we can retrieve from C. In that case, we define ⊆ X , resulting in x y if and only if x ≥X y, and the dominance relation on Ω ˆ Together with (2), this finally leads to the the principle becomes x y ⇒ xSy. monotonicity constraint x ≥X y ⇒ fˆ(x) ≥D fˆ(y) . In this case, doubt is tolerated since x = y ⇒ fˆ(x) = fˆ(y) is a trivial demand, regardless of the cardinality of fˆ(x). This also means that the monotonicity constraint reduces to x >X y ⇒ fˆ(x) ≥D fˆ(y) . (7) Definition 4. We say there is reversed preference inside the representation (fˆ, ≥D ) if 2 )(x >X y ∧ fˆ(x)