Toward a systematic conflict resolution framework for ontologies

We first distinguish between meaning negotiation and conflict resolution. Subsequently, we introduce a preliminary ‘library’ of conflicts and the conflict set data structure that stores the minimum necessary data about such conflicts to be used for resolution, and then proceed to resolution strategies.

Characterizing meaning negotiation and conflict resolution

Negotiating the meaning of the knowledge—classes, properties, and axioms—to be represented in an ontology involves reaching an agreement on: 1) the exact elements required, 2) the domain theory that will provide these elements, and 3) the required ontology language to represent the former. Each item may involve meaning negotiation and conflict resolution. This will be disambiguated and illustrated first, after which potential sources of conflict are identified and the conflict set is introduced.

Types and sources of conflicts

We define informally the concepts of meaning negotiation and conflict resolution, which will be illustrated afterward. Meaning negotiation concerns deliberations to figure out the precise semantics one wants to represent in the ontology. They are all positive choices in the sense of “which of the options is applicable? Then we take that one”. Conflict resolution concerns choosing one option among a set of two or more options, where that choice is deemed the ‘lesser among evils’ for that scenario. It necessarily involves a compromise and making it work requires reengineering something in at least one of the ontologies or as a whole. Subtypes include: Language conflict resolution A conflict arises within the same family of languages or with a m ore distant one. This is either a zero-sum game (i.e., with a winner and a loser) or there may be a joint outside option. Ontological conflict resolution The ontologies adhere to different theories, which may be foundational philosophical decisions that affect the overall structure of the ontology or subject domain arguments with competing theories. This is likely a zero-sum game of either-or (no joint outside option).

They are illustrated in the following example.

Example 2

Meaning negotiation may involve assistance with explanations for the modeler, such as when they do not know whether to give up reflexivity for scalability, to offer them a dialogue alike “if you don’t have reflexivity, you’ll miss these sort of inferences: [example goes here]”, or one can frame negotiation of alternatives as an imperative, e.g., “take either parthood or proper parthood as primitive for your mereological theory, but not both.”.

Conflict resolution applies in many cases; e.g., there are several common types of language conflicts, such as:

A conflict within a language family, such as the Description Logics-based OWL species, is transitivity versus qualified cardinality constraints, because one cannot have both declared on the same object property, as illustrated in the Introduction with biped and has - part.

A syntax-level conflict, which manifests itself when having to merge an ontology represented in the common logic interchange format (CLIF) and another one represented in OWL, or OBO and OWL.

A conflict about a language’s semantics (on top of the syntax issues): when ontologies are represented in different languages where one has a model-theoretic semantics and the other a graph-based one, or even the same overall semantics but one has the unique name assumption (UNA) embedded in the language and the other does not.

What to do then? Besides choosing either, there may be a so-called ‘joint outside option’ (a term from game theory) where neither wins, but there is an alternative option. For instance, instead of debating over transitivity vs qualified cardinality, leave OWL to choose CLIF, or when deliberating CLIF or OWL in an either-or way, one can keep both and move outside either setting and into the framework of the Distributed Ontology, Model and Specification Language (DOL) [17], which is a meta-language that provides the ‘glue’ between ontologies represented in different languages, including more expressive ones. This strategy was illustrated in [5] for mereotopological theories.

The language’s semantics can be an example of a zero-sum game, such as to either embed the UNA in the semantics or not, but it cannot be both, and it affects the reasoner’s deductions. For instance, consider some ontology O1 that contains \(\mathsf \) together with the assertions . Under no-UNA such as with OWL, it will deduce b1=b2, because it is the only way to satisfy the =1 constraint; under UNA, it will deduce the ontology is inconsistent because is violating the =1 constraint in the TBox.

An example of an ontological conflict is a clash in the top-level organization of the ontology, such as between BFO and GFO, and related philosophical differences, such as qualities with qualia vs trope theory to represent attributes [18]. At the subject domain level, this may be, e.g., whether a virus is a living thing or not and competing scientific theories more generally. They do not have a joint outside option because the theories conflict, except returning it to the domain experts to resolve (e.g., to conduct experiments).

What are the sources of such conflicts? The sources can be manifold and six principal cases were discerned, which are non-exclusive and possibly also not an exhaustive list:

Ontological differences between established theories; e.g., extensional mereology vs. minimal mereology and DOLCE vs. BFO as top-level ontology.

Ontological differences at the axiom-level; e.g., not all philosophers agree on whether the parthood relation is antisymmetric [19].

Different modeling styles; e.g., foundational ontology-inspired or conceptual model-influenced, resulting in, e.g., reifying verbs into classes or adding them as object properties, respectively, [20], like the Vaccination/vaccinates mentioned in Fig. 1.

Fig. 1figure1

Sample scenario (summarized) to detect and resolve conflicts in an ontology reuse scenario where the intention is to import ontology O2 into ontology O1

Logic limitations causing conflicts for an ontology, affecting the software ecosystem; e.g., the biped’s has - part being either transitive or have it participate in axioms with qualified cardinality constraints in OWL 2 DL, or facing this clash when trying to merge or integrate two ontologies.

Logic limitations by design, for the purpose of scalability; e.g., there are axioms in one’s ontology that are beyond the desired OWL species, so that one has to choose either to abandon the preferred species or to remove the axioms.

Certain deductions made by the reasoner (excluding modeling mistakes); e.g., an unsatisfiable class resulting from disjoint ancestors that were aligned. While this may also have as source an ontological difference at the axiom-level, it would manifest either after adding the axioms, during test-driven development (TDD) [13], or upon ontology matching attempts.

The first three are, in principle, a priori negotiations by an ontologist, but may manifest only upon ontology matching. Cases 4 and 5 emerge during ontology authoring. The last one may or may not be a priori. They will be elaborated on in the “Resolving conflicts” section further below.

The conflict set

Conflict detection offers opportunities for automation and, even though there is no single way of how conflicts can be detected, some tasks can be carried out with the aid of state-of-the-art ontology development environments (ODEs), which we elaborate on afterward and its requirements emanating from that in the “Discussion” section further below.

The data structure in which the detected conflicts are stored and upon which the resolution of conflicts operates, is called the conflict set, which is generated in all cases where a conflict is detected. Without loss of generality, it is assumed that, when matching more than two ontologies, a conflict set is generated for every pair of ontologies.

The conflict set is described in a context-free grammar in Backus-Naur Form, as shown in Fig. 2, for purposes of genericity, as such a grammar can be implemented easily in most programming and rule languages and enforce verification of correctness of any implementation of the conflict sets. Accordingly, there are two ontologies or two fragments of the same ontology, each identified by an IRI or another identifier and composed of a (possibly singleton) set of axioms. An axiom may adhere to an ontologically well-founded theory, such as ground mereology, or some domain theory. It is briefly illustrated in the following example with the Virus Infectious Diseases Ontology (VIDO) and the COVoc vocabulary, whereas a more comprehensive case is deferred to the case study in the “Results” section.

Fig. 2figure2

Conflict set grammar for recording individual conflict sets in or between ontologies (production rules of most terminals are omitted)

Example 3

The conflict set grammar of Fig. 2 is illustrated in ‘pretty printing’ notation with VIDO and COVoc, noting that we assume that \(\mathsf \sqcap \mathsf \sqsubseteq \bot \) is present in the ontology, following VIDO’s accompanying documentation [21] (though vido.owldoes not have that disjointness axiom), which is explained afterward:

Both ontologies are identified with their respective < IRI >, for one conflict: in virology (the value of the conflict set’s < theory > for VIDO), viruses are acellular structures and not organisms, as shown with the relevant < axiom > number, axiom itself, and < description > thereof. The < inference > over the combined < ontology > (the “\(\mathcal _ \cup \mathcal _\)” part in the box, above) with temporary name \(\mathcal _^\), and thus includes the two alignment axioms (the “ vido:virus≡covoc:Viruses and vido:organism≡covoc:Organism”), is that the respective virus classes are now unsatisfiable (\(} \sqsubseteq \bot \ \text \ } \sqsubseteq \bot \)) and thus the combined ontology is incoherent.

The explanation thereof, which assists toward resolution of this conflict at the subject domain level, will be elaborated on in the next section, and its Fig. 5 in particular.

Resolving conflicts

In practice, conflict resolution often starts with some issue raised by the ODE, and specifically when an axiom is added or an ontology is merged or integrated into the active ontology. Examples of such issues are undecidability, language profile violation, and incoherence. They can be seen as cues indicating that something is wrong with the active ontology. The author then has to find out what raised the issue. Thereby, they may be supported by the ODE. Proceeding that way is not as straightforward as one might expect, because there is no one-to-one correspondence between conflict and issue. Examples of such ‘causal investigations’ will be given in the following subsections. For the rest of this section, the following principal choices are presupposed:

The ontology author adheres to Occam’s razor when developing an ontology for the case at hand: the least expressive language in which the required axioms can be represented fully is preferred over all more expressive ones.

The ontology author wants to capture as much of the semantics of the domain theory as possible.

The ontology author prefers a decidable language over first or higher order logic for representing a domain theory and a coherent ontology over an incoherent one (that suffers from at least one unsatisfiable class).

The first choice is a general principle in many situations in life. The second choice assumes that the author prefers representing a full axiomatization over a partial axiomatization and, by extension, a partial one is better than mere primitives without any axioms. While the third choice may not hold in all situations, we deemed it realistic to include, since most software infrastructure caters for decidable ontology languages and coherent ontologies, and Semantic Web and Knowledge Graph languages in particular.

We now proceed to discuss the conflicts listed in Table 1 in their order of presentation.

Table 1 A selection of conflicts that may emerge during ontology authoring, as a preliminary library of conflictsTable 2 Differences between foundational ontologies (non-exhaustive, partially based on ONSETv1.2 [7])Conflicting top-level and domain theories

If an ontologically well-founded theory underlying some axioms to add or an ontology to integrate is in conflict with the ontology representing the desired theory (nos. 1–3 and 5 in Table 1), then the respective IRIs must be added to the conflict set. This presupposes that the conflict is known and the pair of IRIs is already listed somewhere, for instance, in a library of common conflicts. To give an example, if one wants to add the part-whole relations taxonomy [22] that is aligned with DOLCE to a BFO-aligned IDO [3], then the theory conflict ‘BFO vs. DOLCE’ will be detected through looking up the library of common conflicts. Conflict resolution, in this case, aims at preserving a consistent theory. Since for conflicting theories there is no joint outside option, the ontology author has to decide in favor of one theory and discard the other. Their decision may be informed by the deliberations of what should be represented in the ontology made during meaning negotiating.

ODEs support the import of ontologies and after import, their IRIs can be read from the metadata of the active ontology and looked up manually for common conflicts in a library. Accordingly, this conflict detection approach is straightforward to implement. Uncommon conflicts are harder to detect and therefore harder to resolve. The use of ontology design patterns [23] for a theory would be helpful in automating detection, as would annotations that describe which theory is being represented. In addition, the library of common conflicts may grow upon finding more conflicts, be this automatically or manually curated or both, so that it can prevent the same or a similar conflict from emerging later on in the project.

Taking a step back from practicalities like checking the IRIs, there may be ontological commitments at stake underlying differences in IRIs. A straightforward example of commitments and a foundational ontology that satisfies them would be “Universals, Realist, Concrete entities only”, which BFO meets [24]. One may have other preferences for commitments that do not agree fully. For instance, one may want an ontology for representing particulars, take a realist stance, yet also admit to a multiplicative approach where one admits to a difference between, e.g., the vase and the clay that it is made of. DOLCE and GFO, and SUMO each can only meet two of the three requirements, as illustrated in Fig. 3, and one then either has to change the requirements or accept that not all of them will be met and, in case of the latter option, which one is going to be traded in to compromise.

Fig. 3figure3

Explanation generated by ONSET [7] after having selected a combination of requirements for ontological commitments “Particulars, Realist, yet Multiplicative” that no foundational ontology can fully meet and whose conflict needs to be resolved at this requirements analysis stage

The top-level theory conflict listed in line No. 4 in Table 1 concerns ontologically irreconcilable differences that are embedded in the language that is used to represent the ontology in. This issue is described at length in [25] and its Table 2 lists the assessment of ontological features by ontology language, which can be applied for detecting conflicts. For instance, if one is convinced of the separation between natural language with labeling things one the one hand and the things themselves with their representation in logic on the other, then OBO is a good choice. Contrast this to OWL, which assumes the two are intertwined in that the classes and properties are expected to have human readable names. It is possible to convert one into the other, but it remains a brittle workaround due to the fundamental difference: practically, it acts out in ‘OBO style’ identifiers as names for classes and object properties in an OWL file with a hopeful wish that their labels are rendered in the tools. However, OWL assumes the elements are named and therefore labels are not mandatory, resulting in situations where the ‘OBO style forced into OWL’ occasionally breaks and one is left with only meaningless identifiers.

Conflicting theories at the subject domain level (nos. 5–6 in Table 1) may act out, and be detected and resolved, in two different ways. First, if each domain theory was represented in an ontology and had its own IRI, then one could use the same resolution approach as for the top-level theory issues of nos. 1–3 described above. Second, particularly in the case of fine-grained aspects like the status of an element, then they can be processed and resolved as for axiom-level conflicts, which will be described below and illustrated with a follow-up of Example 3.

Conflicts manifesting themselves in an undecidable language

With conflicts nos 7–8 in Table 1, conflict resolution aims at preserving a decidable ontology language or raising awareness of undecidability when opting for a joint outside option. In the first case, this most often is a zero-sum game: the ontology author has to chose which one(s) of the conflicting axioms in the conflict set to keep and which one(s) to delete. For mereotopological theories, these types of conflicts have been investigated [5]: in most instances, incorporating a full axiomatization renders the active ontology at least undecidable, and possibly also incoherent. To support the author’s decision, some criteria can be established for how to compute the possible resolution choices, such as the following:

The least number of axioms are affected;

The preferred axiom type is identified by assigning weights to classes, properties, or types of constraints;

The least number of inferences are lost.

The least number of axioms affected can be read from the conflict set that was computed in detecting the conflict. Assigning weights to axiom types implies that certain types are a priori considered more valuable than others; e.g., one may weigh existentially quantified properties in axioms more than universally quantified properties and unqualified cardinality more than qualified cardinality. The least number of inferences lost requires an additional step where the inferences of the ontologies are computed and recorded in the conflict set. If undecidability is caused by an ontological conflict at the axiom-level that was not resolved along with conflicting theories (e.g., weak vs. strong supplementation in Example 4), then also the decisions taken when negotiating meaning upfront may serve as a criterion. The authors’ decision and the criteria upon which it is based ideally should be recorded so as to keep track of the decision made and in case the same or a similar conflict emerges later on in the project.

The second case, i.e., opting for a joint outside option, requires that principal choice (iii) (preferring a decidable language) is relaxed. Theories that are represented in different logics can be dealt with by the DOL framework [17]. This includes cases where the resulting logic is undecidable.

State-of-the-art ODEs provide some support for detecting and resolving these kinds of conflicts. To give an example, the OWL API [26] of Protégé 5 [27] issues an error message reporting the conflict arising from a violation of the expressive OWL 2 DL specification, which may render the language undecidable, when it is caused by a non-admissible axiom combination such as those listed in Example 4. In addition, Protégé 5 is equipped with an OWL reasoner, and a diff tool for computing the differences between OWL ontologies is available as a plug-in [28]. In order to compute the number of inferences lost, the axioms inferred from the merged ontology are first computed using the OWL reasoner. This requires that the merged ontology is saved as two decidable versions by removing one conflicting axiom set in exchange for the other. The difference between the sets of inferred axioms is then computed and returned to the modeler to inform them about the consequence of that particular choice.

Example 4

The following three examples are a sampling of mutually exclusive axiom combinations where a modeler thus has to choose either one or the other but they cannot be both in the same ontology:

Within language family: an ontology expressed in OWL 2 DL where transitivity or a role chain is declared on some object property R prohibits the use of any of the following features to be declared on R as well, if the aim is to remain within OWL 2 DL: minimum cardinality, maximum cardinality, exact cardinality, functionality, inverse functionality, reflexivity, irreflexivity, asymmetry, and role disjointness.

Mereology: weak supplementation, expressed as pp(x,y)→∃z(p(z,y)∧¬o(z,x)), in the mereological theory called Minimal Mereology is incompatible with strong supplementation (¬p(y,x)→∃z(p(z,y)∧¬o(z,x))) in Extensional Mereology.

Temporal logics: one has to choose how to ‘see’ time, where it is either discrete such that there is a first and last time point t and a series of successive time points with no time point between any t and t+1 or it is dense such that ∀t,t′∈T,t<t′,∃t″.t<t″<t′, i.e., it is infinitely possible to squeeze another time point between two adjacent time points.

Conflicts manifesting themselves in a language profile violation

This may occur in particular with conflict no 8 of Table 1, but may also appear as a confounding issue in conflict 7. The case where resolution aims at preserving the language to the extent that it is decidable again, while accepting that the original profile is violated, has to be distinguished from the case where the original language profile should be preserved. Presupposing aforementioned principal choices, conflict resolution in the first case aims at capturing as much of the semantics of the domain theory as possible. Since the language profile that was violated may not be the most expressive one, there may be room for a (decidable) joint outside option. To give an example, the ‘overlap’ axiom O(x,y)=def∃z(P(z,x)∧P(z,y)) cannot be expressed in any decidable OWL species, because it does not admit property definitions in the language. That is, the syntax does not permit usage of equivalence for object properties and, in fact, even puts limitations on property subsumption so as to prevent cycles in the hierarchy (called a “regular” role hierarchy [29]). While preserving decidability, the author may still want to state that P(z,x)∧P(z,y) is a sufficient condition for O(x,y) or they may want to state that O(x,y) is a reflexive and symmetric property. Yet, doing so may violate the original language profile. Whether the ontology language still is undecidable with the modified axioms and conditions can be figured out in the same way as described in the previous section, about Conflicts Manifesting Themselves in an Undecidable Language. Weakening the theory step by step this way will end up in a representation such that the language features used remain within a decidable language, since the representation of properties as mere primitives is always possible in OWL and other ontology languages.

In the second case, conflict resolution aims at preserving the original language profile at the expense of relaxing principal choice (ii), i.e., accepting that ‘as much semantics as possible’ is less than anticipated. This applies to conflicts emerging from what is called design for scalability in the “Types and sources of conflicts” section, above. Here, conflict resolution is likely to be a zero-sum game similar to that described in the section about undecidability. For instance, the OWL 2 QL profile is aimed at applications that use very large amounts of data, such as conventional relational database systems, and where query answering is the most important reasoning task. Violating this profile then means accepting that query answering may no longer be implementable by rewriting queries into a standard relational query language [30].

Concerning tool support for these two cases, in the first case, it is the same as that described in the preceding section about undecidability. For the second case, the OWL Species Classifier supports authors of OWL ontologies by listing which axioms violate which OWL species (see footnote 1). An illustration of the benefits of this sort of automation is described in the following example, where the classifier was used to search through the 417 axioms of the CIDO ontology for COVID-19 [31] to check for profile violations, compared to manually searching for a possible needle in a haystack.

Example 5

Practically, medical ontologies for information systems typically do not exceed the OWL 2 EL profile, because of scalability and compatibility with typical OBO Foundry ontologies and SNOMED CT, so let us assume this as a requirement. CIDO [31] is not within OWL 2 EL, however, which is due to a class expression with a universal quantifier on the right-hand side; more specifically, as easily pinpointed by the OWL Species Classifier (see Fig. 4), ‘Yale New Haven Hospital SARS-CoV-2 assay’ \(\sqsubseteq \forall \)’EUA-authorized use at’.’FDA EUA-authorized organization’ is one of the axioms that violate the OWL 2 EL expressiveness restrictions not only in the initial unofficial cido.owl of 14 June 2020 but also still in the most recent release of 31-1-2021 (v1.0.181); see also Fig. 4.

Fig. 4figure4

Section of the OWL classifier, having detected that CIDO_0000020, i.e., Yale New Haven Hospital SARS-CoV-2 assay, is one of the class expressions violating the OWL 2 EL profile

Conflicts manifesting themselves in an incoherent ontology

Conflicts nos. 7–8 in Table 1 may also manifest themselves as an incoherent ontology, with one or more unsatisfiable classes or properties (observe that conflict no. 7 includes any remaining unresolved theory-level conflicts of nos. 1–3, 5, and 6). Conflict resolution in this case aims at preserving or obtaining a coherent ontology. Examples include ontological misspecifications at the axiom-level, such as disjoint ancestors, resulting in unsatisfiable classes (no. 7 in Table 1). Such conflicts manifest only when the automated reasoner reports the deductions; an example is shown in Fig. 5. In the simplest case, they are resolved by keeping some of the conflicting axioms and removing others in a way similar to that described in the section about undecidability, above. Typically, either the disjointness axiom on the ancestors or the subclass axioms on the class may be kept, but not both. State-of-the-art ODEs allow for making deductions. After running the OWL reasoner in Protégé 5.x, for instance, unsatisfiable classes and properties are displayed in red color. In order to find out what made them unsatisfiable, justifications can be computed using the respective plug-in, where a justification is a set of axioms from an ontology that is sufficient for an entailment to hold [11]. In the case of unsatisfiable classes and properties, justifications are computed for entailments with owl:Nothing and owl:bottomObjectProperty on the right-hand side of the inclusion axiom, enabling identification of the source(s) of incoherence.

Fig. 5figure5

Inference visualisation and justification algorithm explanations of [11] as implemented in Protégé v5.5. A: explanation in the case of the conflict described in Example 3; B: explanation of inference with the VIDO and COVoc ontologies as-is with the two alignment axioms on virus and organism

Conflicting modeling styles and language issues

These conflicts arise from source 3 listed at the start of the “Methods” section and concern the ‘other conflicts’, being nos. 9–10 listed in Table 1. Resolving them aims at restructuring (parts of) an ontology such that correspondences with entities of a different ontology can be established. For instance, if the same notion is modeled in one ontology as a class and in another as an object property (see Fig. 1), or even in both ways in the same ontology, and the ontology language does not permit heterogenous alignments, then either the object property has to be reified or the class has to be recast as an object property (see no. 9 in Table 1). Typical examples are object properties such as o1:married-to and o1:has-member and corresponding reifications as o2:Marriage and o2:Member, respectively. A concrete difference in modeling choice in two ontologies in the same domain is illustrated in Example 6.

Example 6

Let us continue with COVID-19 ontologies. Besides the CIDO (v1.0.181), there is also the CODO [32] (v1.3 of 25-9-2020), which focuses predominantly on patient data and therewith alike a practice-oriented application ontology. CODO’s COVID-19 test results are represented as codo:‘laboratory test finding’ ≡ , i.e., the outcomes are instances and, within the context of the other content in the ontology, practically conflated with the diagnosis of the disease. In contrast, CIDO is informed by the BFO top-level ontology with its modeling guidelines and has a cido:‘COVID-19 diagnosis’ class with three subclasses: negative COVID-19 diagnosis conclusion, and one for positive and one for presumptive positive. Aside from disagreeing on names and possibly also the meaning of possible test statuses, the modeling style issue here is that it is an example of class vs. instance modeling conflict of the same notion. Ontologically, the same notion cannot be both a class and an instance and should thus not co-exist in one ontology, hence, one of the two options will have to be chosen if they were to be integrated.

What is recorded in the conflict set depends on the case at hand; for the class vs. property example, these would be the respective axioms to match and the axioms they are used in, which may be found by using a natural language processing-based algorithm with part-of-speech tagging and stemming. Generally speaking, there are two different options of dealing with conflicting modeling styles. The first is to choose one style and then convert instances of the ‘losing’ modeling pattern into the style of the ‘winning’ pattern, where each consists of one or more axioms, and the second option is to keep both styles and match patterns by a set of axioms, rather than by a single bridging axiom, which is referred to as a heterogeneous TBox mapping [20].

The way how ODEs deal with conflicting styles depends on the kind of conflict and the ODE. For instance, Protégé 5 restricts alignments to equivalence and simple subsumption between classes and between object properties. Simply put, it does not provide the necessary means to assert an alignment between a class and an object property. An algorithm for detecting conflicting modeling styles and additional axioms for heterogeneous alignment has been proposed [20], but this is yet to be integrated in an ODE. A joint outside option may be the DOL framework that already does have a logic-based mechanism for heterogeneous alignments [17].

Language conflicts (no. 10 in Table 1) may vary in severity. There are straight-forward differences in naming patterns, such as camel case naming versus using underscores between words in a term, which are detected by the OOPS! tool [33], and one can use a spelling checker to detect US English versus British English orthography and more advanced resources to assist with synonym detection and resolution. It will not yet resolve which English variant to use in an ontology. More detailed labeling of language aspects and a stricter separation of the language layer on top of the ontology layer may assist in resolving this. Promising proposals for such additional modeling are the Ontolex-lemon community standard [34] and the MoLA [

留言 (0)

沒有登入
gif