Cultural knowledge evolution in areas (2024-02-10)
Line van den Berg, Jérôme Euzenat, Class? en classe: jouer avec des classifications pour combiner mathématiques et informatique, Recherches et recherches-actions en didactique de l'informatique 1(1), 2024
Class? est un jeu dans lequel les joueurs doivent organiser leurs cartes en fonction d’une classification cachée. Les cartes posées par les autres joueurs leur permettent de deviner où mettre les leurs. Il a été conçu pour que des écoliers appréhendent que les mêmes objets peuvent être classés de différentes manières et qu’il est possible de transmettre une classification sans l’expliciter. Le jeu fait appel a des notions qui se présentent facilement à l’aide des cartes à jouer comme des ensembles définis par des conditions nécessaires et suffisantes (classes). Cela permet d’introduire des classifications hiérarchiques et des notions algorithmiques (tests de conditions, récursion) pour les manipuler. Enfin, il nécessite de raisonner logiquement sur ces notions. Class? a été joué avec succès par des élèves du CM2 à la seconde. Nous nous sommes donc posé la question de son positionnement en tant que ressource pédagogique. Il apparait tout d’abord qu’il ne semble pas illustrer de concepts particulièrement mis en avant par les programmes officiels de l’éducation nationale. Il offre plutôt une manière alternative de renforcer des notions transversales très importantes en informatique. Nous caractérisons Class? par rapport aux efforts d’informatique sans ordinateurs et à d’autres jeux utilisables à cette fin. Finalement, nous discutons d’une décomposition de Class? en une succession de jeux plus simples permettant d’introduire les notions impliquées l’une après l’autre.
Classification hiérarchique, Apprentissage par le jeu, Informatique sans ordi, Ordres partiels, Conditions nécessaires et suffisantes
Yasser Bourahla, Multi-agent simulation of cultural ontology evolution through interaction, Thèse d'informatique, Université de Grenoble, Grenoble (FR), 2023
Artificial agents, as humans, use their knowledge to behave in an environment and within a society. Humans evolve their knowledge by adapting it in response to interactions with their environment and society. The question that is raised in this thesis is:``can knowledge evolve in a society of artificial agents, as it does in a human society?'' In particular, if agents adapt to improve their social interactions, how can this affect the quality of the population's knowledge about the environment? And how does it affect knowledge diversity? To address the questions, ontology evolution is simulated based on principles from experimental cultural evolution through an experimental framework in which: agents initially learn ontologies, from object samples, which they later adapt by interacting with each other about objects in the environment. Using this experimental framework, we show that: (1) agents reach a state of agreement in their interactions, (2) they improve the quality of their knowledge about the environment, and (3) they preserve the diversity of their knowledge. In order to characterise knowledge evolution through multiple generations, experiments are conducted with agents endowed with reproduction capabilities. Results show that (1) the variation provided by inter-generation transmission allows agents to further improve the quality of their ontologies; (2) agents select the knowledge to be preserved through intra-generation transmission which compensates for the lack of teacher selection in inter-generation transmission; and finally, (3) diversity remains stable from one generation to another. This work not only provides a basis for implementing agents capable of culturally evolving their knowledge, but also suggests that simulating such behavior can serve as a valuable tool for testing hypotheses about human cultural knowledge evolution.
Multi-agent simulation, Adaptive multi-agent systems, Cultural evolution
Jérôme Euzenat, Society = Autonomy + Adaptation, in: Proc. Dagstuhl seminar on Agents on the web, Wadern (DE), (Olivier Boissier, Andrei Ciortea, Andreas Harth, Alessandro Ricci, Danai Vachtsevanou (eds), Agents on the web (seminar 21072), Dagstuhl reports 13(2), 2023), pp86, 2023
What makes a true lively society is the capability of their members to autonomously adapt to others. It is not a set of norms cast in iron, be they programming norms or ’legal norms’. It is a set of beings trying to behave with others. This behaviour may lead to explicit norms that make explicit what should not have/need to be reinvented, but they may well remain implicit, hence continuously adapted. We should design software agents so that they are able to elaborate what drives their (social but not only) behaviours. They should be allowed to try, to make mistakes, and to transmit what they know. This is the ground on which evolution may happen. This capacity is what should be built in agents in order for them to behave without breaking too many things. The goal is not to reach a static equilibrium: in an open-ended agent space there are always opportunities to learn new things, meet new people and visit new places. Hence, rather than the state reached by agents, this is they ability to surf a dynamic disequilibrium that must be sought. This statement is somewhat made for triggering reactions within the seminar. It reacts to the apparent loss of autonomy of agents. It also extend the one I did for the previous seminar.
Andreas Kalaitzakis, Jérôme Euzenat, À quoi sert la spécialisation en évolution culturelle de la connaissance?, in: Maxime Morge (éd), Actes 31e journées francophones sur Systèmes multi-agent (JFSMA), Strasbourg (FR), pp76-85, 2023
Des agents peuvent faire évoluer leurs ontologies en accomplissant conjointement une tâche. Nous considérons un ensemble de tâches dont chaque agent ne considère qu'une partie. Nous supposons que moins un agent considère de tâches, plus la précision de sa meilleure tâche sera élevée. Pour le vérifier, nous simulons différentes populations considérant un nombre de tâches croissant. De manière contre-intuitive, l'hypothèse n'est pas vérifiée. D'une part, lorsque les agents ont une mémoire illimitée, plus un agent considère de tâches, plus il est précis. D'autre part, lorsque les agents ont une mémoire limitée, les objectifs de maximiser la précision de leur meilleures tâches et de s'accorder entre eux sont mutuellement exclusifs. Lorsque les sociétés favorisent la spécialisation, les agents n'améliorent pas leur précision. Cependant, ces agents décideront plus souvent en fonction de leurs meilleures tâches, améliorant ainsi la performance de leur société.
Evolution culturelle de la connaissance, Simulation multi-agents, Spécialisation des agents
Andreas Kalaitzakis, Jérôme Euzenat, Multi-tasking resource-constrained agents reach higher accuracy when tasks overlap, in: Proc. 20th European conference on multi-agents systems (EUMAS), Napoli (IT), (Vadim Malvone, Aniello Murano (eds), Proc. 20th European conference on multi-agents systems (EUMAS), Lecture notes in computer science 14282, 2023), pp425-434, 2023
Agents have been previously shown to evolve their ontologies while interacting over a single task. However, little is known about how interacting over several tasks affects the accuracy of agent ontologies. Is knowledge learned by tackling one task beneficial for another task? We hypothesize that multi-tasking agents tackling tasks that rely on the same properties, are more accurate than multi-tasking agents tackling tasks that rely on different properties. We test this hypothesis by varying two parameters. The first parameter is the number of tasks assigned to the agents. The second parameter is the number of common properties among these tasks. Results show that when deciding for different tasks relies on the same properties, multi-tasking agents reach higher accuracy. This suggests that when agents tackle several tasks, it is possible to transfer knowledge from one task to another.
Cultural knowledge evolution, Knowledge transfer, Multi-tasking
Adriana Luntraru, Value-sensitive knowledge evolution, Master's thesis, Université Grenoble Alpes, Grenoble (FR), 2023
Cultural values are cognitive representations of general objectives, such as independence or mastery, that people use to distinguish whether something is "good" or "bad". More specifically, people may use their values to evaluate alternatives and pick the most compatible one. Cultural values have been previously used in artificial societies of agents with the purpose of replicating and predicting human behavior. However, to the best of our knowledge, they have never been used in the context of cultural knowledge evolution. We consider cooperating agents which adapt their individually learned ontologies by interacting with each other to agree. When two agents disagree during an interaction, one of them needs to adapt its ontology. We use the cultural values of independence, novelty, authority and mastery to influence the choice of which agent adapts in a population of agents sharing the same values. We investigate the effects the choice of cultural values has on the knowledge obtained. Our results show that agents do not improve the accuracy of their knowledge without using the mastery value. Under certain conditions, independence causes the agents to converge to successful interactions faster, and novelty increases knowledge diversity, but both effects come with a large reduction in accuracy. We however did not find any significant effects of authority.
Anaïs Siebers, Intrinsic exploration-motivation in cultural knowledge evolution, Master's thesis, Ruhr Universität, Bochum (DE), 2023
In cultural knowledge evolution simulated by multi-agent simulations, agents can improve the accuracy of their knowledge by interacting with other agents and adapting their knowledge with the aim of agreeing. But their knowledge might be confined to specific areas because they do not have the capacity to explore the world around them. Since intrinsic motivation to explore in artificial agents has already proven to increase exploration, it was researched whether and how agents in simulations of cultural knowledge evolution can be motivated to explore. Moreover, it was tested how far this improves and changes their knowledge. Three different kinds of motivation were investigated: curiosity, creativity and non-exploration. Moreover, intrinsic motivation was modelled with and without reinforcement learning. Agents either explored on their own or picked specific interaction partner(s). It has been shown that it is possible to model agents with intrinsic motivation to explore in cultural knowledge evolution, and that this has a significant effect on the agents’ knowledge. Contrary to the expectations and other studies, this did not lead to an increase in knowledge completeness. Out of all intrinsic motivations, curiosity had the highest accuracy and completeness. Models with reinforcement learning performed similar to direct models. As expected, intrinsic motivation led to faster convergence of the agents’ knowledge, especially with social agents. Heterogeneously motivated agents only had a higher accuracy and completeness than homogeneously motivated agents in specific cases. This thesis can be regarded as a foundation for further investigation into the role of intrinsic motivation in cultural knowledge evolution. Different forms of intrinsic motivation or different reinforcement learning techniques could be tested. Additionally, intrinsic motivation at different stages of the experiment or in different ratios, for example curious agents and agents with no motivation, could be investigated in more detail. Lastly, agents could teach other agents things they explored a lot.
Cultural knowledge evolution, Intrinsic motivation, Exploration, Artificial curiosity, Computational creativity, Multi-agent simulation
Line van den Berg, Manuel Atencia, Jérôme Euzenat, Raising awareness without disclosing truth, Annals of mathematics and artificial intelligence 91(4):431-464, 2023
Agents use their own vocabularies to reason and talk about the world. Public signature awareness is satisfied if agents are aware of the vocabularies, or signatures, used by all agents they may, eventually, interact with. Multi-agent modal logics and in particular Dynamic Epistemic Logic rely on public signature awareness for modeling information flow in multi-agent systems. However, this assumption is not desirable for dynamic and open multi-agent systems because (1) it prevents agents to use unique signatures other agents are unaware of, (2) it prevents agents to openly extend their signatures when encountering new information, and (3) it requires that all future knowledge and beliefs of agents are bounded by the current state. We propose a new semantics for awareness that enables us to drop public signature awareness. This semantics is based on partial valuation functions and weakly reflexive relations. Dynamics for raising public and private awareness are then defined in such a way as to differentiate between becoming aware of a proposition and learning its truth value. With this, we show that knowledge and beliefs are not affected through the raising operations.
Awareness, Raising awareness, Dynamic epistemic logic, Partial valuations, Multi-agent systems
Yasser Bourahla, Manuel Atencia, Jérôme Euzenat, Knowledge transmission and improvement across generations do not need strong selection, in: Piotr Faliszewski, Viviana Mascardi, Catherine Pelachaud, Matthew Taylor (eds), Proc. 21st ACM international conference on Autonomous Agents and Multi-Agent Systems (AAMAS), (Online), pp163-171, 2022
Agents have been used for simulating cultural evolution and cultural evolution can be used as a model for artificial agents. Previous results have shown that horizontal, or intra-generation, knowledge transmission allows agents to improve the quality of their knowledge to a certain level. Moreover, variation generated through vertical, or inter-generation, transmission allows agents to exceed that level. Such results were obtained under specific conditions such as the drastic selection of agents allowed to transmit their knowledge, seeding the process with correct knowledge or introducing artificial noise during transmission. Here, we question the necessity of such measures and study their impact on the quality of transmitted knowledge. For that purpose, we combine the settings of two previous experiments and relax these conditions (no strong selection of teachers, no fully correct seed, no introduction of artificial noise). The rationale is that if interactions lead agents to improve their overall knowledge quality, this should be sufficient to ensure correct knowledge transmission, and that transmission mechanisms are sufficiently imperfect to produce variation. In this setting, we confirm that vertical transmission improves on horizontal transmission even without drastic selection and oriented learning. We also show that horizontal transmission is able to compensate for the lack of parent selection if it is maintained for long enough. This means that it is not necessary to take the most successful agents as teachers, neither in vertical nor horizontal transmission, to cumulatively improve knowledge.
Ontology, Multi-agent social simulation, Multi-agent learning, Knowledge diversity
Yasser Bourahla, Manuel Atencia, Jérôme Euzenat, Transmission de connaissances et sélection, in: Valérie Camps (éd), Actes 30e journées francophones sur Systèmes multi-agent (JFSMA), Saint-Étienne (FR), pp63-72, 2022
Les agents peuvent être utilisés pour simuler l'évolution culturelle et l'évolution culturelle peut être utilisée comme modèle pour les agents artificiels. Des expériences ont montré que la transmission intragénérationnelle des connaissances permet aux agents d'en améliorer la qualité. De plus, sa transmission intergénérationnelle permet de dépasser ce niveau. Ces résultats ont été obtenus dans des conditions particulières: sélection drastique des agents transmetant leurs connaissances, initialisation avec des connaissances correctes ou introduction de bruit lors de la transmission. Afin d'étudier l'impact de ces mesures sur la qualité de la connaissance transmise, nous combinons les paramètres de deux expériences précédentes et relâchons ces conditions. Ce dispositif confirme que la transmission verticale permet d'améliorer la qualité de la connaissance obtenue par transmission horizontale même sans sélection drastique et apprentissage orienté. Il montre également qu'une transmission intragénérationnelle suffisante peut compenser l'absence de sélection parentale.
Simulation sociale multi-agents, Évolution culturelle, Transmission des connaissances, Génération d'agents, Évolution culturelle des connaissances
Yasser Bourahla, Jérôme David, Jérôme Euzenat, Meryem Naciri, Measuring and controlling knowledge diversity, in: Tiago Prince Sales, Maria Hedblom, He Tan, Lucía Gómez Álvarez, Rafael Peñaloza, Srdjan Vesic (eds), Proc. 1st JOWO workshop on formal models of knowledge diversity (FMKD), Jönköping (SE), 2022
Assessing knowledge diversity may be useful for many purposes. In particular, it is necessary to measure diversity in order to understand how it arises or is preserved; it is also necessary to control it in order to measure its effects. Here we consider measuring knowledge diversity using two components: (a) a diversity measure taking advantage of (b) a knowledge difference measure. We present the general principles and various candidates for such components. We discuss how these measures may be used to generate populations of agents with controlled levels of knowledge diversity.
Knowledge diversity, Diversity measure, Ontology dissimilarity, Diversity control, Entropy
Yasser Bourahla, Manuel Atencia, Jérôme Euzenat, Inter-generation knowledge transmission without individual selection, in: Proc. 4th conference on Conference of the Cultural evolution society, Aarhus (DK), 2022
cultural knowledge evolution, vertical transmission, horizontal transmission, multi-agent simulation, knowledge accuracy
Jérôme Euzenat, Can AI systems culturally evolve their knowledge?, in: Proc. 4th conference on Conference of the Cultural evolution society, Aarhus (DK), 2022
agent-based models, computational cultural knowledge evolution, artificial intelligence
Line van den Berg, Jérôme Euzenat, The small Class? gamebook, Pedagogical material, 2022
Class? is an enjoyable card game aiming at grouping colourful cards into meaningful classes. It illustrates facets of reasoning with classifications. In order to introduce Class? progressively, this small gamebook provides a sequence of games before getting to the Class? game itself and beyond. The games are presented in increasing order of difficulty so that a game will benefit from mastering of previous ones.
Class?, Classification, Game
Yasser Bourahla, Manuel Atencia, Jérôme Euzenat, Knowledge improvement and diversity under interaction-driven adaptation of learned ontologies, in: Ulle Endriss, Ann Nowé, Frank Dignum, Alessio Lomuscio (eds), Proc. 20th ACM international conference on Autonomous Agents and Multi-Agent Systems (AAMAS), London (UK), pp242-250, 2021
When agents independently learn knowledge, such as ontologies, about their environment, it may be diverse, incorrect or incomplete. This knowledge heterogeneity could lead agents to disagree, thus hindering their cooperation. Existing approaches usually deal with this interaction problem by relating ontologies, without modifying them, or, on the contrary, by focusing on building common knowledge. Here, we consider agents adapting ontologies learned from the environment in order to agree with each other when cooperating. In this scenario, fundamental questions arise: Do they achieve successful interaction? Can this process improve knowledge correctness? Do all agents end up with the same ontology? To answer these questions, we design a two-stage experiment. First, agents learn to take decisions about the environment by classifying objects and the learned classifiers are turned into ontologies. In the second stage, agents interact with each other to agree on the decisions to take and modify their ontologies accordingly. We show that agents indeed reduce interaction failure, most of the time they improve the accuracy of their knowledge about the environment, and they do not necessarily opt for the same ontology.
Ontology, Multi-agent social simulation, Multi-agent learning, Knowledge diversity
Jérôme Euzenat, The web as a culture broth for agents and people to grow knowledge, in: Proc. Dagstuhl seminar on Autonomous agents on the web, Wadern (DE), (Olivier Boissier, Andrei Ciortea, Andreas Harth, Alessandro Ricci (eds), Autonomous agents on the web (seminar 21072), Dagstuhl reports 11(1), 2021), pp40-41, 2021
Line van den Berg, Manuel Atencia, Jérôme Euzenat, A logical model for the ontology alignment repair game, Autonomous agents and multi-agent systems 35(2):32, 2021
Ontology alignments enable agents to communicate while preserving heterogeneity in their knowledge. Alignments may not be provided as input and should be able to evolve when communication fails or when new information contradicting the alignment is acquired. The Alignment Repair Game (ARG) has been proposed for agents to simultaneously communicate and repair their alignments through adaptation operators when communication failures occur. ARG has been evaluated experimentally and the experiments showed that agents converge towards successful communication and improve their alignments. However, whether the adaptation operators are formally correct, complete or redundant could not be established by experiments. We introduce a logical model, Dynamic Epistemic Ontology Logic (DEOL), that enables us to answer these questions. This framework allows us (1) to express the ontologies and alignments used via a faithful translation from ARG to DEOL, (2) to model the ARG adaptation operators as dynamic modalities and (3) to formally define and establish the correctness, partial redundancy and incompleteness of the adaptation operators in ARG.
The refine operator is not partially redundant with respect to Agent b (because it has no way to detect the incoherence from the announcement alone).
Ontology alignment, Alignment repair, Multi-agent systems, Agent communication, Dynamic Epistemic Logic
Line van den Berg, Cultural knowledge evolution in dynamic epistemic logic, Thèse de mathématiques-informatique, Université de Grenoble, Grenoble (FR), October 2021
To reason and talk about the world, agents may use their own distinct vocabularies, structured into knowledge representations, also called ontologies. In order to communicate, they use alignments: translations between terms of their ontologies. aHowever, ontologies may change, requiring their alignments to evolve accordingly. Experimental cultural evolution offers a framework to study the mechanisms of their knowledge evolution. It has been applied to the evolution of alignments in the Alignment Repair Game (ARG). Experiments have shown that, through ARG, agents improve their alignments and reach successful communication. Yet, these experiments are not sufficient to understand the formal properties of cultural knowledge evolution. This thesis bridges experimental cultural knowledge evolution with a theoretical model of cultural knowledge evolution in logic. This is achieved by introducing Dynamic Epistemic Ontology Logic and defining a faithful translation of ARG in DEOL that (a) encodes the ontologies, (b) maps agents' ontologies and alignments to knowledge and beliefs, and (c) captures the adaptation operators through announcements and conservative upgrades. This model shows that all but one adaptation operator are correct, they are incomplete and some are partially redundant. Three differences between the ARG agents and their logical model explain these results, leading to an independent model of awareness based on partial valuations and weakly reflexive relations. An alternative model of ARG is then defined under which the formal properties are re-examined, showing that this model is closer to the original game. This is a first step towards defining a theoretical model of cultural knowledge evolution.
Dynamic epistemic logic, Ontology alignments, Cultural knowledge evolution
Jérôme Euzenat, A map without a legend: the semantic web and knowledge evolution, Semantic web journal 11(1):63-68, 2020
The current state of the semantic web is focused on data. This is a worthwhile progress in web content processing and interoperability. However, this does only marginally contribute to knowledge improvement and evolution. Understanding the world, and interpreting data, requires knowledge. Not knowledge cast in stone for ever, but knowledge that can seamlessly evolve; not knowledge from one single authority, but diverse knowledge sources which stimulate confrontation and robustness; not consistent knowledge at web scale, but local theories that can be combined. We discuss two different ways in which semantic web technologies can greatly contribute to the advancement of knowledge: semantic eScience and cultural knowledge evolution.
Semantic web, Linked data, Big data, Open data, Knowledge representation, Knowledge, Ontology, Machine learning, Reproducible research, eScience, Cultural evolution
Line van den Berg, Manuel Atencia, Jérôme Euzenat, Agent ontology alignment repair through dynamic epistemic logic, in: Bo An, Neil Yorke-Smith, Amal El Fallah Seghrouchni, Gita Sukthankar (eds), Proc. 19th ACM international conference on Autonomous Agents and Multi-Agent Systems (AAMAS), Auckland (NZ), pp1422-1430, 2020
Ontology alignments enable agents to communicate while preserving heterogeneity in their information. Alignments may not be provided as input and should be able to evolve when communication fails or when new information contradicting the alignment is acquired. In the Alignment Repair Game (ARG) this evolution is achieved via adaptation operators. ARG was evaluated experimentally and the experiments showed that agents converge towards successful communication and improve their alignments. However, whether the adaptation operators are formally correct, complete or redundant is still an open question. In this paper, we introduce a formal framework based on Dynamic Epistemic Logic that allows us to answer this question. This framework allows us (1) to express the ontologies and alignments used, (2) to model the ARG adaptation operators through announcements and conservative upgrades and (3) to formally establish the correctness, partial redundancy and incompleteness of the adaptation operators in ARG.
The refine operator is not partially redundant with respect to Agent b (because it has no way to detect the incoherence from the announcement alone).
Ontology alignment, Alignment repair, Agent communication, Dynamic Epistemic Logic
Line van den Berg, Manuel Atencia, Jérôme Euzenat, Unawareness in multi-agent systems with partial valuations, in: Proc. 10th AAMAS workshop on Logical Aspects of Multi-Agent Systems (LAMAS), Auckland (NZ), 2020
Public signature awareness is satisfied if agents are aware of the vocabulary, propositions, used by other agents to think and talk about the world. However, assuming that agents are fully aware of each other's signatures prevents them to adapt their vocabularies to newly gained information, from the environment or learned through agent communication. Therefore this is not realistic for open multi-agent systems. We propose a novel way to model awareness with partial valuations that drops public signature awareness and can model agent signature unawareness, and we give a first view on defining the dynamics of raising and forgetting awareness on this framework.
Awareness, Dynamic Epistemic Logic, Partial valuations, Multi-agent systems
Jérôme Euzenat, Replicator-interactor in experimental cultural knowledge evolution, in: Proc. 2nd JOWO workshop on Interaction-Based Knowledge Sharing (WINKS), Graz (AT), 2019
Cultural evolution may be studied at a `macro' level, inspired from population dynamics, or at a `micro' level, inspired from genetics. The replicator-interactor model generalises the genotype-phenotype distinction of genetic evolution. Here, we consider how it can be applied to cultural knowledge evolution experiments. In particular, we consider knowledge as replicator and the behaviour it induces as interactor. We show that this requires to address problems concerning transmission. We discuss the introduction of horizontal transmission within the replicator-interactor model and/or differential reproduction within cultural evolution experiments.
Line van den Berg, Epistemic alignment repair, in: Proc. 31st ESSLLI student session, Riga (LV), 2019
Ontology alignments enable interoperability between heterogeneous information resources. The Alignment Repair Game (ARG) specifically provides a way for agents to simultaneously communicate and improve the alignment when a communication failure occurs. This is achieved through applying adaptation operators that provide a revision strategy for agents to resolve failures with minimum information loss. In this paper, we explore how closely these operators resemble logical dynamics. We develop a variant of Dynamic Epistemic Logic called DEOL to capture the dynamics of ARG by modeling ontologies as knowledge and alignments as belief with respect to the plausibility relation. The dynamics of ARG are then achieved through announcements and conservative upgrades. With the representation of ARG in DEOL, we formally establish the limitations and the redundancy of the adaptation operators. More precisely, that for a complete logical reasoner, replace, addjoin and refine are redundant for one or both agents in the game and that add would be replaced by addjoin in all cases.
Ontology alignment, Alignment repair, Dynamic Epistemic Logic
Jérôme Euzenat, De la langue à la connaissance: approche expérimentale de l'évolution culturelle, Bulletin de l'AFIA 100:9-12, 2018
Jérôme Euzenat, Interaction-based ontology alignment repair with expansion and relaxation, in: Proc. 26th International Joint Conference on Artificial Intelligence (IJCAI), Melbourne (VIC AU), pp185-191, 2017
Agents may use ontology alignments to communicate when they represent knowledge with different ontologies: alignments help reclassifying objects from one ontology to the other. These alignments may not be perfectly correct, yet agents have to proceed. They can take advantage of their experience in order to evolve alignments: upon communication failure, they will adapt the alignments to avoid reproducing the same mistake. Such repair experiments had been performed in the framework of networks of ontologies related by alignments. They revealed that, by playing simple interaction games, agents can effectively repair random networks of ontologies. Here we repeat these experiments and, using new measures, show that previous results were underestimated. We introduce new adaptation operators that improve those previously considered. We also allow agents to go beyond the initial operators in two ways: they can generate new correspondences when they discard incorrect ones, and they can provide less precise answers. The combination of these modalities satisfy the following properties: (1) Agents still converge to a state in which no mistake occurs. (2) They achieve results far closer to the correct alignments than previously found. (3) They reach again 100% precision and coherent alignments.
The results reported in this paper for operators addjoin and refadd are not accurate, due to a software error. The results reported were worse than they should have been. Updated results can be found in [
20180308-NOOR], [
20180311-NOOR] and [
20180529-NOOR].
Jérôme Euzenat, Crafting ontology alignments from scratch through agent communication, in: Proc. 20th International Conference on Principles and practice of multi-agent systems (PRIMA), Nice (FR), (Bo An, Ana Bazzan, João Leite, Serena Villata, Leendert van der Torre (eds), Proc. 20th International Conference on Principles and practice of multi-agent systems (PRIMA), Lecture notes in computer science 10621, 2017), pp245-262, 2017
Agents may use different ontologies for representing knowledge and take advantage of alignments between ontologies in order to communicate. Such alignments may be provided by dedicated algorithms, but their accuracy is far from satisfying. We already explored operators allowing agents to repair such alignments while using them for communicating. The question remained of the capability of agents to craft alignments from scratch in the same way. Here we explore the use of expanding repair operators for that purpose. When starting from empty alignments, agents fails to create them as they have nothing to repair. Hence, we introduce the capability for agents to risk adding new correspondences when no existing one is useful. We compare and discuss the results provided by this modality and show that, due to this generative capability, agents reach better results than without it in terms of the accuracy of their alignments. When starting with empty alignments, alignments reach the same quality level as when starting with random alignments, thus providing a reliable way for agents to build alignment from scratch through communication.
Ontology alignment, Alignment repair, Cultural knowkedge evolution, Agent simulation, Coherence, Network of ontologies
Jérôme Euzenat, Knowledge diversity under socio-environmental pressure, in: Michael Rovatsos (ed), Investigating diversity in AI: the ESSENCE project, 2013-2017, Deliverable, ESSENCE, 62p., 2017, pp28-30
Experimental cultural evolution has been convincingly applied to the evolution of natural language and we aim at applying it to knowledge. Indeed, knowledge can be thought of as a shared artefact among a population influenced through communication with others. It can be seen as resulting from contradictory forces: internal consistency, i.e., pressure exerted by logical constraints, against environmental and social pressure, i.e., the pressure exerted by the world and the society agents live in. However, adapting to environmental and social pressure may lead agents to adopt the same knowledge. From an ecological perspective, this is not particularly appealing: species can resist changes in their environment because of the diversity of the solutions that they can offer. This problem may be approached by involving diversity as an internal constraint resisting external pressure towards uniformity.