• Refine Query
  • Source
  • Publication year
  • to
  • Language
  • 20
  • 13
  • 1
  • 1
  • Tagged with
  • 38
  • 38
  • 15
  • 15
  • 10
  • 9
  • 9
  • 8
  • 8
  • 7
  • 7
  • 6
  • 6
  • 6
  • 6
  • About
  • The Global ETD Search service is a free service for researchers to find electronic theses and dissertations. This service is provided by the Networked Digital Library of Theses and Dissertations.
    Our metadata is collected from universities around the world. If you manage a university/consortium/country archive and want to be added, details can be found on the NDLTD website.
11

Revisão de crenças em ACTL usando verificação de modelos limitada / Belief revision in ACTL using bounded model checking

Hora, Bruno Vercelino da 03 August 2017 (has links)
Uma importante etapa do desenvolvimento de software é o de levantamento e análise dos requisitos. Porém, durante esta etapa podem ocorrer inconsistências que prejudicarão o andamento do projeto. Além disso, após finalizada a especificação, o cliente pode querer acrescentar ou modificar as funcionalidades do sistema. Tudo isso requer que a especificação do software seja revista, mas isso é altamente custoso, tornando necessário um processo automatizado para simplificar tal revisão. Para lidar com este problema, uma das abordagens utilizadas tem sido o processo de Revisão de Crenças, juntamente com o processo de Verificação de Modelos. O objetivo deste trabalho é utilizar o processo de revisão de crenças e verificação de modelos para avaliar especificações de um projeto procurando inconsistências, utilizando o fragmento universal da Computation Tree Logic (CTL), conhecido como ACTL, e revisá-las gerando sugestões de mudanças na especificação. A nossa proposta é traduzir para lógica clássica tanto o modelo (especificação do software) quanto a propriedade a ser revisada, e então aplicar um resolvedor SAT para verificar a satisfazibilidade da fórmula gerada. A partir da resposta do resolvedor SAT, iremos gerar sugestões válidas de mudanças para a especificação, fazendo o processo de tradução reversa da lógica clássica para o modelo original. / The objective of this work is to join the proccess of belief revision and model checking to evaluate project specifications looking for inconsistences, using the universal fragment of Computation Tree Logic (CTL), known as ACTL, and revise them generating changes suggestions in the specification. Our approach will translate the model (software specification) and the property to be revised to classical logic. Then we will apply a SAT solver to verify the generated formulas satsifability. From the SAT solver answer, we will create changes valid suggestions to the specification making the translation back from classical logic to the original model. To generate the changes suggestions, we proposed a framework based on heuristics where different approaches and decisions can be implemented, aiming a better application for each project scope. We implemented a basic heuristic as an example and used it to test the implementation to analise the proposed algorithm
12

Methodologies for Belief Revision in Multi-agent Systems / Metodologias de Revisão de Crenças em Sistemas Multi-agente

Malheiro, Maria Benedita Campos Neves January 1999 (has links)
The goal of this thesis is twofold: first, we want to present the distributed belief accommodation and revision model for multi-agent systems that has been developed and, second, we wish to show its applicability to an appropriate domain. The Distributed Belief Accommodation & Revision model, called DeBAteR model, was developed for co-operative heterogeneous multi-agent systems used to model inherently dynamic distributed problems. In these systems, although the agents are able to detect changes both in the environment and in the problem specifications, each agent has only a partial view of the global picture. As a result the information that represents the current state of affairs is dynamic, incomplete and sometimes uncertain. This non-monotonic kind of data is called beliefs ? a belief is a piece of data that is held as correct as long as no contradicting evidence is found or presented. Each agent is expected to include an assumption based truth maintenance module for representing properly this type of data. Our main effort was concentrated on the task of maintaining the system's information, which consists of updating, revising and accommodating the represented beliefs. Belief updating is necessary for including the changes detected by the agents both in the environment and/or in the problem specifications. Belief revision is essential for solving the inconsistencies detected among the represented beliefs. Belief accommodation and revision is crucial for integrating the multiple disparate perspectives regarding the same data items, which may occur whenever there is overlap of expertise domains between the agents. In order to solve the information conflicts that result from the detection of inconsistencies between distinct beliefs or within multi-perspective beliefs we conceived the DeBATeR model. The DeBAteR is fully distributed, provides individual belief autonomy and is made of two methodologies: the pro-active belief accommodation and revision methodology and the delayed belief revision methodology. Whilst the first methodology is used to solve domain independent conflicts, the second methodology was devised for solving domain dependent conflicts. Both methodologies use argumentation for, in the case of the domain independent conflicts, choosing the most credible perspective between the existing multiple perspectives of a belief, and, in the case of the domain dependent conflicts, finding the best alternative belief support set for the affected concepts. These methodologies are distributed and their scope may be internal or collective. The DeBAteR model main contributions are: (i) the pro-active methodology conceived for solving domain independent conflicts and (ii) the capability, not only to represent and maintain individual beliefs and joint beliefs, but also to accommodate, rationally maintain and make use of multi-perspective beliefs. Finally, we describe the developed decision support multi-agent system for choosing adequate project locations, called DIPLOMAT ? Dynamic and Interactive Project Location Test bed, which has the ability of accommodating and revising the represented beliefs according to the DeBAteR model methodologies. / O objectivo desta dissertação é duplo: por um lado, pretendemos dar a conhecer o modelo de revisão e acomodação de crenças para sistemas multi-agente por nós desenvolvido e, por outro, procuramos ilustrar a sua validade descrevendo a aplicação que realizámos. O modelo, designado DeBAteR ? Distributed Belief Accommodation & Revision, destina-se a sistemas multi-agente cooperativos e heterogéneos que modelam problemas inerentemente distribuídos e dinâmicos. Neste tipo de sistemas, a informação que representa o ambiente é dinâmica (os agentes possuem a capacidade de constatar alterações no ambiente e/ou nas condições do problema) e, muitas vezes, incompleta (os agentes possuem visões parcelares da realidade) e/ou incerta. Este tipo de informação, de carácter não definitivo, designa-se por crenças ? uma crença é uma convicção tida como correcta enquanto não for posta em causa por alguma evidência. Cada agente, a fim de representar e manipular crenças, foi enriquecido com um módulo específico de manutenção de consistência baseado em suposições. O nosso esforço concentrou-se na tarefa de manutenção (actualização, revisão e acomodação) da informação do sistema. A actualização de crenças é essencial para incorporar as alterações que os agentes detectam no ambiente e/ou nas condições do problema. A revisão de crenças é indispensável para resolver inconsistências (conflitos) entre as crenças representadas. A acomodação e revisão simultânea de crenças é imprescindível para a integração das múltiplas perspectivas díspares que surgem em relação a um mesmo item de informação (crenças pluri-perspectiva) quando existe sobreposição de domínios de especialidade entre os agentes. Para tentar solucionar estes conflitos entre crenças concebemos um modelo distribuído que assegura autonomia individual de crença. O modelo de acomodação e revisão de crenças DeBAteR é composto por duas metodologias: a metodologia pró-activa de acomodação e revisão de crenças pluri-perspectiva e a metodologia retardada de revisão de crenças. A primeira, destina-se a resolver conflitos independentes do domínio e a segunda destina-se à resolução de conflitos dependentes do domínio. Estas metodologias são suportadas quase integralmente por um sistema de argumentação que procura, no caso dos conflitos independentes do domínio, escolher a perspectiva mais credível e, no caso dos conflitos dependentes do domínio, encontrar o melhor conjunto alternativo de suporte para os conceitos afectados. Esta actividade é descentralizada e pode decorrer quer no âmbito intra-agente, quer no âmbito inter-agente. É ainda de realçar: (i) o carácter pró-activo da resolução dos conflitos independentes do domínio (crenças pluri-perspectiva) e (ii) a capacidade de, não só, representar e manter crenças de âmbito individual (crenças suportadas por apenas um agente) e crenças conjuntas (crenças suportadas por vários agentes), mas também, de sintetizar, manter racionalmente e utilizar crenças pluri-perspectiva. Por último, descrevemos o sistema multi-agente de apoio à decisão no domínio da localização de empreendimentos desenvolvido, denominado DIPLOMAT ? Dynamic and Interactive Project Location Test bed, o qual possui a capacidade de acomodar e rever crenças de acordo com as metodologias concebidas no âmbito do modelo DeBAteR.
13

Dynamics of argumentation frameworks / Dynamique des systèmes d'argumentation

Mailly, Jean-Guy 30 September 2015 (has links)
Cette thèse traite du problème de l'intégration d'une nouvelle information dans un système d'argumentation abstrait. Un tel système est un graphe orienté dont les nœuds représentent les arguments, et les arcs représentent les attaques entre arguments. Il existe divers moyen de décider quels arguments sont acceptés par l'agent qui utilise un tel système pour représenter ses croyances.Il peut arriver dans la vie d'un agent qu'il soit confronté à une information du type "tel argument devrait être accepté", alors que c'est en contradiction avec ses croyances actuelles, représentées par son système d'argumentation.Nous avons étudié dans cette thèse diverses approches pour intégrer une information à un système d'argumentation.Notre première contribution est une adaptation du cadre AGM pour la révision de croyances, habituellement utilisé lorsque les croyances de l'agent sont représentées dans un formalisme logique. Nous avons notamment adapté les postulats de rationalité proposés dans le cadre AGM pour pouvoir caractériser des opérateurs de révision de systèmes d'argumentation, et nous avons proposé différents moyens de générer les systèmes d'argumentation résultant de la révision.Nous avons ensuite proposé d'utiliser la révision AGM comme un outil pour réviser les systèmes d'argumentation. Il s'agit cette fois-ci d'une approche par encodage en logique du système d'argumentation, qui permet d'utiliser les opérateurs de révision usuels pour obtenir le résultat souhaité.Enfin, nous avons étudié le problème du forçage d'un ensemble d'arguments (comment modifier le système pour qu'un ensemble donné soit une extension). Nous avons proposé une nouvelle famille d'opérateurs qui garantissent le succès de l'opération, contrairement aux opérateurs de forçage existants, et nous avons montré qu'une traduction de nos approches en problèmes de satisfaction ou d'optimisation booléenne permet de développer des outils efficaces pour calculer le résultat du forçage. / This thesis tackles the problem of integrating a new piece of information in an abstract argumentation framework. Such a framework is a directed graph such that its nodes represent the arguments, and the directed edges represent the attacks between arguments. There are different ways to decide which arguments are accepted by the agent who uses such a framework to represent her beliefs.An agent may be confronted with a piece of information such that "this argument should be accepted", which is in contradiction with her current beliefs, represented by her argumentation framework.In this thesis, we have studied several approaches to incorporate a piece of information in an argumentation framework.Our first contribution is an adaptation of the AGM framework for belief revision, which has been developed for characterizing the incorporation of a new piece of information when the agent's beliefs are represented in a logical setting. We have adapted the rationality postulates from the AGM framework to characterize the revision operators suited to argumentation frameworks, and we have identified several ways to generate the argumentation frameworks resulting from the revision.We have also shown how to use AGM revision as a tool for revising argumentation frameworks. Our approach uses a logical encoding of the argumentation framework to take advantage of the classical revision operators, for deriving the expected result.At last, we have studied the problem of enforcing a set of arguments (how to change an argumentation framework so that a given set of arguments becomes an extension). We have developed a new family of operators which guarantee the success of the enforcement process, contrary to the existing approaches, and we have shown that a translation of our approaches into satisfaction and optimization problems makes possible to develop efficient tools for computing the result of the enforcement.
14

Revisão de crenças em ACTL usando verificação de modelos limitada / Belief revision in ACTL using bounded model checking

Bruno Vercelino da Hora 03 August 2017 (has links)
Uma importante etapa do desenvolvimento de software é o de levantamento e análise dos requisitos. Porém, durante esta etapa podem ocorrer inconsistências que prejudicarão o andamento do projeto. Além disso, após finalizada a especificação, o cliente pode querer acrescentar ou modificar as funcionalidades do sistema. Tudo isso requer que a especificação do software seja revista, mas isso é altamente custoso, tornando necessário um processo automatizado para simplificar tal revisão. Para lidar com este problema, uma das abordagens utilizadas tem sido o processo de Revisão de Crenças, juntamente com o processo de Verificação de Modelos. O objetivo deste trabalho é utilizar o processo de revisão de crenças e verificação de modelos para avaliar especificações de um projeto procurando inconsistências, utilizando o fragmento universal da Computation Tree Logic (CTL), conhecido como ACTL, e revisá-las gerando sugestões de mudanças na especificação. A nossa proposta é traduzir para lógica clássica tanto o modelo (especificação do software) quanto a propriedade a ser revisada, e então aplicar um resolvedor SAT para verificar a satisfazibilidade da fórmula gerada. A partir da resposta do resolvedor SAT, iremos gerar sugestões válidas de mudanças para a especificação, fazendo o processo de tradução reversa da lógica clássica para o modelo original. / The objective of this work is to join the proccess of belief revision and model checking to evaluate project specifications looking for inconsistences, using the universal fragment of Computation Tree Logic (CTL), known as ACTL, and revise them generating changes suggestions in the specification. Our approach will translate the model (software specification) and the property to be revised to classical logic. Then we will apply a SAT solver to verify the generated formulas satsifability. From the SAT solver answer, we will create changes valid suggestions to the specification making the translation back from classical logic to the original model. To generate the changes suggestions, we proposed a framework based on heuristics where different approaches and decisions can be implemented, aiming a better application for each project scope. We implemented a basic heuristic as an example and used it to test the implementation to analise the proposed algorithm
15

Revisão de modelos formais de sistemas de estados finitos / Revision of formal models finite state systems

Thiago Carvalho de Sousa 26 March 2007 (has links)
Neste trabalho apresentamos uma implementação de revisão de crenças baseada em comparação de modelos (estados) em uma ferramenta de verificação automática de sistemas de estados finitos. Dada uma fórmula (na lógica CTL) inconsistente com o modelo do sistema, revisamos esse modelo de tal maneira que essa fórmula temporal se torne verdadeira. Como temos oito operadores temporais (AG, AF, AX, AU, EG, EF, EX e EU), foram criados algoritmos especícos para cada um deles. Como o modelo do sistema deriva do seu código na linguagem SMV, a sua revisão passa obrigatoriamente por mudanças na sua descrição. A nossa implementação contempla três tipos de mudanças: acréscimo de linhas, eliminação de linhas e mudança no estado inicial, sendo que as duas primeiras provocam modicações nas transições entre os estados que compõe o modelo. Alguns testes foram aplicados para comprovar a contribuição da revisão de crenças (revisão de modelos) como ferramenta de auxílio ao usuário durante o processo de modelagem de sistemas. / In this work we present an implementation of belief revision based on comparison of models (states) in a tool for automatic verication of nite state systems. Given a formula (in the language of CTL) inconsistent with the model of the system, we revise this model in such way that the temporal formula becomes valid. As we have eight temporal operators (AG, AF, AX, AU, EG, EF, EX and EU), specic algorithms for each one of them have been created. As the model of the system is related with its code in SMV language, its revision forces changes in its description. Our implementation contemplates three types of change: addition of lines, elimination of lines and change in the initial state, where the rst two cause modications in the transitions between the states of the model. Some tests were applied to prove the contribution of the belief revision (model revision) as aid-tool to the user during the process of systems modeling.
16

Revisão de modelos CTL / CTL Model Revision

Paulo de Tarso Guerra Oliveira 16 December 2010 (has links)
Verificação de modelos é uma das mais eficientes técnicas de verificação automática de sistemas. No entanto, apesar de poder lidar com verificações complexas, as ferramentas de verificação de modelos usualmente não fornecem informação alguma sobre como reparar inconsistências nestes modelos. Nesta dissertação, mostramos que abordagens desenvolvidas para a atualização de modelos CTL inconsistentes não são capazes de lidar com todos os tipos de alterações em modelos. Introduzimos então o conceito de revisão de modelos: uma abordagem baseada em revisão de crenças para o reparo de modelos inconsistentes em um contexto estático. Relacionamos nossa proposta com trabalhos clássicos em revisão de crenças. Definimos um operador de revisão de modelos e mostramos que este obedece postulados de racionalidade clássico de revisão de crenças. Propomos um algoritmo de revisão com base no algoritmo utilizado pela abordagem de atualização de modelos. Discutimos sobre problemas e limites do algoritmo proposto, e mostramos que essa estratégia de adaptação não é uma solução apropriada. / Model checking is one of the most robust techniques in automated system verification. But, although this technique can handle complex verifications, model checking tools usually do not give any information on how to repair inconsistent system models. In this dissertation, we show that approaches developed for CTL model update cannot deal with all kinds of model changes. We introduce the concept of CTL model revision: an approach based on belief revision to handle system inconsistency in a static context. We relate our proposal to classical works on belief revision. We define an operator for model revision and we show that it obeys the classical rationality postulates of belief revision. We propose an algorithm for model revision based on the algorithm used by the model update approach. We discuss problems and limitations of our proposed algorithm and show that this strategy of adaptation is not an appropriate solution.
17

Exploring Teaching Regimes to Change Preschoolers' Mistaken Beliefs about Sinking Objects

Baker, Heather R. January 2013 (has links)
No description available.
18

Belief Revision in Dynamic Abducers through Meta-Abduction

Bharathan, Vivek 14 September 2010 (has links)
No description available.
19

Belief Change in Reasoning Agents / Axiomatizations, Semantics and Computations

Jin, Yi 26 January 2007 (has links) (PDF)
The capability of changing beliefs upon new information in a rational and efficient way is crucial for an intelligent agent. Belief change therefore is one of the central research fields in Artificial Intelligence (AI) for over two decades. In the AI literature, two different kinds of belief change operations have been intensively investigated: belief update, which deal with situations where the new information describes changes of the world; and belief revision, which assumes the world is static. As another important research area in AI, reasoning about actions mainly studies the problem of representing and reasoning about effects of actions. These two research fields are closely related and apply a common underlying principle, that is, an agent should change its beliefs (knowledge) as little as possible whenever an adjustment is necessary. This lays down the possibility of reusing the ideas and results of one field in the other, and vice verse. This thesis aims to develop a general framework and devise computational models that are applicable in reasoning about actions. Firstly, I shall propose a new framework for iterated belief revision by introducing a new postulate to the existing AGM/DP postulates, which provides general criteria for the design of iterated revision operators. Secondly, based on the new framework, a concrete iterated revision operator is devised. The semantic model of the operator gives nice intuitions and helps to show its satisfiability of desirable postulates. I also show that the computational model of the operator is almost optimal in time and space-complexity. In order to deal with the belief change problem in multi-agent systems, I introduce a concept of mutual belief revision which is concerned with information exchange among agents. A concrete mutual revision operator is devised by generalizing the iterated revision operator. Likewise, a semantic model is used to show the intuition and many nice properties of the mutual revision operator, and the complexity of its computational model is formally analyzed. Finally, I present a belief update operator, which takes into account two important problems of reasoning about action, i.e., disjunctive updates and domain constraints. Again, the updated operator is presented with both a semantic model and a computational model.
20

Belief Change in Reasoning Agents: Axiomatizations, Semantics and Computations

Jin, Yi 17 January 2007 (has links)
The capability of changing beliefs upon new information in a rational and efficient way is crucial for an intelligent agent. Belief change therefore is one of the central research fields in Artificial Intelligence (AI) for over two decades. In the AI literature, two different kinds of belief change operations have been intensively investigated: belief update, which deal with situations where the new information describes changes of the world; and belief revision, which assumes the world is static. As another important research area in AI, reasoning about actions mainly studies the problem of representing and reasoning about effects of actions. These two research fields are closely related and apply a common underlying principle, that is, an agent should change its beliefs (knowledge) as little as possible whenever an adjustment is necessary. This lays down the possibility of reusing the ideas and results of one field in the other, and vice verse. This thesis aims to develop a general framework and devise computational models that are applicable in reasoning about actions. Firstly, I shall propose a new framework for iterated belief revision by introducing a new postulate to the existing AGM/DP postulates, which provides general criteria for the design of iterated revision operators. Secondly, based on the new framework, a concrete iterated revision operator is devised. The semantic model of the operator gives nice intuitions and helps to show its satisfiability of desirable postulates. I also show that the computational model of the operator is almost optimal in time and space-complexity. In order to deal with the belief change problem in multi-agent systems, I introduce a concept of mutual belief revision which is concerned with information exchange among agents. A concrete mutual revision operator is devised by generalizing the iterated revision operator. Likewise, a semantic model is used to show the intuition and many nice properties of the mutual revision operator, and the complexity of its computational model is formally analyzed. Finally, I present a belief update operator, which takes into account two important problems of reasoning about action, i.e., disjunctive updates and domain constraints. Again, the updated operator is presented with both a semantic model and a computational model.

Page generated in 0.0736 seconds