Spelling suggestions: "subject:"orientedarchitecture"" "subject:"revêtementsd'architecture""
211 |
MidGov : middleware para governo eletrônico baseado em gradee computacionais / MidGov : middleware for electronic government based on grid computingSilva, Geraldo Magela, 1985- 10 March 2013 (has links)
Orientador: Edmundo Roberto Mauro Madeira / Dissertação (mestrado) ¿ Universidade Estadual de Campinas, Instituto de Computação / Made available in DSpace on 2018-08-24T01:33:20Z (GMT). No. of bitstreams: 1
Silva_GeraldoMagela_M.pdf: 6211710 bytes, checksum: 5a88c999846d2e6c89c5a5487e7cbed2 (MD5)
Previous issue date: 2013 / Resumo: Agências governamentais ao redor do mundo estão realizando grandes investimentos na utilização de Tecnologia da Informação e Comunicação em suas atividades. Essa tendência, conhecida como Governo eletrônico, impulsiona grande demanda por pesquisas cujo foco principal é o desenvolvimento de aplicações destinadas a um governo mais transparente e colaborativo. Aplicações para esse tipo de cenário introduzem uma série de desafios que precisam ser enfrentados, incluindo maior interoperabilidade entre sistemas, escalabilidade, questões de segurança, entre outros. Nesse sentido, o paradigma de Arquitetura Orientada a Serviços (SOA) apresenta-se como uma interessante proposta para mitigar a heterogeneidade dos serviços prestados pelas diversas entidades envolvidas. Além disso, computação em grade pode ser considerada uma solução promissora para aplicações de middleware em Governo eletrônico, graças à sua alta capacidade de armazenamento e processamento, além de sua recente orientação a serviços, tornando-a uma poderosa ferramenta para aplicações intra-domínio. Considerando esses desafios, este trabalho propõe uma plataforma para aplicações de Governo eletrônico em sistemas em grades utilizando serviços de suporte fornecidos pelo Globus Toolkit 4 (GT4) no contexto da Web Semântica. O trabalho inclui a implementação de um protótipo do middleware e sua validação através de um cenário de aplicação / Abstract: Government agencies around the world are making large investments in the use of Information and Communication Technology in their activities. This trend, known as electronic government, drives a demand for research focused on development of applications aimed at a more transparent and collaborative government. Applications for this type of scenario pose a series of challenges to be faced, including greater interoperability between systems, scalability, and security issues, among others. In this sense, the paradigm of Service-Oriented Architecture (SOA) presents itself as an interesting proposal to mitigate the heterogeneity of services provided by various involved entities. Furthermore, grid computing can be considered a promising solution for middleware applications in e-Government due to its high storage and processing capacity, and also its recent service orientation, making it a powerful tool for intra-domain applications. Considering these challenges, this dissertation proposes a platform for e-Government applications on grid computing, using the support services provided by the Globus Toolkit 4 (GT4) in the context of the Semantic Web. The work includes the implementation of a middleware prototype and its validation through an application scenario / Mestrado / Ciência da Computação / Mestre em Ciência da Computação
|
212 |
Service-Oriented Integration of Component and Organizational MultiAgent Models / Intégration orientée services des approches composants logiciels et systèmes MultiAgents OrganisationnellesAboud, Nour 04 December 2012 (has links)
Les travaux présentés dans cette thèse concernent des problématiques d'architecture logicielle multi-domaines pour le développement d’applications distribuées. Ces applications sont caractérisées aujourd’hui comme des systèmes ouverts, complexes, hétérogènes et à large échelle. Les approches traditionnelles, telles que l’approche orienté objet, n’offrent plus un paradigme de conception suffisant pour appréhender la complexité de tels systèmes. Ces nouvelles tendances ont conduit à l’émergence d’approches de plus haut niveau telles que les approches orientées services, composants ou agents. Chacune de ces approches offrent des intérêts et des caractéristiques propres dans le développement d’applications distribuées. Les services offrent une abstraction et une interopérabilité à large échelle. Abstraction dans le sens où un service permet de spécifier un élément fonctionnel sans préciser comment cet élément est implémenté. Les composants sont une approche robuste basée sur la composition et la réutilisation d’éléments clairement définis par leurs interfaces. Les agents sont eux des éléments présentant un comportement dynamique dirigé par un but et des interactions de haut niveau avec les autres agents formant l’application, vue comme une organisation de services collaboratifs. D’un point de vue conceptuel, le service peut donc être perçu comme le modèle « métier » de l’application, alors que les composants et les agents constituent un modèle d’implémentation. L’étude de ces différents domaines et des modèles associés, a montré que les approches composants et agents sont complémentaires, les points forts d’une approche représentant les faiblesses de l’autre. Face à ce constat, il nous est paru intéressant d’intégrer ces deux approches, au sein d’une même démarche de conception. Cela permet, d’une part, qu’une approche puisse bénéficier des intérêts de l’autre et d’autre part, d’utiliser conjointement des agents et des composants dans la conception d'une même application. La démarche que nous avons adoptée consiste à considérer les services comme pivot d’interaction afin de rendre possible l’interopérabilité des agents et des composants. Pour supporter cette démarche, nous avons défini un processus de conception basé sur l’Ingénierie Des Modèles qui contient quatre modèles conceptuels (Domain Specific language) dont l’intérêt est de mettre l’accent sur les concepts de services et d’interaction. Nous avons ainsi défini un modèle de services, un modèle de composants et un modèle d’agents. Enfin, un modèle mixte appelé CASOM, Component Agent Service Oriented Model, permet de spécifier une application via une combinaison des trois domaines précédents. Ensuite, des règles de correspondances ont été définies entre les quatre modèles pour pouvoir par exemple transformer une spécification agents en une spécification composants ou mixte. L'implémentation de ces transformations a été réalisée en langage ATL (ATLAS Transformation Language). / The presented work considers problems related to multi-domain software architecture for the development of distributed applications. These applications are large-scaled, heterogeneous, open and complex software systems. Traditional approaches such as object-oriented are no longer sufficient to represent such complex systems. These trends lead to the emergence of higher-level approaches such as service-oriented, components or agents. Each one of these approaches offers interests and characteristics in the development of distributed applications. Services provide an abstraction and interoperability in a large scale. Abstraction is in the sense that a service can specify a functional element without specifying how this element is implemented. The components are a robust approach based on composition and reusability through their clearly defined interfaces. Agents are elements which are characterized by dynamic goal directed behaviours and high-level interactions with other agents forming the application, seen as an organization for collaborative services. From a conceptual point of view, the service can be seen as the “business” model of an application, while components and agents are the implementation models. The study of these different domains, with their related models, showed that the components and agents approaches are complementary; the strengths of one approach overcome the weaknesses of the other. Therefore, we are interested in the integration of these two approaches in a single design approach. This allows an approach to benefit from the interests of the other, on one hand and the use of agents and components jointly in the design of an application on the other hand. To reach our objective, we consider services as pivot of interaction between agents and components. The result of our analysis leads us to develop a design process based on Model-Driven Engineering which contains four conceptual models (Domain Specific Languages) with the main interest of focusing on the concepts of services and interaction. We then defined a service, component and agent models. Finally, a hybrid model called CASOM, Component Agent Service Oriented Model, was proposed that allows application specification via a combination of the three domains. Then, mapping rules have been defined between the four models in order to transform agents specification into components specification or mixed. The implementation of these transformations was done in ATL language (ATLAS Transformation Language).
|
213 |
CCS - Collect, Convert and Send : Designing and implementing a system for data portability and media migration to mobile devicesGustafsson, Jonas, Alserin, Fredrik January 2006 (has links)
In this thesis we will identify which are the desired features and functionalities for implementing a system capable of acting as an information bridge for content available in the “wired” Internet to be delivered to mobile devices. We will also explore how to design and build such a system based on the specifications within parts of the MUSIS project. The MUSIS’ system development is used as a base of the work described in this thesis and the experiences from those efforts are used in order to design a system with more focus on data portability and media migration. During the development of the MUSIS platform, problems related to system upgrading, i.e. adding new ad-hoc functionalities were discovered. Due to the fact that a user-centred design approach was taken, this was essential in the project. To solve some of these issues, we propose a new component-based system with a high level of scalability and re-usability. We name this system Collect, Convert and Send, CCS. The system shall be seen as a base that can be used as a core system for different projects where interoperability of content between different platforms, devices or systems is important. The implementation of the system is based on the use cases and those theoretical aspects and ideas related to component software, interoperability, media migration and metadata in a Web service context. The results of our efforts give some indications that the use of component software gives a foundation for a service-oriented architecture.
|
214 |
SOA and QualityPeng, Qian, Fan, Yang Qing January 2008 (has links)
This thesis emphasizes on investigating the relationship between the quality attributes and service oriented architecture (SOA). Due to quality attributes requirements drive the design of software architecture, it is necessary to maintain the positive quality of SOA and improve the negative quality of SOA. This thesis gives an introduction to SOA, Enterprise Service Bus (ESB) and MULE. Then, it covers information on quality of systems and tactics for achieving each quality attribute. Finally, we discuss the quality of SOA in detail, and illustrate how to set up a SOA and how to improve its quality using a case of an order for supermarket. / Order system
|
215 |
DIGITAL TRANSFORMATION : HOW APIS DRIVE BUSINESS MODEL CHANGE AND INNOVATION / DIGITAL TRANSFORMATION : HOW APIS DRIVE BUSINESS MODEL CHANGE AND INNOVATIONHellbe, Simon, Leung, Peter January 2015 (has links)
Over the years, information technology has created opportunities to improve and extend businesses and to start conducting business in new ways. With the evolution of IT, all businesses and industries are becoming increasingly digitized. This process, or coevolution, of IT and business coming together is called digital transformation. One of the recent trends in this digital transformation is the use of application programmable interfaces (APIs). APIs are standardized digital communication interfaces, used for communication and exchange of information between systems, services and devices (such as computers, smartphones and connected machines). API communication is one of the foundational building blocks in recent disruptive technology trends such as mobile and cloud computing. The purpose of this study is to gain an understanding of the business impact that is created in digital transformation related to the use of APIs. To investigate this novel area, an exploratory study is performed where a frame of reference with an exploratory framework is created based on established academic literature. The exploratory framework consists of three main parts which cover the research questions, including Business Drivers, Business Model Change & Innovation and Challenges & Limitations related to API-enabled digital transformation. The framework is used to gather empirical data consisting of two types, interviews (primary data) and contemporary reports (secondary data). Interviews are performed with API-utilizing companies, consulting firms and IT solution providers and contemporary reports are published by consulting and technology research and advisory firms. Two main business drivers are identified in the study. The first is Understanding & Satisfying Customer Needs which is derived from companies experiencing stronger and changing demands for automated, personalized value-adding services. This requires higher degree of integration across channels and organizations. The second driver is Business Agility, which derives from higher requirements on adapting to changing environments while maintaining operational efficiency. Cost Reduction is also mentioned as a third and secondary driver, as a positive side-effect in combination with the other drivers. The identified impact on business models is that business model innovation is mostly happening in the front-end of business model towards customers. Several examples also exist of purely API-enabled businesses that sell services or manage information exchanges over APIs. The challenges and limitations identified are mostly classic challenges of using IT in businesses and not specific to use of APIs, where the general consensus is that IT and business need to become more integrated, and that strategy and governance for API-initiatives need to be established.
|
216 |
Service oriented architecture governance tools within information securityMokgosi, Letlhogonolo 07 June 2012 (has links)
M.Tech. / Service Oriented Architecture has many advantages. For example, organisations can align business with Information Technology, reuse the developed functionality, reduce development and maintain cost for applications. Organisations adopt Service Oriented Architecture with the aim of automating and integrating business processes. However, it has information security vulnerabilities that should be considered. For example, applications exchange information across the Internet, where it can be tampered with. Information security is therefore one of the crucial qualities that need to be satisfied within information systems. This dissertation addresses the issue of information security within Service Oriented Architecture applications. Some organisations rely on Service Oriented Architecture governance tools when securing information in their Service Oriented Architecture environment. However, they may purchase them without investigating whether they include information security. The aim of this dissertation is to analyse whether these tools include information security. Each tool is benchmarked against the five information security services, defined by the ISO 7498/2 document and including identification and authentication, authorisation, confidentiality, integrity and non-repudiation. The dissertation concludes with a table summarising the results. This dissertation offers decision-makers information that can assist them in analysing whether Service Oriented Architecture governance tools includes information security. It also assists organisations to be aware of security vulnerabilities within Service Oriented Architecture applications, and the consequences that may arise if information security measures are ignored.
|
217 |
Interface-Based Software Development / Interface-Based Software DevelopmentRais, Aziz January 2017 (has links)
Even though there are many software development and project management methodologies available, research and practice both show that IT software development projects still fail, and that the quality of software products does not always meet customers' expectations. There might be multiple causes for such failures, but some of these reasons can be seen to influence or create others. Therefore, the larger the project is, the higher its risk of failure, especially if the teams involved work remotely (distributed and outsourced). This increase in project complexity is considered the motivation for this paper. Similarly, there are other factors that can result in a project's failure and customers' dissatisfaction regarding software quality. All such factors identified by research conducted by organizations specializing in this area are analyzed in order to identify a common root of IT project failures. Once the root causes of these failures have been identified and analyzed, the goal of the Interface based software development methodology is to solve them. The solution offered by an Interface based software development methodology is to improve understanding of software requirements and to describe these requirements with interfaces in an object-oriented way. Interface based software development will support and drive development towards service-oriented architecture (SOA) and component-based development (CBD). The goal of interface based software development is to increase software testability and maintainability and to make it more easily feasible to execute various software development processes in parallel.
|
218 |
Podpora webových služeb v prostředí .NET framework / Web services support in the .NET frameworkFischer, Roman January 2008 (has links)
The work targets the very important part of service oriented architecture -- web services. It is focused at one of its practical implementation, from the Microsoft Corporation, especially in the last two releases of .NET Framework. The main target of this work is comprehensive summary and demonstration of Microsoft web services development. The principles of web services and their base in the .NET Framework are described here. At the beginning of this work the web services are put into the scope of service oriented architecture and their main principles are described. Subsequently the .NET Framework is explained with its main principles and with the comparison to the JAVA technology. This part is finished with the explanation of classic ASMX web services. The main part of this work focuses at detailed description of the theoretical and practical aspects of web services in the Windows Communication Foundation and the Windows Workflow Foundation context. Even the development of Windows Forms application is not left out, because they also have done a large progress and they are often consumers of web services. The options of securing web services in the .NET Framework are shown both theoretically and practically including authentication and authorization. The methods of transactional behavior are shown too. The work also focuses at the orchestration of web services including the explanation of workflows with the deployment example. At the end of the work there is revealed how to guarantee interoperability with other technologies, how to manage web services and how to ensure their versioning.
|
219 |
Intéropérabilité sémantique entre les outils de traitement d'images en neuro-imagerie / Semantic interoperability between image processing tools in neuroimagingWali, Bacem 21 June 2013 (has links)
Le domaine de la recherche en neuroimagerie nécessite de pouvoir partager, réutiliser et comparer les outils de traitement d'images des différents laboratoires. Cependant la tâche de partage de traitement sous forme de services et leur composition sous forme de workflow reste une tâche difficile et trop souvent complexe. Ceci est dû dans la plupart des cas à l'hétérogénéité des services et des plateformes qui diffèrent au niveau de leurs conceptions et de leurs implémentations. Nous travaillons dans le cadre du projet NeuroLOG, une initiative cherchant à construire un système fédéré pour le partage de données et d'outils de traitement dans le domaine de la neuroimagerie. Il adopte une approche ontologique pour assurer la médiation et le partage de ressources entre les différents collaborateurs. Notre travail de thèse vise à compléter la médiation pour assurer le partage et la composition des outils de traitement d'images et à fournir aux utilisateurs spécialistes et non-spécialistes du domaine de la neuroimagerie une plateforme de composition de service ergonomique et facile à utiliser. Nous utilisons pour cela les techniques du web sémantique afin de remédier aux différents problèmes d'interopérabilité et de cohérence de ressources utilisées et produites. La première solution proposée se fonde sur une extension de la plateforme OWL-S. Elle a été adaptée aux différents services web de la plateforme de neuroimagerie. On a déduit que finalement les outils qui ne possèdent pas le format de services web et une description conforme au standard WSDL ne peuvent pas être enchaînés sous forme de workflow. A partir de là, nous avons proposé une autre approche pour effectuer la composition de services de traitement d'images. Elle se se fonde sur un nouveau modèle ontologique de composition de services qui répond aux exigences de la neuroimagerie, qui s'articule bien avec l'ontologie de domaine OntoNeuroLOG et qui pourra remédier aux différents problèmes rencontrés lors de l'élaboration de la première approche. Ce travail a permit de remédier à la fois aux problèmes d'hétérogénéité des descripteurs des services et à l'interopérabilité des services selon les contraintes de la neuroimagerie au sein de la plateforme NeuroLOG. / The field of neuroimaging research requires the ability to share, reuse and compare image processing tools coming from different laboratories. However, sharing treatment as services and composing them as workflows, is usually difficult and a complex task. This is due in most cases to the heterogeneity of services and platforms with regards to their conception and their implementation. We work within the NeuroLOG project, which aims at developing a middleware to federate data repositories and to facilitate the sharing and reuse of processing tools to analyze the shared images. It adopts an ontological approach for data and tools mediation and for sharing resources. This work aims to provide tools mediation to enhance the sharing and composition of image processing tools and provide non-specialist and expert users of neuroimaging field with an ergonomic and easy to use composition platform. We have chosen to use the Semantic Web techniques to address the various problems of resource interoperability and consistency. The first proposed solution is based on an extension of the OWL-S framework. It has been adapted to the various web services of our neuroimaging platform. We finally concluded that services that haven't the WSDL standard as descriptor could not be chained as workflow. So, we have proposed a new approach to compose image processing tools. It is based on a new ontological model for service composition that meets the requirements of the neuroimaging domain and the constraints of our domain ontology OntoNeuroLOG and addresses the various problems encountered in the development of the first approach. This work led to solve the two major problems in the composition of services; the heterogeneity of services descriptors and the interoperability of services according to the constraints within the NeuroLOG platform.
|
220 |
Système de gestion de flux pour l'Internet des objets intelligents / Data stream management system for the future internet of thingsBillet, Benjamin 19 March 2015 (has links)
L'Internet des objets (ou IdO) se traduit à l'heure actuelle par l'accroissement du nombre d'objets connectés, c'est-à-dire d'appareils possédant une identité propre et des capacités de calcul et de communication de plus en plus sophistiquées : téléphones, montres, appareils ménagers, etc. Ces objets embarquent un nombre grandissant de capteurs et d'actionneurs leur permettant de mesurer l'environnement et d'agir sur celui-ci, faisant ainsi le lien entre le monde physique et le monde virtuel. Spécifiquement, l'Internet des objets pose plusieurs problèmes, notamment du fait de sa très grande échelle, de sa nature dynamique et de l'hétérogénéité des données et des systèmes qui le composent (appareils puissants/peu puissants, fixes/mobiles, batteries/alimentations continues, etc.). Ces caractéristiques nécessitent des outils et des méthodes idoines pour la réalisation d'applications capables (i) d'extraire des informations utiles depuis les nombreuses sources de données disponibles et (ii) d'interagir aussi bien avec l'environnement, au moyen des actionneurs, qu'avec les utilisateurs, au moyen d'interfaces dédiées. Dans cette optique, nous défendons la thèse suivante : en raison de la nature continue des données (mesures physiques, évènements, etc.) et leur volume, il est important de considérer (i) les flux comme modèle de données de référence de l'Internet des objets et (ii) le traitement continu comme modèle de calcul privilégié pour transformer ces flux. En outre, étant donné les préoccupations croissantes relatives à la consommation énergétique et au respect de la vie privée, il est préférable de laisser les objets agir au plus près des utilisateurs, si possible de manière autonome, au lieu de déléguer systématiquement l'ensemble des tâches à de grandes entités extérieures telles que le cloud. À cette fin, notre principale contribution porte sur la réalisation d'un système distribué de gestion de flux de données pour l'Internet des objets. Nous réexaminons notamment deux aspects clés du génie logiciel et des systèmes distribués : les architectures de services et le déploiement. Ainsi, nous apportons des solutions (i) pour l'accès aux flux de données sous la forme de services et (ii) pour le déploiement automatique des traitements continus en fonction des caractéristiques des appareils. Ces travaux sont concrétisés sous la forme d'un intergiciel, Dioptase, spécifiquement conçu pour être exécuté directement sur les objets et les transformer en fournisseurs génériques de services de calcul et de stockage.Pour valider nos travaux et montrer la faisabilité de notre approche, nous introduisons un prototype de Dioptase dont nous évaluons les performances en pratique. De plus, nous montrons que Dioptase est une solution viable, capable de s'interfacer avec les systèmes antérieurs de capteurs et d'actionneurs déjà déployés dans l'environnement. / The Internet of Things (IoT) is currently characterized by an ever-growing number of networked Things, i.e., devices which have their own identity together with advanced computation and networking capabilities: smartphones, smart watches, smart home appliances, etc. In addition, these Things are being equipped with more and more sensors and actuators that enable them to sense and act on their environment, enabling the physical world to be linked with the virtual world. Specifically, the IoT raises many challenges related to its very large scale and high dynamicity, as well as the great heterogeneity of the data and systems involved (e.g., powerful versus resource-constrained devices, mobile versus fixed devices, continuously-powered versus battery-powered devices, etc.). These challenges require new systems and techniques for developing applications that are able to (i) collect data from the numerous data sources of the IoT and (ii) interact both with the environment using the actuators, and with the users using dedicated GUIs. To this end, we defend the following thesis: given the huge volume of data continuously being produced by sensors (measurements and events), we must consider (i) data streams as the reference data model for the IoT and (ii) continuous processing as the reference computation model for processing these data streams. Moreover, knowing that privacy preservation and energy consumption are increasingly critical concerns, we claim that all the Things should be autonomous and work together in restricted areas as close as possible to the users rather than systematically shifting the computation logic into powerful servers or into the cloud. For this purpose, our main contribution can be summarized as designing and developing a distributed data stream management system for the IoT. In this context, we revisit two fundamental aspects of software engineering and distributed systems: service-oriented architecture and task deployment. We address the problems of (i) accessing data streams through services and (ii) deploying continuous processing tasks automatically, according to the characteristics of both tasks and devices. This research work lead to the development of a middleware layer called Dioptase, designed to run on the Things and abstract them as generic devices that can be dynamically assigned communication, storage and computation tasks according to their available resources. In order to validate the feasability and the relevance of our work, we implemented a prototype of Dioptase and evaluated its performance. In addition, we show that Dioptase is a realistic solution which can work in cooperation with legacy sensor and actuator networks currently deployed in the environment.
|
Page generated in 0.1001 seconds