Spelling suggestions: "subject:"metadata."" "subject:"datadata.""
421 |
Estudo sobre aplicação de web semântica e visualização em dados abertos: proposta de portal para transparência em séries históricas de indicadores sociais, econômicos e financeiros, públicos na web / Study on application of semantic web and viewing in open dataGarcia, Paulo Sérgio Rangel 22 November 2011 (has links)
Made available in DSpace on 2016-04-29T14:23:04Z (GMT). No. of bitstreams: 1
Paulo Sergio Rangel Garcia.pdf: 2904735 bytes, checksum: 788a6453214c9f0a41c9629423741347 (MD5)
Previous issue date: 2011-11-22 / This paper aims to discuss a model for constructing a transparent portal that contains historical trends of public social, economical and financial indicators, preferably if collected at e-Gov websites. In order to reach that objective, we researched extensively within Manuel Castells, Tim Berners-Lee, Vinton Cerf, Karin K. Breitman, Martha Gabriel and Fernanda Viegas s work, amongst others, about semantic web, ontologies, search engine optimization and data visualization.
Starting from a study of the historical evolution of internet, we try to understand how was possible to forge this context of highly productive disorganized dynamics that provided the conditions of the massive availability of information and data that we see today. By utilizing concepts and technologies such as metadata, ontologies, web services, software agents, data visualization and optimization of research mechanisms, we discuss techniques and methodologies that could make collecting data as fast as possible feasible, as well as putting it together in a way that either humans or other computer systems may easily understand. We also analyze the current e-Gov services provided by the Brazilian government by evaluating its characteristics, tendencies and possibilities for evolution and growth by utilizing data and research provided by NIC.BR and W3C BR.
Thus, based on research and the author s own experience, this dissertation will try to propose a model for a portal that allows human interaction and integration with other computer systems for historical trends collected via web and formats and provides this data to society in a structured way.
Finally, we indicate future researches that may contribute to the discussion and development of websites and services that apropriate public open data / Este trabalho tem como objetivo discutir um modelo para a construção de portal que dê transparência a séries históricas de indicadores sociais, econômicos e financeiros, públicos e disponíveis na web. Para se chegar a esse objetivo, foram realizados estudos em autores como Manuel Castells, Tim Berners-Lee, Vinton Cerf, Karin K. Breitman, Martha Gabriel e Fernanda Viegas entre outros, sobre web semântica, ontologias, otimização em mecanismos de buscas e visualização de dados.
A partir do estudo da evolução histórica da Internet, buscou-se compreender como foi possível formar esse contexto de dinamismo não ordenado, com alta produção, criando condições para a alta disponibilidade de dados e informações que existem hoje. Abordando conceitos e tecnologias relacionadas à web semântica, ontologias, web services, agentes de software, visualização de dados e otimização dos mecanismos de buscas, discutiu-se técnicas e metodologias que viabilizem a sua construção de maneira a permitir que tais dados sejam coletados e que estejam disponíveis em formatos compreensíveis para seres humanos e integráveis a outros sistemas computacionais. Foi também analisado o
e-Gov atual, especialmente no Brasil, avaliando suas características, tendências e oportunidades de evolução utilizando dados e estudos disponibilizados pelo NIC.BR e W3C BR.
Assim, com base nesses estudos e na experiência do autor, esta dissertação tem o objetivo de propor um modelo de portal que permita a interatividade humana e integração com outros sistemas computacionais, das séries históricas públicas coletadas, formatando e disponibilizando-as de maneira estruturada para a sociedade.
` Por fim, é feita a indicação de pesquisas futuras, as quais venham contribuir para a discussão e o desenvolvimento de sites e serviços que apropriem dados abertos públicos.
|
422 |
Dos dados aos formatos - Um modelo teórico para o design do sistema narrativo no jornalismo digital / From Data to Formats - A theoretical model to design the narrative system in the digital journalism contextBertocchi, Daniela 17 February 2014 (has links)
Já não vemos a possibilidade de sustentar a ideia do formato notícia como o produto final da cadeia de produção jornalística digital. O desajuste que aqui enxergamos não está na notícia, ou em sua estrutura tradicional, ou em sua função social, mas no seu entendimento como produto final, como efeito de uma causa, como desfecho, epílogo ou conclusão de um processo. O desajuste está no termo final, que pressupõe um início. Já não vemos o formato narrativo digital como um produto hereditário de uma cadeia linear e fechada de produção no jornalismo. No lugar do controle e da linearidade, proporemos neste estudo observar a produção narrativa dentro de um processo circular típico do jornalismo pós-industrial. Em vez de fechamentos de formatos estáticos, notaremos continuidades com formatos adaptativos e mais leveza e imponderabilidade no lugar de estruturas narrativas rígidas e sólidas. Esse é, de fato, um estudo que abraça o desconforto e a beleza das incertezas que nascem da relação forma-conteúdo-tela das narrativas digitais no jornalismo. O que nos interessará aqui será criar um instrumento teórico para lidar com essa condição complexa. Assim, propomos um modelo teórico capaz de expandir a compreensão do que seja a narrativa digital jornalística, inaugurando um modo sistêmico de se pensar o desenho das narrativas no jornalismo digital e fornecendo assim um entendimento novo para este fenômeno. Neste estudo o termo \"narrativa\" não irá, portanto, se referir apenas ao que vemos nas telas de nossos dispositivos tecnológicos. Expandimos neste trabalho a perspectiva da narrativa clássica que toma a narrativa jornalística como o agenciamento dos fatos e a colocamos como agenciamento coletivo entre os estratos do sistema narrativo ou, de forma mais reduzida, assumimos a narrativa como sistema narrativo. Expansão, aqui, significa dilatação da visão e não negação ou discordância em relação às construções e percursos teóricos que precedem esta pesquisa. Para a construção do modelo teórico, partimos da narratologia pós-clássica, da moderna teoria dos sistemas e do modelo JDBD (Jornalismo Digital de Base de Dados). Observamos, nesta jornada, que o agenciamento entre os estratos do sistema narrativo realiza-se de forma coletiva por diversos atores: jornalistas, engenheiros, designers, webmasters, especialistas em negócios, arquitetos de informação, usuários, robôs, softwares, algoritmos, entre muitos outros; e que o jornalista atua sobretudo nas camadas de frontend do sistema. Familiarizando-se com camadas mais subterrâneas do sistema narrativo, a antenarrativa (dados e metadados), o jornalista abre oportunidades para melhor comunicar suas histórias no ciberespaço, interfaceando formatos que provoquem experiências ricas para suas audiências. Em nosso entendimento, o jornalista é, portanto, potencialmente um designer da experiência narrativa. / We no longer see the digital narrative format as a product of a linear hereditary and closed production in digital journalism. We no longer see the the form of news, their format, what we see in the screens, as a final product. The imbalance that we see here is not in the news itself or in its traditional structure, or even in its social function, but in understand it as an end product, as an effect or a cause, as an outcome, the epilogue or the conclusion of a process. The misfit is in this notion: the end, which implies a beginning. nstead of control and linearity, we propose in this study to observe the storytelling production inside a circular process, a loop that is typical of a post-industrial journalism. Instead of closing static formats, we will notice continuities with adaptive shapes. More lightness and weightlessness rather than rigid and solid narrative structures. In fact, this is a study that embraces both the discomfort and the beauty of the uncertainties that arise from the relationship between forms, contents and screens when we analyse digital storytelling in journalism context. What concern us here shall be create a theoretical apparatus to address this complex condition. In this way, this study opens up a systemic way to think about the design of narratives in digital journalism and thus provide a new insight of this phenomenon. The term storytelling will not, therefore, refer only to what we see on the screens of our technological devices. Here we expand the perspective of classical narrative (the arrangement of the events) and place it as the collective assemblage among the strata of narrative system. Or in other words: in this study, narrative is narrative system. It is worth to emphasize that when we say \" expansion of perspective,\" we mean dilation of vision and no denial nor disagreement with the theoretical constructs and pathways that precede this research .For the construction of the theoretical model, we start from the post-classical narratology, the modern theory of systems and the JDBD model. We observed, in this journey, the arrangement of the narrative system is held collectively by several actors: journalists, engineers, designers, webmasters, business specialists, information architects, users, robots, software, algorithms, among many others. And the journalist operates in some of the strata of the storytelling system, especially in the frontend layers. If familiar with subterranean layers of narrative system, the antenarrative (data and metadata), journalists provide opportunities to better communicate their stories in cyberspace, interfacing formats which cause rich experiences for its audiences. In our understanding, the journalist is potentially a narrative experience designer.
|
423 |
Dynamic Data Citation Service-Subset Tool for Operational Data ManagementSchubert, Chris, Seyerl, Georg, Sack, Katharina January 2019 (has links) (PDF)
In earth observation and climatological sciences, data and their data services grow on a daily
basis in a large spatial extent due to the high coverage rate of satellite sensors, model calculations, but
also by continuous meteorological in situ observations. In order to reuse such data, especially data
fragments as well as their data services in a collaborative and reproducible manner by citing the origin
source, data analysts, e.g., researchers or impact modelers, need a possibility to identify the exact
version, precise time information, parameter, and names of the dataset used. A manual process would
make the citation of data fragments as a subset of an entire dataset rather complex and imprecise to
obtain. Data in climate research are in most cases multidimensional, structured grid data that can
change partially over time. The citation of such evolving content requires the approach of "dynamic
data citation". The applied approach is based on associating queries with persistent identifiers. These
queries contain the subsetting parameters, e.g., the spatial coordinates of the desired study area or the
time frame with a start and end date, which are automatically included in the metadata of the newly
generated subset and thus represent the information about the data history, the data provenance,
which has to be established in data repository ecosystems. The Research Data Alliance Data Citation
Working Group (RDA Data Citation WG) summarized the scientific status quo as well as the state of
the art from existing citation and data management concepts and developed the scalable dynamic
data citation methodology of evolving data. The Data Centre at the Climate Change Centre Austria
(CCCA) has implemented the given recommendations and offers since 2017 an operational service
on dynamic data citation on climate scenario data. With the consciousness that the objective of this
topic brings a lot of dependencies on bibliographic citation research which is still under discussion,
the CCCA service on Dynamic Data Citation focused on the climate domain specific issues, like
characteristics of data, formats, software environment, and usage behavior. The current effort beyond
spreading made experiences will be the scalability of the implementation, e.g., towards the potential
of an Open Data Cube solution.
|
424 |
Text Augmentation: Inserting markup into natural language text with PPM ModelsYeates, Stuart Andrew January 2006 (has links)
This thesis describes a new optimisation and new heuristics for automatically marking up XML documents, and CEM, a Java implementation, using PPM models. CEM is significantly more general than previous systems, marking up large numbers of hierarchical tags, using n-gram models for large n and a variety of escape methods. Four corpora are discussed, including the bibliography corpus of 14682 bibliographies laid out in seven standard styles using the BibTeX system and marked up in XML with every field from the original BibTeX. Other corpora include the ROCLING Chinese text segmentation corpus, the Computists' Communique corpus and the Reuters' corpus. A detailed examination is presented of the methods of evaluating mark up algorithms, including computation complexity measures and correctness measures from the fields of information retrieval, string processing, machine learning and information theory. A new taxonomy of markup complexities is established and the properties of each taxon are examined in relation to the complexity of marked up documents. The performance of the new heuristics and optimisation are examined using the four corpora.
|
425 |
Pixgis : An Application for Map-centric Management of Photographs with Embedded Spatial MetadataSjödin, Erik January 2006 (has links)
<p>This thesis presents Pixgis; a novel application for map-centric management of photographs. In short Pixgis is an interactive environment in which photographs may be discovered, viewed and managed through maps. With Pixgis finding photographs from a specific location or of a particular structure is as easy as finding the location or structure on a map. As Pixgis simultaneously displays maps, photographs and spatial metadata it also enables users to analyze photographs in new manners. This thesis work illustrates the benefits of applications for map-centric management of photographs, exposes the problems one faces when implementing such applications and presents novel solutions to many of these problems. The thesis also elaborates on spatial metadata and methods for acquisition of photographs with embedded spatial metadata.</p>
|
426 |
Pixgis : An Application for Map-centric Management of Photographs with Embedded Spatial MetadataSjödin, Erik January 2006 (has links)
This thesis presents Pixgis; a novel application for map-centric management of photographs. In short Pixgis is an interactive environment in which photographs may be discovered, viewed and managed through maps. With Pixgis finding photographs from a specific location or of a particular structure is as easy as finding the location or structure on a map. As Pixgis simultaneously displays maps, photographs and spatial metadata it also enables users to analyze photographs in new manners. This thesis work illustrates the benefits of applications for map-centric management of photographs, exposes the problems one faces when implementing such applications and presents novel solutions to many of these problems. The thesis also elaborates on spatial metadata and methods for acquisition of photographs with embedded spatial metadata.
|
427 |
Individual Information Adaptation Based on Content DescriptionWallin, Erik Oskar January 2004 (has links)
<p>Todays increasing information supply raises the needfor more effective and automated information processing whereindividual information adaptation (personalization) is onepossible solution. Earlier computer systems for personalizationlacked the ability to easily define and measure theeffectiveness of personalization efforts. Numerous projectsfailed to live up to the their expectations, and the demand forevaluation increased.</p><p>This thesis presents some underlying concepts and methodsfor implementing personalization in order to increase statedbusiness objectives. A personalization system was developedthat utilizes descriptions of information characteristics(metadata) to perform content based filtering in anon-intrusive way.</p><p>Most of the described measurement methods forpersonalization in the literature are focused on improving theutility for the customer. The evaluation function of thepersonalization system described in this thesis takes thebusiness operators standpoint and pragmatically focuseson one or a few measurable business objectives. In order toverify operation of the personalization system, a functioncalled bifurcation was created. The bifurcation functiondivides the customers stochastically into two or morecontrolled groups with different personalizationconfigurations. Bygiving one of the controlled groups apersonalization configuration that deactivates thepersonalization, a reference group is created. The referencegroup is used to measure quantitatively objectives bycomparison with the groups with active personalization.</p><p>Two different companies had their websites personalized andevaluated: one of Swedens largest recruitment servicesand the second largest Swedish daily newspaper. The purposewith the implementations was to define, measure, and increasethe business objectives. The results of the two case studiesshow that under propitious conditions, personalization can bemade to increase stated business objectives.</p><p><b>Keywords:</b>metadata, semantic web, personalization,information adaptation, one-to-one marketing, evaluation,optimization, personification, customization,individualization, internet, content filtering, automation.</p>
|
428 |
CCS - Collect, Convert and Send : Designing and implementing a system for data portability and media migration to mobile devicesGustafsson, Jonas, Alserin, Fredrik January 2006 (has links)
<p>In this thesis we will identify which are the desired features and functionalities for implementing a system capable of acting as an information bridge for content available in the “wired” Internet to be delivered to mobile devices. We will also explore how to design and build such a system based on the specifications within parts of the MUSIS project. The MUSIS’ system development is used as a base of the work described in this thesis and the experiences from those efforts are used in order to design a system with more focus on data portability and media migration.</p><p>During the development of the MUSIS platform, problems related to system upgrading, i.e. adding new ad-hoc functionalities were discovered. Due to the fact that a user-centred design approach was taken, this was essential in the project. To solve some of these issues, we propose a new component-based system with a high level of scalability and re-usability. We name this system Collect, Convert and Send, CCS. The system shall be seen as a base that can be used as a core system for different projects where interoperability of content between different platforms, devices or systems is important.</p><p>The implementation of the system is based on the use cases and those theoretical aspects and ideas related to component software, interoperability, media migration and metadata in a Web service context. The results of our efforts give some indications that the use of component software gives a foundation for a service-oriented architecture.</p>
|
429 |
Προηγμένες τεχνικές και αλγόριθμοι εξόρυξης γνώσης για την προσωποποίηση της πρόσβασης σε δικτυακούς τόπους / Advanced techniques and algorithms of knowledge mining from Web SitesΓιαννακούδη, Θεοδούλα 16 May 2007 (has links)
Η προσωποποίηση του ιστού είναι ένα πεδίο που έχει κερδίσει μεγάλη προσοχή όχι μόνο στην ερευνητική περιοχή, όπου πολλές ερευνητικές μονάδες έχουν ασχοληθεί με το πρόβλημα από διαφορετικές μεριές, αλλά και στην επιχειρησιακή περιοχή, όπου υπάρχει μία ποικιλία εργαλείων και εφαρμογών που διαθέτουν ένα ή περισσότερα modules στη διαδικασία της εξατομίκευσης. Ο στόχος όλων αυτών είναι, εξερευνώντας τις πληροφορίες που κρύβονται στα logs του εξυπηρετητή δικτύου να ανακαλύψουν τις αλληλεπιδράσεις μεταξύ των επισκεπτών των ιστότοπων και των ιστοσελίδων που περιέχονται σε αυτούς. Οι πληροφορίες αυτές μπορούν να αξιοποιηθούν για τη βελτιστοποίηση των δικτυακών τόπων, εξασφαλίζοντας έτσι αποτελεσματικότερη πλοήγηση για τον επισκέπτη και διατήρηση του πελάτη στην περίπτωση του επιχειρηματικού τομέα. Ένα βασικό βήμα πριν την εξατομίκευση αποτελεί η εξόρυξη χρησιμοποίησης από τον ιστό, ώστε να αποκαλυφθεί τη γνώση που κρύβεται στα log αρχεία ενός web εξυπηρετητή. Εφαρμόζοντας στατιστικές μεθόδους και μεθόδους εξόρυξης δεδομένων στα web log δεδομένα, μπορούν να προσδιοριστούν ενδιαφέροντα πρότυπα που αφορούν τη συμπεριφορά πλοήγησης των χρηστών, όπως συστάδες χρηστών και σελίδων και πιθανές συσχετίσεις μεταξύ web σελίδων και ομάδων χρηστών. Τα τελευταία χρόνια, γίνεται μια προσπάθεια συγχώνευσης του περιεχομένου του ιστού στη διαδικασία εξόρυξης χρησιμοποίησης, για να επαυξηθεί η αποτελεσματικότητα της εξατομίκευσης. Το ενδιαφέρον σε αυτή τη διπλωματική εργασία εστιάζεται στο πεδίο της εξόρυξης γνώσης για τη χρησιμοποίηση δικτυακών τόπων και πώς η διαδικασία αυτή μπορεί να επωφεληθεί από τα χαρακτηριστικά του σημασιολογικού ιστού. Αρχικά, παρουσιάζονται τεχνικές και αλγόριθμοι που έχουν προταθεί τα τελευταία χρόνια για εξόρυξη χρησιμοποίησης από τα log αρχεία των web εξυπηρετητών. Έπειτα εισάγεται και ο ρόλος του περιεχομένου στη διαδικασία αυτή και παρουσιάζονται δύο εργασίες που λαμβάνουν υπόψη και το περιεχόμενο των δικτυακών τόπων: μία τεχνική εξόρυξης χρησιμοποίησης με βάση το PLSA, η οποία δίνει στο τέλος και τη δυνατότητα ενοποίησης του περιεχομένου του ιστού και ένα σύστημα προσωποποίησης το οποίο χρησιμοποιεί το περιεχόμενο του ιστοτόπου για να βελτιώσει την αποτελεσματικότητα της μηχανής παραγωγής προτάσεων. Αφού αναλυθεί θεωρητικά το πεδίο εξόρυξης γνώσης από τα logs μέσα από την περιγραφή των σύγχρονων τεχνικών, προτείνεται το σύστημα ORGAN-Ontology-oRiented usaGe ANalysis- το οποίο αφορά στη φάση της ανάλυσης των log αρχείων και την εξόρυξη γνώσης για τη χρησιμοποίηση των δικτυακών τόπων με άξονα τη σημασιολογία του ιστοτόπου. Τα σημασιολογικά χαρακτηριστικά του δικτυακού τόπου έχουν προκύψει με τεχνικές εξόρυξης δεδομένων από το σύνολο των ιστοσελίδων και έχουν σχολιαστεί από μία OWL οντολογία. Το ORGAN παρέχει διεπαφή για την υποβολή ερωτήσεων σχετικών με την επισκεψιμότητα και τη σημασιολογία των σελίδων, αξιοποιώντας τη γνώση για το site, όπως αναπαρίσταται πάνω στην οντολογία. Περιγράφεται διεξοδικά ο σχεδιασμός, η ανάπτυξη και η πειραματική αξιολόγηση του συστήματος και σχολιάζονται τα αποτελέσματα του. / Web personalization is a domain which has gained great momentum not only in the research area, where many research units have addressed the problem form different perspectives, but also in the industrial area, where a variety of modules for the personalization process is available. The objective is, researching the information hidden in the web server log files to discover the interactions between web sites visitors and web sites pages. This information can be further exploited for web sites optimization, ensuring more effective navigation for the user and client retention in the industrial case. A primary step before the personalization is the web usage mining, where the knowledge hidden in the log files is revealed. Web usage mining is the procedure where the information stored in the Web server logs is processed by applying statistical and data mining techniques such as clustering, association rules discovery, classification, and sequential pattern discovery, in order to reveal useful patterns that can be further analyzed. Recently, there has been an effort to incorporate Web content in the web usage mining process, in order to enhance the effectiveness of personalization. The interest in this thesis is focused on the domain of the knowledge mining for usage of web sites and how this procedure can get the better of attributes of the semantic web. Initially, techniques and algorithms that have been proposed lately in the field of web usage mining are presented. After, the role of the context in the usage mining process is introduced and two relevant works are presented: a usage mining technique based on the PLSA model, which may integrate attributes of the site content, and a personalization system which uses the site content in order to enhance a recommendation engine. After analyzing theoretically the usage mining domain, a new system is proposed, the ORGAN, which is named after Ontology-oRiented usaGe ANalysis. ORGAN concerns the stage of log files analysis and the domain of knowledge mining for the web site usage based on the semantic attributes of the web site. The web site semantic attributes have resulted from the web site pages applying data mining techniques and have been annotated by an OWL ontology. ORGAN provides an interface for queries submission concerning the average level of visitation and the semantics of the web site pages, exploiting the knowledge for the site, as it is derived from the ontology. There is an extensive description of the design, the development and the experimental evaluation of the system.
|
430 |
Klausimynų projektavimo šablonų kalba / A pattern language for questionnaire designŽilinskas, Tomas 16 January 2007 (has links)
This thesis describes a methodology presented in a pattern language for design of questionnaire / survey information systems. It is intended to be used by professional programmers with negligible experience in the domain of questionnaire / survey software. A system of individual and discrete design patterns of various types is interconnected by varying relations enabling the user to navigate effortlessly and to implement a customizable level of expertise contained within the language. The design patterns contain expert knowledge about construction and contents of a universal high-end questionnaire / survey information system. Data structures, graphical user interface and psychological ramifications of questions’ formulations are discussed in detail. The descriptions of two different prototypes of information systems are supplied. They were created using the expertise of the pattern language and serve as a proof of eligibility for it. The innovation in the thesis is driven by the absence of pattern language for questionnaire design, usage of various types of design patterns / relations in the created language and the very design of universal metadata-based questionnaire information system described in the created language.
|
Page generated in 0.0667 seconds