• Refine Query
  • Source
  • Publication year
  • to
  • Language
  • 147
  • 38
  • 16
  • 13
  • 8
  • 7
  • 7
  • 5
  • 4
  • 4
  • 4
  • 3
  • 2
  • 2
  • 1
  • Tagged with
  • 301
  • 301
  • 76
  • 60
  • 55
  • 47
  • 43
  • 38
  • 38
  • 37
  • 37
  • 33
  • 31
  • 30
  • 29
  • About
  • The Global ETD Search service is a free service for researchers to find electronic theses and dissertations. This service is provided by the Networked Digital Library of Theses and Dissertations.
    Our metadata is collected from universities around the world. If you manage a university/consortium/country archive and want to be added, details can be found on the NDLTD website.
31

Visualization for frequent pattern mining

Carmichael, Christopher Lee 03 April 2013 (has links)
Data mining algorithms analyze and mine databases for discovering implicit, previously unknown and potentially useful knowledge. Frequent pattern mining algorithms discover sets of database items that often occur together. Many of the frequent pattern mining algorithms represent the discovered knowledge in the form of a long textual list containing these sets of frequently co-occurring database items. As the amount of discovered knowledge can be large, it may not be easy for most users to examine and understand such a long textual list of knowledge. In my M.Sc. thesis, I represent both the original database and the discovered knowledge in pictorial form. Specifically, I design a new interactive visualization system for viewing the original transaction data (which are then fed into the frequent pattern mining engine) and for revealing the interesting knowledge discovered from the transaction data in the form of mined patterns.
32

Criminal careers and the crime drop in Scotland, 1989-2011 : an exploration of conviction trends across age and sex

Matthews, Benjamin Michael January 2017 (has links)
Rates of recorded crime have been falling in many countries in Western Europe, including Scotland, since the early 1990s. This marks the reversal of a trend of increasing levels of crime seen since the 1950s. Despite this important recent change, most analyses of the ‘crime drop’ have focused on recorded crime or victimisation rates aggregated to national or regional level. It is little known how patterns of offending or conviction have changed at the individual level. As a result it is not known how the crime drop is manifest in changing offending or conviction rates, or how patterns of criminal careers have changed over this period. The aim of this thesis is to explore trends in convictions across a number of criminal careers parameters – the age-crime curve, prevalence and frequency, polarisation and conviction pathways – over the course of the crime drop in Scotland. The results presented here are based on a secondary analysis of the Scottish Offenders Index, a census of convictions in Scottish courts, between 1989 and 2011. Analysis is conducted using a range of descriptive statistical techniques to examine change across age, sex and time. Change in the age-crime curve is analysed using data visualisation techniques and descriptive statistics. Standardisation and decomposition analysis is used to analyse the effects of prevalence, frequency and population change. Trends in conviction are also examined between groups identified statistically using Latent Class Analysis to assess the polarisation of convictions, and trends in the movement between these groups over time provides an indication of changing pathways of conviction. This thesis finds a sharp contrast between falling rates of conviction for young people, particularly young men, and increases in conviction rates for those between their mid-twenties and mid-forties, with distinct periods of change between 1989- 2000, 2000-2007 and 2007-2011. These trends are driven primarily by changes in the prevalence of conviction, and result in an increasingly even distribution of convictions over age. Analysis across latent classes shows some evidence of convictions becoming less polarised for younger men and women but increasingly polarised for older men and women. Similarities in trends analysed across latent classes between men and women of the same age suggest that the process driving these trends is broadly similar within age groups. Increases in conviction rates for those over 21 are explained by both greater onset of conviction and higher persistence in conviction, particularly between 1998 and 2004. The results of this thesis suggest that explanations of the crime drop must have a greater engagement with contrasting trends across age and sex to be able to properly explain falling conviction rates. These results also reinforce the need for criminal careers research to better understand the impact of recent changes social context on patterns of convictions over people’s lives. The distinct periods identified in these results suggest a potential effect of changes in operation of the justice system in Scotland leading to high rates of convictions in the early 2000s. However, the descriptive focus of this analysis and its reliance upon administrative data from a single country mean this thesis cannot claim to definitively explain these trends. As a result, replication of this research in another jurisdiction is encouraged to assess whether trends identified are particular to Scotland.
33

Visual Analytics Tool for the Global Change Assessment Model

January 2015 (has links)
abstract: The Global Change Assessment Model (GCAM) is an integrated assessment tool for exploring consequences and responses to global change. However, the current iteration of GCAM relies on NetCDF file outputs which need to be exported for visualization and analysis purposes. Such a requirement limits the uptake of this modeling platform for analysts that may wish to explore future scenarios. This work has focused on a web-based geovisual analytics interface for GCAM. Challenges of this work include enabling both domain expert and model experts to be able to functionally explore the model. Furthermore, scenario analysis has been widely applied in climate science to understand the impact of climate change on the future human environment. The inter-comparison of scenario analysis remains a big challenge in both the climate science and visualization communities. In a close collaboration with the Global Change Assessment Model team, I developed the first visual analytics interface for GCAM with a series of interactive functions to help users understand the simulated impact of climate change on sectors of the global economy, and at the same time allow them to explore inter comparison of scenario analysis with GCAM models. This tool implements a hierarchical clustering approach to allow inter-comparison and similarity analysis among multiple scenarios over space, time, and multiple attributes through a set of coordinated multiple views. After working with this tool, the scientists from the GCAM team agree that the geovisual analytics tool can facilitate scenario exploration and enable scientific insight gaining process into scenario comparison. To demonstrate my work, I present two case studies, one of them explores the potential impact that the China south-north water transportation project in the Yangtze River basin will have on projected water demands. The other case study using GCAM models demonstrates how the impact of spatial variations and scales on similarity analysis of climate scenarios varies at world, continental, and country scales. / Dissertation/Thesis / Masters Thesis Computer Science 2015
34

Interpretations of Data in Ethical vs. Unethical Data Visualizations

January 2017 (has links)
abstract: This paper presents the results of an empirical analysis of deceptive data visualizations paired with explanatory text. Data visualizations are used to communicate information about important social issues to large audiences and are found in the news, social media, and the Internet (Kirk, 2012). Modern technology and software allow people and organizations to easily produce and publish data visualizations, contributing to data visualizations becoming more prevalent as a means of communicating important information (Sue & Griffin, 2016). Ethical transgressions in data visualizations are the intentional or unintentional use of deceptive techniques with the potential of altering the audience’s understanding of the information being presented (Pandey et al., 2015). While many have discussed the importance of ethics in data visualization, scientists have only recently started to look at how deceptive data visualizations affect the reader. This study was administered as an on-line user survey and was designed to test the deceptive potential of data visualizations when they are accompanied by a paragraph of text. The study consisted of a demographic questionnaire, chart familiarity assessment, and data visualization survey. A total of 256 participants completed the survey and were evenly distributed between a control (non-deceptive) survey or a test (deceptive) survey in which participant were asked to observe a paragraph of text and data visualization paired together. Participants then answered a question relevant to the observed information to measure how they perceived the information to be. The individual differences between demographic groups and their responses were analyzed to understand how these groups reacted to deceptive data visualizations compared to the control group. The results of the study confirmed that deceptive techniques in data visualizations caused participants to misinterpret the information in the deceptive data visualizations even when they were accompanied by a paragraph of explanatory text. Furthermore, certain demographics and comfort levels with chart types were more susceptible to certain types of deceptive techniques. These results highlight the importance of education and practice in the area of data visualizations to ensure deceptive practices are not utilized and to avoid potential misinformation, especially when information can be called into question. / Dissertation/Thesis / Masters Thesis Technical Communication 2017
35

VRMol - um ambiente virtual distribuído para visualização e análise de moléculas de proteínas. / VRMol - a distributed virtual enviroment to visualize and analyze molecules of proteins.

Ildeberto Aparecido Rodello 12 February 2003 (has links)
Este trabalho utiliza conceitos de Realidade Virtual e Sistemas Distribuídos para desenvolver um Ambiente Virtual Distribuído para visualização e análise de moléculas de proteínas, denominado VRMol. O sistema foi implementado com a linguagem Java, incluindo as APls Java 3D e Java RMI, visando permitir que pesquisadores geograficamente dispersos troquem informações de uma maneira rápida e eficiente, acelerando a pesquisa e discussão remotas. Assim, foram desenvolvidos uma interface gráfica com Java 3D e um conjunto de métodos para troca de mensagens de acordo com o modelo de comunicação cliente/servidor, com Java RMI. Além disso, o sistema também permite a utilização de alguns dispositivos de entrada não convencionais como joystick e luvas. / This work use the Virtual Reality and the Distributed Systems concepts to develop a Distributed Virtual Environment to visualize and analyze molecules of proteins, called VRMol. The system was implemented with the Java programming language, including the Java 3D and Java RMI APIs, aiming to allow geographically disperse researches exchange information in a quick and efficient way, speeding up the remote research and discussion. Thus, was developed a graphical interface with Java 3D and a set of methods to exchange messages according to a client/server communication model with Java RMI. Furthermore, the system also allows the use of some non-conventional input devices as joysticks and gloves.
36

A visual training based approach to surface inspection

Niskanen, M. (Matti) 18 June 2003 (has links)
Abstract Training a visual inspection device is not straightforward but suffers from the high variation in material to be inspected. This variation causes major difficulties for a human, and this is directly reflected in classifier training. Many inspection devices utilize rule-based classifiers the building and training of which rely mainly on human expertise. While designing such a classifier, a human tries to find the questions that would provide proper categorization. In training, an operator tunes the classifier parameters, aiming to achieve as good classification accuracy as possible. Such classifiers require lot of time and expertise before they can be fully utilized. Supervised classifiers form another common category. These learn automatically from training material, but rely on labels that a human has set for it. However, these labels tend to be inconsistent and thus reduce the classification accuracy achieved. Furthermore, as class boundaries are learnt from training samples, they cannot in practise be later adjusted if needed. In this thesis, a visual based training method is presented. It avoids the problems related to traditional training methods by combining a classifier and a user interface. The method relies on unsupervised projection and provides an intuitive way to directly set and tune the class boundaries of high-dimensional data. As the method groups the data only by the similarities of its features, it is not affected by erroneous and inconsistent labelling made for training samples. Furthermore, it does not require knowledge of the internal structure of the classifier or iterative parameter tuning, where a combination of parameter values leading to the desired class boundaries are sought. On the contrary, the class boundaries can be set directly, changing the classification parameters. The time need to take such a classifier into use is small and tuning the class boundaries can happen even on-line, if needed. The proposed method is tested with various experiments in this thesis. Different projection methods are evaluated from the point of view of visual based training. The method is further evaluated using a self-organizing map (SOM) as the projection method and wood as the test material. Parameters such as accuracy, map size, and speed are measured and discussed, and overall the method is found to be an advantageous training and classification scheme.
37

Cartogram Visualization: Methods, Applications, and Effectiveness

Nusrat, Sabrina, Nusrat, Sabrina January 2017 (has links)
Cartograms are value-by-area maps which modify geographic regions, such as countries, in proportion to some variable of interest, such as population. These are popular georeferenced data visualizations that have been used for over a century to illustrate patterns and trends in the world around us. A wide variety of cartogram types exist, that were designed to optimize different cartogram dimensions, such as geographic accuracy and statistical accuracy. This work surveys cartogram research in visualization, cartography and geometry, covering a broad spectrum of different cartogram types: from the traditional rectangular cartograms, to Dorling and diffusion cartograms. Based on prior work in visualization and cartography, I propose a task taxonomy of cartograms, and describe a study of cartograms based on quantitative metric-based comparisons, task-based time-and-error evaluation, and subjective preference and feedback analysis. For these evaluations, I considered four major types of cartograms which allowed us to compare and analyze the evaluation strategies and discuss the implications of the surprising outcomes. In the context of maps, the ability to recall information shown in the map is one of the important factors in determining effectiveness. In spite of some early studies that involved cartograms, the memorability of different cartogram types has not been investigated. In order to create effective data presentations, we first need to understand what makes a visualization memorable. I investigate the memorability of contiguous and Dorling cartograms, both in terms of recognition of the map and recall of data. Finally, I describe bivariate cartograms, a technique specifically designed to allow for the simultaneous comparison of two geo-statistical variables. Traditional cartograms are designed to show only a single statistical variable, but in practice, it is often useful to show two variables (e.g., the total sales for two competing companies) simultaneously. Bivariate cartograms make it easy to find geographic patterns and outliers in a pre-attentive way. They are most effective for showing two variables from the same domain (e.g., population in two different years, sales for two different companies), although they can also be used for variables from different domains (e.g., population and income). I also describe a small-scale evaluation of the proposed techniques that indicates bivariate cartograms are especially effective for finding geo-statistical patterns, trends and outliers.
38

Towards a Cloud-based Data Analysis and Visualization System

Li, Zhongli January 2016 (has links)
In recent years, increasing attentions are paid on developing exceptional technologies for efficiently processing massive collection of heterogeneous data generated by different kinds of sensors. While we have observed great successes of utilizing big data in many innovative applications, the need on integrating information poses new challenges caused by the heterogeneity of the data. In this thesis, we target at geo-tagged data, and propose a cloud based platform named City Digital Pulse (CDP), where a unified mechanism and extensible architecture are provided to facilitate the various aspects in big data analysis, ranging from data acquisition to data visualization. We instantiate the proposed system using multi-model data collected from two social platforms, Twitter and Instagram, which include plenty of geo-tagged messages. Data analysis is performed to detect human affections from the user uploaded content. The emotional information in big social data can be uncovered by using a multi-dimension visualization interface, based on which users can easily grasp the evolving of human affective status within a given geographical area, and interact with the system. This offers costless opportunities to improve the decision making in many critical areas. Both the proposed architecture and algorithm are empirically demonstrated to be able to achieve real-time big data analysis.
39

A Flexible Service-Oriented Approach to Address Hydroinformatic Challenges in Large-Scale Hydrologic Predictions

Souffront Alcantara, Michael Antonio 01 December 2018 (has links)
Water security is defined as a combination of water for achieving our goals as a society, and an acceptable level of water-related risks. Hydrologic modeling can be used to predict streamflow and aid in the decision-making process with the goal of attaining water security. Developed countries usually have their own hydrologic models; however, developing countries often lack hydrologic models due to factors such as the maintenance, computational costs, and technical capacity needed to run models. A global streamflow prediction system (GSPS) would help decrease vulnerabilities in developing countries and fill gaps in areas where no local models exist by providing extensive results that can be filtered for specific locations. The development of a GSPS has been deemed a grand challenge of the hydrologic community. To this end, many scientists and engineers have started to develop large-scale systems to an acceptable degree of success. Renowned models like the Global Flood Awareness System (GloFAS), the US National Water Model (NWM), and NASA's Land Assimilation System (LDAS) are proof that our ability to model large areas has improved remarkably. Even so, during this evolution the hydrologic community has started to realize that having a large-scale forecasting system does not make it immediately useful. New hydroinformatic challenges have surfaced that prevent these models from reaching their full potential. I have divided these challenges in four main categories: big data, data communication, adoption, and validation. I present a description of the background leading to the development of a GSPS including existing models, and the components needed to create an operational system. A case study with the NWM is also presented where I address the big data and data communication challenges by developing cyberinfrastructure and accessibility tools such as web applications and services. Finally, I used the GloFAS-RAPID model to create a forecasting system covering Africa, North America, South America, and South Asia using a service-oriented approach that includes the development of web applications, and services for providing improved data accessibility, and helping address adoption and validation challenges. I have developed customized services in collaboration with countries that include Argentina, Bangladesh, Colombia, Peru, Nepal, and the Dominican Republic. I also conducted validation tests to ensure that results are acceptable. Overall, a model-agnostic approach to operationalize a GSPS and provide meaningful results at the local level is provided with the potential to allow decision makers to focus on solving some of the most pressing water-related issues we face as a society.
40

JupyterLab_Voyager: A Data Visualization Enhancement in JupyterLab

Zhang, Ji 01 June 2018 (has links)
With the emergence of big data, scientific data analysis and visualization (DAV) tools are critical components of the data science software ecosystem; the usability of these tools is becoming extremely important to facilitate next-generation scientific discoveries. JupyterLab has been considered as one of the best polyglot, web-based, open-source data science tools. As the next phase of extensible interface for the classic iPython Notebooks, this tool supports interactive data science and scientific computing across multiple programming languages with great performances. Despite these advantages, previous heuristics evaluation studies have shown that JupyterLab has some significant flaws in the data visualization side. The current DAV system in JupyterLab heavily relies on users’ understanding and familiarity with certain visualization libraries, and doesn’t support the golden visual-information-seeking mantra of “overview first, zoom and filter, then details-on-demand”. These limitations often lead to a workflow bottleneck at the start of a project. In this thesis, we present ‘JupyterLab_Voyager’, an extension for JupyterLab that provides a graphical user interface (GUI) for data visualization operations and couples faceted browsing with visualization recommendation to support exploration of multivariate, tabular data, as a solution to improve the usability of the DAV system. The new plugin works with various types of datasets in the JupyterLab ecosystem; using the plugin you can perform a high-level graphical analysis of fields within your dataset sans coding without leaving the JupyterLab environment. It helps analysts learn about the dataset and engage in both open-ended exploration and target specific answers from the dataset. User testings and evaluations demonstrated that this implementation has good usability and significantly improves the DAV system in JupyterLab.

Page generated in 0.0562 seconds