• Refine Query
  • Source
  • Publication year
  • to
  • Language
  • 29
  • 4
  • 4
  • 3
  • 1
  • 1
  • 1
  • 1
  • Tagged with
  • 48
  • 48
  • 48
  • 9
  • 8
  • 8
  • 7
  • 7
  • 6
  • 6
  • 6
  • 6
  • 6
  • 5
  • 5
  • About
  • The Global ETD Search service is a free service for researchers to find electronic theses and dissertations. This service is provided by the Networked Digital Library of Theses and Dissertations.
    Our metadata is collected from universities around the world. If you manage a university/consortium/country archive and want to be added, details can be found on the NDLTD website.
11

Clustering Multilingual Documents: A Latent Semantic Indexing Based Approach

Lin, Chia-min 09 February 2006 (has links)
Document clustering automatically organizes a document collection into distinct groups of similar documents on the basis of their contents. Most of existing document clustering techniques deal with monolingual documents (i.e., documents written in one language). However, with the trend of globalization and advances in Internet technology, an organization or individual often generates/acquires and subsequently archives documents in different languages, thus creating the need for multilingual document clustering (MLDC). Motivated by its significance and need, this study designs a Latent Semantic Indexing (LSI) based MLDC technique. Our empirical evaluation results show that the proposed LSI-based multilingual document clustering technique achieves satisfactory clustering effectiveness, measured by both cluster recall and cluster precision.
12

Probabilistic Latent Semantic Analysis Based Framework For Hybrid Social Recommender Systems

Eryol, Erkin 01 June 2010 (has links) (PDF)
Today, there are user annotated internet sites, user interaction logs, online user communities which are valuable sources of information concerning the personalized recommendation problem. In the literature, hybrid social recommender systems have been proposed to reduce the sparsity of the usage data by integrating the user related information sources together. In this thesis, a method based on probabilistic latent semantic analysis is used as a framework for a hybrid social recommendation system. Different data hybridization approaches on probabilistic latent semantic analysis are experimented. Based on this flexible probabilistic model, network regularization and model blending approaches are applied on probabilistic latent semantic analysis model as a solution for social trust network usage throughout the collaborative filtering process. The proposed model has outperformed the baseline methods in our experiments. As a result of the research, it is shown that the proposed methods successfully model the rating and social trust data together in a theoretically principled way.
13

Text Summarization Using Latent Semantic Analysis

Ozsoy, Makbule Gulcin 01 February 2011 (has links) (PDF)
Text summarization solves the problem of presenting the information needed by a user in a compact form. There are different approaches to create well formed summaries in literature. One of the newest methods in text summarization is the Latent Semantic Analysis (LSA) method. In this thesis, different LSA based summarization algorithms are explained and two new LSA based summarization algorithms are proposed. The algorithms are evaluated on Turkish and English documents, and their performances are compared using their ROUGE scores.
14

A Reference Architecture for Providing Latent Semantic Analysis Applications in Distributed Systems. Diploma Thesis

Dietl, Reinhard 12 1900 (has links) (PDF)
With the increasing availability of storage and computing power, Latent Semantic Analysis (LSA) has gained more and more significance in practice over the last decade. This diploma thesis aims to develop a reference architecture which can be utilised to provide LSA based applications in a distributed system. It outlines the underlying problems of generation, processing and storage of large data objects resulting from LSA operations, the problems arising from bringing LSA into a distributed context, suggests an architecture for the software components necessary to perform the tasks, and evaluates the applicability to real world scenarios, including the implementation of a classroom scenario as a proof-of-concept. (author's abstract) / Series: Theses / Institute for Statistics and Mathematics
15

Supporting students in the analysis of case studies for professional ethics education

2015 January 1900 (has links)
Intelligent tutoring systems and computer-supported collaborative environments have been designed to enhance human learning in various domains. While a number of solid techniques have been developed in the Artificial Intelligence in Education (AIED) field to foster human learning in fundamental science domains, there is still a lack of evidence about how to support learning in so-called ill-defined domains that are characterized by the absence of formal domain theories, uncertainty about best solution strategies and teaching practices, and learners' answers represented through text and argumentation. This dissertation investigates how to support students' learning in the ill-defined domain of professional ethics through a computer-based learning system. More specifically, it examines how to support students in the analysis of case studies, which is a common pedagogical practice in the ethics domain. This dissertation describes our design considerations and a resulting system called Umka. In Umka learners analyze case studies individually and collaboratively that pose some ethical or professional dilemmas. Umka provides various types of support to learners in the analysis task. In the individual analysis it provides various kinds of feedback to arguments of learners based on predefined system knowledge. In the collaborative analysis Umka fosters learners' interactions and self-reflection through system suggestions and a specifically designed visualization. The system suggestions offer learners the chance to consider certain helpful arguments of their peers, or to interact with certain helpful peers. The visualization highlights similarities and differences between the learners' positions, and illustrates the learners' level of acceptance of each other's positions. This dissertation reports on a series of experiments in which we evaluated the effectiveness of Umka's support features, and suggests several research contributions. Through this work, it is shown that despite the ill-definedness of the ethics domain, and the consequent complications of text processing and domain modelling, it is possible to build effective tutoring systems for supporting students' learning in this domain. Moreover, the techniques developed through this research for the ethics domain can be readily expanded to other ill-defined domains, where argument, qualitative analysis, metacognition and interaction over case studies are key pedagogical practices.
16

Deep Web Collection Selection

King, John Douglas January 2004 (has links)
The deep web contains a massive number of collections that are mostly invisible to search engines. These collections often contain high-quality, structured information that cannot be crawled using traditional methods. An important problem is selecting which of these collections to search. Automatic collection selection methods try to solve this problem by suggesting the best subset of deep web collections to search based on a query. A few methods for deep Web collection selection have proposed in Collection Retrieval Inference Network system and Glossary of Servers Server system. The drawback in these methods is that they require communication between the search broker and the collections, and need metadata about each collection. This thesis compares three different sampling methods that do not require communication with the broker or metadata about each collection. It also transforms some traditional information retrieval based techniques to this area. In addition, the thesis tests these techniques using INEX collection for total 18 collections (including 12232 XML documents) and total 36 queries. The experiment shows that the performance of sample-based technique is satisfactory in average.
17

Educational Technology: A Comparison of Ten Academic Journals and the New Media Consortium Horizon Reports for the Period of 2000-2017

Morel, Gwendolyn 12 1900 (has links)
This exploratory and descriptive study provides an increased understanding of the topics being explored in both published research and industry reporting in the field of educational technology. Although literature in the field is plentiful, the task of synthesizing the information for practical use is a massive undertaking. Latent semantic analysis was used to review journal abstracts from ten highly respected journals and the New Media Consortium Horizon Reports to identify trends within the publications. As part of the analysis, 25 topics and technologies were identified in the combined corpus of academic journals and Horizon Reports. The journals tended to focus on pedagogical issues whereas the Horizon Reports tended to focus on technological aspects in education. In addition to differences between publication types, trends over time are also described. Findings may assist researchers, practitioners, administrators, and policy makers with decision-making in their respective educational areas.
18

Measuring Semantic Relatedness Using Salient Encyclopedic Concepts

Hassan, Samer 08 1900 (has links)
While pragmatics, through its integration of situational awareness and real world relevant knowledge, offers a high level of analysis that is suitable for real interpretation of natural dialogue, semantics, on the other end, represents a lower yet more tractable and affordable linguistic level of analysis using current technologies. Generally, the understanding of semantic meaning in literature has revolved around the famous quote ``You shall know a word by the company it keeps''. In this thesis we investigate the role of context constituents in decoding the semantic meaning of the engulfing context; specifically we probe the role of salient concepts, defined as content-bearing expressions which afford encyclopedic definitions, as a suitable source of semantic clues to an unambiguous interpretation of context. Furthermore, we integrate this world knowledge in building a new and robust unsupervised semantic model and apply it to entail semantic relatedness between textual pairs, whether they are words, sentences or paragraphs. Moreover, we explore the abstraction of semantics across languages and utilize our findings into building a novel multi-lingual semantic relatedness model exploiting information acquired from various languages. We demonstrate the effectiveness and the superiority of our mono-lingual and multi-lingual models through a comprehensive set of evaluations on specialized synthetic datasets for semantic relatedness as well as real world applications such as paraphrase detection and short answer grading. Our work represents a novel approach to integrate world-knowledge into current semantic models and a means to cross the language boundary for a better and more robust semantic relatedness representation, thus opening the door for an improved abstraction of meaning that carries the potential of ultimately imparting understanding of natural language to machines.
19

Lexical semantic richness : effect on reading comprehension and on readers' hypotheses about the meanings of novel words

Duff, Dawna Margaret 01 May 2015 (has links)
Purpose: This study investigates one possible reason for individual differences in vocabulary learning from written context. A Latent Semantic Analysis (LSA) model is used to motivate the prediction of a causal relationship between semantic knowledge for words in a text and the quality of their hypotheses about the semantics of novel words, an effect mediated by reading comprehension. The purpose of this study was to test this prediction behaviorally, using a within subject repeated measures design to control for other variables affecting semantic word learning. Methods: Participants in 6th grades (n=23) were given training to increase semantic knowledge of words from one of two texts, counterbalanced across participants. After training, participants read untreated and treated texts, which contained six nonword forms. Measures were taken of reading comprehension (RC) and the quality of the readers' hypotheses about the semantics of the novel words (HSNW). Text difficulty and semantic informativeness of the texts about nonwords were controlled. Results: All participants had increases in semantic knowledge of taught words after intervention. For the group as a whole, RC scores were significantly higher in the treated than untreated condition, but HSNW scores were not significantly higher in the treated than untreated condition. Reading comprehension ability was a significant moderator of the effect of treatment on HSNW. A subgroup of participants with lower scores on a standardized reading comprehension measure (n=6) had significantly higher HSNW and RC scores in the treated than untreated condition. Participants with higher standardized reading comprehension scores (n=17) showed no effect of treatment on either RC or HSNW. Difference scores for RC and difference scores for HSNW were strongly related, indicating that within subjects, there is a relationship between RC and HSNW. Conclusions: The results indicate that for a subgroup of readers with weaker reading comprehension, intervention to enhance lexical semantic richness had a substantial and significant effect on both their reading comprehension and on the quality of hypotheses that they generated about the meanings of novel words. Neither effect was found for a subgroup of readers with stronger reading comprehension. Clinical and educational implications are discussed.
20

Simulating Expert Clinical Comprehension: Adapting Latent Semantic Analysis to Accurately Extract Clinical Concepts From Psychiatric Narrative

Cohen, Trevor, Blatter, Brett, Patel, Vimla 01 December 2008 (has links)
Cognitive studies reveal that less-than-expert clinicians are less able to recognize meaningful patterns of data in clinical narratives. Accordingly, psychiatric residents early in training fail to attend to information that is relevant to diagnosis and the assessment of dangerousness. This manuscript presents cognitively motivated methodology for the simulation of expert ability to organize relevant findings supporting intermediate diagnostic hypotheses. Latent Semantic Analysis is used to generate a semantic space from which meaningful associations between psychiatric terms are derived. Diagnostically meaningful clusters are modeled as geometric structures within this space and compared to elements of psychiatric narrative text using semantic distance measures. A learning algorithm is defined that alters components of these geometric structures in response to labeled training data. Extraction and classification of relevant text segments is evaluated against expert annotation, with system-rater agreement approximating rater-rater agreement. A range of biomedical informatics applications for these methods are suggested.

Page generated in 0.1156 seconds