51 |
Creating Knowledge Graphs using Distributional Semantic ModelsSandelius, Hugo January 2016 (has links)
This report researches a method for creating knowledge graphs, a specific way of structuring information, using distributional semantic models. Two different algorithms for selecting graph edges and two different algorithms for labelling edges are tried, and variations of those are evaluated. We perform experiments comparing our knowledge graphs with existing manually constructed knowledge graphs of high quality, with respect to graph structure and edge labels. We find that the algorithms usually produces graphs with a structure similar to that of manually constructed knowledge graphs, as long as the data set is sufficiently large and general, and that the similarity of edge labels to manually chosen edge labels vary widely depending on input.
|
52 |
Natural Language Inference Transfer Learning in a Multi-Task Contract Dataset : In the Case of ContractNLI: a Document Information Extraction SystemTang, Yiu Kei January 2023 (has links)
This thesis investigates the enhancement of legal contract Natural Language Inference (NLI) classification through supervised fine-tuning on general domain NLI, in the case of ContractNLI and Span NLI BERT (Koreeda and Manning, 2021), a multi-task document information extraction dataset and framework. Annotated datasets of a specific professional domain are scarce due to the high time and labour cost required to create them. Since NLI is a simple yet effective task in inducing and evaluating natural language understanding (NLU) abilities in language models, there is potential in leveraging abundant general domain NLI datasets to aid information extraction and classification for legal contracts. This work evaluates the impact of transfer learning from Adversarial NLI (Nie et al.,2020) from the general domain to ContractNLI, via sequential and mixed batch fine-tuning. The study also extends its investigation to the effects of the model’s evidence identification component on NLI, by fine-tuning on the Contract Understanding Atticus Dataset (Hendrycks et al., 2021). The results highlight the benefits of fine-tuning with general domain NLI data, particularly for hypotheses in the target task with balanced entailment and contradiction training examples. In addition, the study demonstrates the reciprocal relationship between evidence identification and NLI classification, where improvements in the former enhance the accuracy of the latter. With NLI being more commonly applied to information extraction settings in specialised domains, this work sheds light on the potential impacts of existing general domain NLI resources in stepping up classification performance in specific domains.
|
53 |
Generating Conceptual Metaphoric ParaphrasesGotting, Olof January 2021 (has links)
Metaphoric Paraphrase generation is a relatively new and unexplored Natural Language Generation task. The aim of the task is to develop computational systems that paraphrase literal sentences into cogent metaphoric ones. Challenges in the field include representation of common sense knowledge and ensuring meaning retention when dealing with phrases that are dissimilar in their literal sense. This thesis will deal with the specific task of paraphrasing literal adjective phrases into metaphoric noun phrases, taking into consideration the preceding context of the adjective phrase. Two different systems were developed as part of this study. The systems are identical, apart from the fact that one is endowed with a knowledge representation based on Conceptual Metaphor Theory. The paraphrases generated by the systems, along with paraphrases written by a native speaker of English, were scored on the parameters of meaning retention and creativity by a crowd-sourced panel. Both systems were able to generate cogent metaphoric paraphrases, although fairly unreliably compared to the human. The system endowed with Conceptual Metaphor Theory knowledge got a lower average meaning retention score and a higher average creativity score than the system without Conceptual Metaphor Theory knowledge representation. In addition to that it was found that less similarity in sentence embeddings of literal sentences and metaphoric paraphrases of them correlates with a higher level of perceived meaning retention and a lower perceived creativity of the metaphoric paraphrase. It was also found that less difference in GPT-2 log probability between literal sentences and metaphoric paraphrases of them correlates with humans evaluating the paraphrases as less creative.
|
54 |
The Influence of M-BERT and Sizes on the Choice of Transfer Languages in ParsingZhang, Yifei January 2021 (has links)
In this thesis, we explore the impact of M-BERT and different transfer sizes on the choice of different transfer languages in dependency parsing. In order to investigate our research questions, we conduct a series of experiments on the treebanks in Universal Dependencies with UUParser. The main conclusions and contributions of this study are as follows: First, we train a variety of languages in several different scripts with M-BERT being added into the parsing framework, which is one of the most state-of-the-art deep learning models based on the Transformer architecture. In general, we get advancing results with M-BERT compared with the randomly initialized embedding in UUParser. Second, since it is a common way to choose a source language, which is 'close' to the target language in cross-lingual parsing, we try to explore what 'close' languages actually are, as there is not a definition for 'close'. In our study, we explore how strongly the parsing results are correlated with the different linguistic distances between the source and target languages. The relevant data is queried from URIEL Database. We find that the parsing performance is more dependent on inventory, syntactic and featural distance than on the geographic, genetic and phonological distance in zero-shot experiments. In the few-shot prediction, the parsing accuracy shows stronger correlation with inventory and syntactic distance than with others. Third, we vary the training sizes in few-shot experiments with M-BERT being added to see how the parsing results are influenced. We find that it is very obvious that few-shot experiments outperform zero-shot experiments. With the source sizes being cut, all parsing scores decrease. However, we do not see a linear drop of the results.
|
55 |
Text Normalization for Text-to-SpeechZhang, Zhaorui January 2023 (has links)
Text normalization plays a crucial role in text-to-speech systems by ensuring that the input text is in an appropriate format and consists of standardized words prior to grapheme-to-phoneme conversion for text-to-speech. The aim of this study was to assess the performance of five text normalization systems based on different methods. These text normalization systems were evaluated on the English Google text normalization dataset. The evaluation was based on the similarity between the ground truth and normalized outputs from each text normalization system. Since multiple ground truth issues occurred during the evaluation, the original similarity scores needed to be manually re-scored. The re-scoring was employed on a sample data semi-randomly extracted from the evaluation dataset. According to the results, the accuracy of these text normalization systems can be ranked as follows: the Duplex system, the Hybrid system, the VT system, the RS system, and the WFST system. For the two rule-based systems from ReadSpeaker, the VT system performed slightly better than the RS system, with a slight difference in the original similarity score. By analyzing the error patterns produced during the normalization process, the study provided valuable insights into the strengths and limitations of these systems. The findings of this study contribute to the refinement of internal rules, leading to improved accuracy and effectiveness of text normalization in text-to-speech applications.
|
56 |
Applying Hierarchical Classifiers to Japanese Text for Emotion ClassificationKurita, Masaki January 2024 (has links)
This thesis demonstrates that hierarchical classifiers that organize a classifica-tion task across multiple, sequential stages substantially improve emotion clas-sification accuracy compared to flat classifiers that classify text in a single step, effectively recognizing underrepresented emotions and classifying similar emo-tions within Japanese social media texts, using the WRIME dataset. The study also investigates different training data formats—emotion intensity lists and sin-gle primary emotion labels—and reveals that while classifiers trained with lists generally achieve better overall performance, single-label classifiers excel inidentifying underrepresented emotions and those expressed with high intensity.We also explore the impact of sentence length and emotion intensity on classifierperformance, finding that hierarchical classifiers are more effective with textsthat exhibit higher emotional intensities and shorter sentences. These findingsdemonstrate the potential of hierarchical approaches to address the complexities of emotion classification tasks, contributing valuable insights into the develop-ment of more effective natural language processing systems. Future research directions include enhancing the computational efficiency of hierarchical classi-fiers and comparing the hierarchical classifiers implemented with various deep learning approaches.
|
57 |
Optimizing Speech Recognition for Low-Resource Languages: Northern SothoPrzezdziak, Agnieszka January 2024 (has links)
In this thesis, the development of an automatic speech recognition (ASR) system for Northern Sotho, a low-resource language in South Africa, is investigated. Low-resource languages face challenges such as limited linguistic data and insufficient computational resources. In an attempt to alleviate these challenges, the multilingual Wav2Vec2-XLSR model is fine-tuned using Northern Sotho speech data with two main strategies to improve ASR performance: inclusion of background noise during training and semi-supervised learning with additional generated labels. An additional dataset compiled from news in Northern Sotho is used for evaluation of the models. The experiments demonstrate that moderate levels of background noise can enhance model robustness, though excessive noise degrades performance, particularly on clean data. Semi-supervised learning with generated labels proves beneficial, especially when working with smaller labelled datasets, though optimal results are always achieved with large, in-domain labelled datasets. The last finding is confirmed by the additional news dataset, which proved extremely challenging, with high error rates achieved by models trained on clean data and limited benefits of noise augmentation.
|
58 |
Word Alignment by Re-using Parallel PhrasesHolmqvist, Maria January 2008 (has links)
In this thesis we present the idea of using parallel phrases for word alignment. Each parallel phrase is extracted from a set of manual word alignments and contains a number of source and target words and their corresponding alignments. If a parallel phrase matches a new sentence pair, its word alignments can be applied to the new sentence. There are several advantages of using phrases for word alignment. First, longer text segments include more context and will be more likely to produce correct word alignments than shorter segments or single words. More importantly, the use of longer phrases makesit possible to generalize words in the phrase by replacing words by parts-of-speech or other grammatical information. In this way, the number of words covered by the extracted phrases can go beyond the words and phrases that were present in the original set of manually aligned sentences. We present experiments with phrase-based word alignment on three types of English–Swedish parallel corpora: a software manual, a novel and proceedings of the European Parliament. In order to find a balance between improved coverage and high alignment accuracy we investigated different properties of generalised phrases to identify which types of phrases are likely to produce accurate alignments on new data. Finally, we have compared phrase-based word alignments to state-of-the-art statistical alignment with encouraging results. We show that phrase-based word alignments can be used to enhance statistical word alignment. To evaluate word alignments an English–Swedish reference set for the Europarl corpus was constructed. The guidelines for producing this reference alignment are presented in the thesis.
|
59 |
Intercepting OpenGL calls for rendering on 3D displayde Vahl, Joel January 2005 (has links)
<p>An OpenGL applications usually renders to a single frame. Multi-view or 3D displays on the other hand, needs more more images representing different viewing directions on the same scene, but modifying a large number of applications would be unsuitable and problematic. However, intercepting and modifying these calls before they reach the GPU would dramatically decrease the amount of work needed to support a large number of applications on a new type of multi-view or 3D display. This thesis describes different ways on intercepting, enqueueing and replaying these calls to support rendering form different view points. Intercepting with both an own implementation of opengl32.dll and an OpenGL driver is discussed, and enqueueing using classes, function pointers and enumeration of functions is tried. The different techniques are discussed quickly with the focus being a working implementation. This resulting in an fully blown OpenGL interceptor with the ability to enqueue and replay a frame multiple times while modifying parameters such as the projection matrix. This implementation uses an own implementation of opengl32.dll that is placed in the application directory to be loaded before the real one. Enqueueing is performed by enumerating all OpenGL calls, pushing this enumeration value and all call data to a list Replaying is done by reading the same list and calling the function pointer associated with the enumeration value with the data in the list.</p>
|
60 |
Video coding using compressed transportation plans / Videokodning med komprimerade transportplanerLissing, Johan January 2007 (has links)
<p>A transportation plan is a byproduct from the calculation of the Kantorovich distance between two images. It describes a transformation from one of the images to the other. This master thesis shows how transportation plans can be used for video coding and how to process the transportation plans to achieve a good bitrate/quality ratio. Various parameters are evaluated using an implemented transportation plan video coder.</p><p>The introduction of transform coding with DCT proves to be very useful, as it reduces the size of the resulting transportation plans. DCT coding roughly gives a 10-fold decrease in bitrate with maintained quality compared to the nontransformed transportation plan coding.</p><p>With the best settings for transportation plan coding, I was able to code a test sequence at about 5 times the bitrate for MPEG coding of the same sequence with similar quality.</p><p>As video coding using transportation plans is a very new concept, the thesis is ended with conclusions on the test results and suggestions for future research in this area.</p>
|
Page generated in 0.0941 seconds