• Refine Query
  • Source
  • Publication year
  • to
  • Language
  • 1010
  • 191
  • 86
  • 71
  • 34
  • 19
  • 18
  • 13
  • 11
  • 10
  • 8
  • 7
  • 5
  • 5
  • 4
  • Tagged with
  • 1805
  • 1805
  • 1571
  • 667
  • 584
  • 474
  • 412
  • 391
  • 277
  • 269
  • 262
  • 237
  • 230
  • 219
  • 206
  • About
  • The Global ETD Search service is a free service for researchers to find electronic theses and dissertations. This service is provided by the Networked Digital Library of Theses and Dissertations.
    Our metadata is collected from universities around the world. If you manage a university/consortium/country archive and want to be added, details can be found on the NDLTD website.
41

A shallow parser based on closed-class words to capture relations in biomedical text

Leroy, Gondy, Chen, Hsinchun, Martinez, Jesse D. 06 1900 (has links)
Artificial Intelligence Lab, Department of MIS, University of Arizona / Natural language processing for biomedical text currently focuses mostly on entity and relation extraction. These entities and relations are usually pre-specified entities, e.g., proteins, and pre-specified relations, e.g., inhibit relations. A shallow parser that captures the relations between noun phrases automatically from free text has been developed and evaluated. It uses heuristics and a noun phraser to capture entities of interest in the text. Cascaded finite state automata structure the relations between individual entities. The automata are based on closed-class English words and model generic relations not limited to specific words. The parser also recognizes coordinating conjunctions and captures negation in text, a feature usually ignored by others. Three cancer researchers evaluated 330 relations extracted from 26 abstracts of interest to them. There were 296 relations correctly extracted from the abstracts resulting in 90% precision of the relations and an average of 11 correct relations per abstract.
42

A Natural Language Search Interface for Accommodation Queries

Chavanne, Erin 01 January 2015 (has links)
Services that once required human interaction are now completed with the click of a few buttons. In general, this allows for a more streamlined process for activities such as sending messages (email or text messages), filing taxes, or even shopping for groceries. In terms of searching for hotels and travel accommodations however, this process has not proven to be the most effective as the speed and efficiency is hindered by the interface through which this information is available. Choosing a travel specific site, filling in the required fields, combing through results for the desired specifications, and then possibly repeating the process elsewhere, does not provide the ability for the user to express the entirety of their preferences for the accommodation and is therefore not an effective method for searching. Natural language search provides a more accessible and intuitive interface for accommodation searching. Instead of specifying fields that may not encompass the the entirety of the desired search, the user is able to express all of the aspects in a single, natural language, search. In this project, we propose a natural language search interface for accommodations such as hotels, hostels, or apartments. Data acquired through Amazon Mechanical Turk is used to create a system for extracting various accommodation fields. Zilyo and Expedia APIs are then queried for real-time accommodation listings. These results are then adjusted based on the specifics of the search that were not included in the original query. A natural language search of this kind is not only more accessible on the consumer end, but provides data that pertains directly to the the entirety of the intended search.
43

Analysing coherence of intention in natural language dialogue

Mc Kevitt, Paul January 1991 (has links)
No description available.
44

Simplifying natural language for aphasic readers

Devlin, Siobhan Lucy January 1999 (has links)
No description available.
45

The role of document structure in text generation

Bouayad-Agha, Nadjet January 2001 (has links)
No description available.
46

Knowledge intensive natural language generation with revision /

Cline, Ben E. January 1994 (has links)
Thesis (Ph. D.)--Virginia Polytechnic Institute and State University, 1994. / Vita. Abstract. Includes bibliographical references (leaves 143-146). Also available via the Internet.
47

A caption-based natural-language interface handling descriptive captions for a multimedia database system /

Dulle, John David. January 1990 (has links)
Thesis (M.S. in Computer Science)--Naval Postgraduate School, June 1990. / Thesis Advisor(s): Lum, Vincent Y. ; Rowe, Neil C. "June 1990." Description based on signature page. DTIC Identifiers: Interfaces, natural language, databases, theses. Author(s) subject terms: Natural language processing, multimedia database system, natural language interface, descriptive captions. Includes bibliographical references (p. 27).
48

Word sense selection in texts an integrated model /

Kwong, Oi Yee. January 1900 (has links)
Thesis (Ph. D.)--University of Cambridge, 2000. / Cover title. "September 2000." Includes bibliographical references.
49

Predicting Depression and Suicide Ideation in the Canadian Population Using Social Media Data

Skaik, Ruba 30 June 2021 (has links)
The economic burden of mental illness costs Canada billions of dollars every year. Millions of people suffer from mental illness, and only a fraction receives adequate treatment. Identifying people with mental illness requires initiation from those in need, available medical services, and professional experts’ time allocation. These resources might not be available all the time. The common practice is to rely on clinical data, which is generally collected after the illness is developed and reported. Moreover, such clinical data is incomplete and hard to obtain. An alternative data source is conducting surveys through phone calls, interviews, or mail, but this is costly and time-consuming. Social media analysis has brought advances in leveraging population data to understand mental health problems. Thus, analyzing social media posts can be an essential alternative for identifying mental disorders throughout the Canadian population. Big data research of social media may also endorse standard surveillance approaches and provide decision-makers with usable information. More precisely, social media analysis has shown promising results for public health assessment and monitoring. In this research, we explore the task of automatically analysing social media textual data using Natural Language Processing (NLP) and Machine Learning (ML) techniques to detect signs of mental health disorders that need attention, such as depression and suicide ideation. Considering the lack of comprehensive annotated data in this field, we propose a methodology for transfer learning to utilize the information hidden in a training sample and leverage it on a different dataset to choose the best-generalized model to be applied at the population level. We also present evidence that ML models designed to predict suicide ideation using Reddit data can utilize the knowledge they encoded to make predictions on Twitter data, even though the two platforms differ in the purpose, structure, and limitations. In our proposed models, we use feature engineering with supervised machine learning algorithms (such as SVM, LR, RF, XGBoost, and GBDT), and we compare their results with those of deep learning algorithms (such as LSTM, Bi-LSTM, and CNNs). We adopt the CNN model for depression classification that obtained the highest F1-score on the test dataset (0.898) and 0.941 recall. This model is later used to estimate the depression level of the population. For suicide ideation detection, we used the CNN model with pre-trained fastText word embeddings and linguistic features (LIWC). The model achieved an F1-score of 0.936 and a recall of 0.88 to predict suicide ideation at the user-level on the test set. To compare our models’ predictions with official statics, we used 2015-2016 population based Canadian Community Health Survey (CCHS) on Mental Health and Well-being conducted by Statistics Canada. The data is used to estimate depression and suicidality in Canadian provinces and territories. For depression, (n=53,050) respondents filled in the Patient Health Questionnaire-9 (PHQ-9) from 8 provinces/territories. Each survey respondent with a score ≥ 10 on the PHQ-9 was interpreted as having moderate to severe depression because this score is frequently used as a screening cut-point. The weighted percentage of depression prevalence during 2015 for females and males of the age between 15 to 75 was 11.5% and 8.1%, respectively (with 54.2% females and 45.8% males). Our model was applied on a population-representative dataset that contains 24,251 Twitter users who posted 1,735,200 tweets during 2015 with a Pearson correlation of 0.88 for both sex and age within the seven provinces and NT territory included in the CCHS. An age correlation of 0.95 was calculated for age and sex (separately) and our model estimated that 10% of the sample dataset has evidence of depression (58.3% females and 41.7% males). For the second task, suicide ideation, Statistics Canada (2015) estimated the total number of people who reported serious suicidal thoughts as 3,396,700 persons, i.e., 9.514% of the total population, whereas our models estimated 10.6% of the population sample were at risk of suicide ideation (59% females and 41% males). The Pearson correlation coefficients between the actual suicide ideation within the last 12 months and the predicted model for each province per age, sex, and both more than 0.62, which indicates a reasonable correlation.
50

Grammars for generating isiXhosa and isiZulu weather bulletin verbs

Mahlaza, Zola January 2018 (has links)
The Met Office has investigated the use of natural language generation (NLG) technologies to streamline the production of weather forecasts. Their approach would be of great benefit in South Africa because there is no fast and large scale producer, automated or otherwise, of textual weather summaries for Nguni languages. This is because of, among other things, the complexity of Nguni languages. The structure of these languages is very different from Indo-European languages, and therefore we cannot reuse existing technologies that were developed for the latter group. Traditional NLG techniques such as templates are not compatible with 'Bantu' languages, and existing works that document scaled-down 'Bantu' language grammars are also not sufficient to generate weather text. In pursuance of generating weather text in isiXhosa and isiZulu - we restricted our text to only verbs in order to ensure a manageable scope. In particular, we have developed a corpus of weather sentences in order to determine verb features. We then created context free verbal grammar rules using an incremental approach. The quality of these rules was evaluated using two linguists. We then investigated the grammatical similarity of isiZulu verbs with their isiXhosa counterparts, and the extent to which a singular merged set of grammar rules can be used to produce correct verbs for both languages. The similarity analysis of the two languages was done through the developed rules' parse trees, and by applying binary similarity measures on the sets of verbs generated by the rules. The parse trees show that the differences between the verb's components are minor, and the similarity measures indicate that the verb sets are at most 59.5% similar (Driver-Kroeber metric). We also examined the importance of the phonological conditioning process by developing functions that calculate the ratio of verbs that will require conditioning out of the total strings that can be generated. We have found that the phonological conditioning process affects at least 45% of strings for isiXhosa, and at least 67% of strings for isiZulu depending on the type of verb root that is used. Overall, this work shows that the differences between isiXhosa and isiZulu verbs are minor, however, the exploitation of these similarities for the goal of creating a unified rule set for both languages cannot be achieved without significant maintainability compromises because there are dependencies that exist in one language and not the other between the verb's 'modules'. Furthermore, the phonological conditioning process should be implemented in order to improve generated text due to the high ratio of verbs it affects.

Page generated in 0.0643 seconds