• Refine Query
  • Source
  • Publication year
  • to
  • Language
  • 244
  • 124
  • 44
  • 38
  • 31
  • 29
  • 24
  • 24
  • 13
  • 7
  • 6
  • 6
  • 5
  • 5
  • 5
  • Tagged with
  • 626
  • 626
  • 144
  • 130
  • 119
  • 114
  • 92
  • 88
  • 86
  • 81
  • 80
  • 76
  • 69
  • 65
  • 65
  • About
  • The Global ETD Search service is a free service for researchers to find electronic theses and dissertations. This service is provided by the Networked Digital Library of Theses and Dissertations.
    Our metadata is collected from universities around the world. If you manage a university/consortium/country archive and want to be added, details can be found on the NDLTD website.
141

Contribution à la méthode de conception inventive par l'extraction automatique de connaissances des textes de brevets d'invention / Toward an automatic extraction of inventive design method knowledge from patents

Souili, Wendemmi Moukassa Achille 31 August 2015 (has links)
Les brevets d’invention titres de propriété industrielle confèrent à leurs titulaires le monopole de l’invention brevetée. On peut y trouver une sorte d’historique de l’évolution de l’artefact. Dans ce contexte le concepteur est très souvent amené à faire des recherches dans les documents de brevets afin de bénéficier des connaissances qui y sont contenues en vue de structurer le processus inventif. Développée pour assister les concepteurs dans leur démarche d’innovation, la Méthode de Conception Inventive (MCI), s’inscrit dans le modèle de la dialectique. La MCI a précisé les concepts entrant en jeu dans la description des évolutions des systèmes techniques et des artefacts. Ces items intéressent bien souvent les concepteurs et sont essentiels à la compréhension du problème sous-jacent et à la collecte de toutes les caractéristiques sur lesquelles on peut agir ; et de l’effet de leurs variations sur l’artefact. Cette thèse consiste d’abord à analyser le document de brevet d’un point de vue linguistique, afin d’en connaitre la typologie. Il s’agit, ensuite, de repérer dans le document de brevets les connaissances susceptibles d’être utiles à la MCI et à les formaliser sous forme de programme informatique. L’approche que nous proposons est issue du text-mining. Elle est à base de marqueurs linguistiques et utilise des patrons lexico-syntaxiques issus du domaine du traitement automatique des langues. Cette méthode d’extraction des concepts utiles à la MCI permet l’établissement d’une sorte de cartographie initiale des évolutions passées et possibles des caractéristiques de l’artefact. L’intérêt est en outre de faciliter grandement l’analyse préliminaire des connaissances relative au dit artefact. / Patents are industrial property titles that give their holders a monopoly over the patented invention. It is possible to find a sort of history of the evolution of the artifact. In this context the designer often like to do research in patent documents in order to benefit from the knowledge contained inside to structure the inventive process. Developed to assist designers in their innovation approach, the Inventive Design Method (IDM) is part of the pattern of dialectic. IDM has clarified the concepts at stake in the description of the evolution of technical systems and artifact. These items often interest designers and are essential to understanding the underlying problem and collecting of all features on which to act; and the effect of variations on the artifact. This thesis, firstly, deals with patent document analysis from a linguistic point of view, in order to know its typology. Then, it is possible to identify in the patent document, the knowledge likely to be useful to IDM and formalize it as a computer program. The approach proposed in this paper is based on text mining techniques. It uses a method based on linguistic markers using lexical and syntactic patterns from the field of natural language processing. This method of extraction of useful concepts for IDM allows the establishment of a kind of initial mapping of past and possible changes in the future of the artifact characteristics. The interest is also to greatly facilitate the preliminary analysis of knowledge on the said artifact.
142

CSR i VD:n har ordet – En kvantitativ innehållsanalys på OMXS30 / CSR in the CEO-letters - A quantitative content analysis on OMXS30

Nilsson, Henrik, Palmgren, Marcus, Ngorsungnoen, Martin January 2017 (has links)
Årsredovisningar är ett ofta förekommande objekt inom redovisningsforskningen. Årsredovisningarna är en viktig kommunikationskanal mellan företaget och dess olika intressenter. Rapportens ändamål är primärt att ge intressenterna information om företagets resultat och finansiella ställning. Årsredovisningarnas karaktär har genom åren utvecklats och berikats med bland annat bilder och kompletterande textuella avsnitt. I denna studie har den narrativa delen av årsredovisningar ’VD:n har ordet’ studerats. Avsnittet utgör en viktig del av årsredovisningens frivilliga element, och är en av de mest lästa delarna. Samtidigt är ’VD:n har ordet’, till skillnad från flertalet andra delar i årsredovisningen, oreglerad. Detta ger företaget och den verkställande direktören stora möjligheter att kommunicera ut legitimitetskapande budskap till dess intressenter. Tidigare forskning visar att aktiviteter kopplade till CSR har en positiv påverkan på intressenternas inställning gentemot företaget. CSR är ett rymligt begrepp som innefattar relationen och ansvarstagandet mellan företag och samhälle. Det har visat sig ha stor betydelse för företagen att lyckas signalera dessa aktiviteter till intressenterna. Denna studie syftar till att ge en bättre inblick i hur svenska företag använder ’VD:n har ordet’ för att lyfta fram CSR. Vidare undersöks om det har skett en ökning av CSR-begrepp i ’VD:n har ordet’. Studien uppvisar att det inte finns något samband mellan företagens vinstmarginal och andel CSR-begrepp i ’VD:n har ordet’. Det går inte heller att urskilja ett samband mellan andel CSR-begrepp och andelen begrepp kopplade till resultat. Däremot kunde vi bevisa att det skett en linjär ökning av CSR-begrepp mellan år 2006 och 2015. Begrepp kopplade till resultat har minskat, om än inte linjärt. / Annual reports are a frequently used item in the accounting research. The annual reports are an important communication channel between the company and its various stakeholders. The purpose of the report is to give the stakeholders information about the company's earnings and financial position. The character of the annual reports has evolved over the years and has been enriched with, among other things, pictures and additional textual sections. In this study, the narrative part of the annual reports ‘CEO-letter’ has been studied. The section constitutes an important part of the annual report's optional elements, and is one of the most read sections. At the same time, the ‘CEO-letter’, unlike most other part of the annual report, is unregulated. This fact gives the company and the CEO a great opportunity to communicate legitimacy-creating messages to its stakeholders. Previous research shows that activities related to CSR have a positive impact on stakeholders' attitude towards the company. CSR is a spacious concept that includes the relationship and responsivities between the company and its society. It has been found to be of great importance for companies to successfully signal these activities to stakeholders. This study aims at giving a better insight into how Swedish companies use the 'CEO-letter' to highlight CSR. Further investigations are made if there has been an increase in the use of words related to CSR in the 'CEO-letters’. This study shows that there is no correlation between the company's profit margin and the share of CSR concepts in the ‘CEO-letter’. In addition, we could not discern a relationship between the share of CSR concepts and the proportion of concepts linked to results in the ‘CEO-letters’. On the other hand, we could prove that there has been a linear increase in CSR concepts over the past ten years. Furthermore, concepts linked to results have decreased, albeit not in a linear manner. Please note that the thesis language is in Swedish.
143

Finding conflicting statements in the biomedical literature

Sarafraz, Farzaneh January 2012 (has links)
The main archive of life sciences literature currently contains more than 18,000,000 references, and it is virtually impossible for any human to stay up-to-date with this large number of papers, even in a specific sub-domain. Not every fact that is reported in the literature is novel and distinct. Scientists report repeat experiments, or refer to previous findings. Given the large number of publications, it is not surprising that information on certain topics is repeated over a number of publications. From consensus to contradiction, there are all shades of agreement between the claimed facts in the literature, and considering the volume of the corpus, conflicting findings are not unlikely. Finding such claims is particularly interesting for scientists, as they can present opportunities for knowledge consolidation and future investigations. In this thesis we present a method to extract and contextualise statements about molecular events as expressed in the biomedical literature, and to find those that potentially conflict each other. The approach uses a system that detects event negations and speculation, and combines those with contextual features (e.g. type of event, species, and anatomical location) to build a representational model for establishing relations between different biological events, including relations concerning conflicts. In the detection of negations and speculations, rich lexical, syntactic, and semantic features have been exploited, including the syntactic command relation. Different parts of the proposed method have been evaluated in a context of the BioNLP 09 challenge. The average F-measures for event negation and speculation detection were 63% (with precision of 88%) and 48% (with precision of 64%) respectively. An analysis of a set of 50 extracted event pairs identified as potentially conflicting revealed that 32 of them showed some degree of conflict (64%); 10 event pairs (20%) needed a more complex biological interpretation to decide whether there was a conflict. We also provide an open source integrated text mining framework for extracting events and their context on a large-scale basis using a pipeline of tools that are available or have been developed as part of this research, along with 72,314 potentially conflicting molecular event pairs that have been generated by mining the entire body of accessible biomedical literature. We conclude that, whilst automated conflict mining would need more comprehensive context extraction, it is feasible to provide a support environment for biologists to browse potential conflicting statements and facilitate data and knowledge consolidation.
144

Discovering relations between indirectly connected biomedical concepts: Research Article

Tsatsaronis, George, Weissenborn, Dirk, Schroeder, Michael 04 January 2016 (has links)
BACKGROUND: The complexity and scale of the knowledge in the biomedical domain has motivated research work towards mining heterogeneous data from both structured and unstructured knowledge bases. Towards this direction, it is necessary to combine facts in order to formulate hypotheses or draw conclusions about the domain concepts. This work addresses this problem by using indirect knowledge connecting two concepts in a knowledge graph to discover hidden relations between them. The graph represents concepts as vertices and relations as edges, stemming from structured (ontologies) and unstructured (textual) data. In this graph, path patterns, i.e. sequences of relations, are mined using distant supervision that potentially characterize a biomedical relation. RESULTS: It is possible to identify characteristic path patterns of biomedical relations from this representation using machine learning. For experimental evaluation two frequent biomedical relations, namely \'has target\', and \'may treat\', are chosen. Results suggest that relation discovery using indirect knowledge is possible, with an AUC that can reach up to 0.8, a result which is a great improvement compared to the random classification, and which shows that good predictions can be prioritized by following the suggested approach. CONCLUSIONS: Analysis of the results indicates that the models can successfully learn expressive path patterns for the examined relations. Furthermore, this work demonstrates that the constructed graph allows for the easy integration of heterogeneous information and discovery of indirect connections between biomedical concepts.
145

Release of the MySQL based implementation of the CTS protocol

Tiepmar, Jochen January 2016 (has links)
In a project called "A Library of a Billion Words" we needed an implementation of the CTS protocol that is capable of handling a text collection containing at least 1 billion words. Because the existing solutions did not work for this scale or were still in development I started an implementation of the CTS protocol using methods that MySQL provides. Last year we published a paper that introduced a prototype with the core functionalities without being compliant with the specifications of CTS (Tiepmar et al., 2013). The purpose of this paper is to describe and evaluate the MySQL based implementa-tion now that it is fulfilling the specifications version 5.0 rc.1 and mark it as finished and ready to use. Fur-ther information, online instances of CTS for all de-scribed datasets and binaries can be accessed via the projects website1. Reference Tiepmar J, Teichmann C, Heyer G, Berti M and Crane G. 2013. A new Implementation for Canonical Text Services. in Proceedings of the 8th Workshop on Language Technology for Cultural Heritage, Social Sciences, and Humanities (LaTeCH).
146

Status Quo der Textanalyse im Rahmen der Business Intelligence

Schieber, Andreas, Hilbert, Andreas January 2014 (has links)
Vor dem Hintergrund der Zunahme unstrukturierter Daten für Unternehmen befasst sich dieser Beitrag mit den Möglichkeiten, die durch den Einsatz der Business Intelligence für Unternehmen bestehen, wenn durch gezielte Analyse die Bedeutung dieser Daten erfasst, gefiltert und ausgewertet werden können. Allgemein ist das Ziel der Business Intelligence die Unterstützung von Entscheidungen, die im Unternehmen (auf Basis strukturierter Daten) getroffen werden. Die zusätzliche Auswertung von unstrukturierten Daten, d.h. unternehmensinternen Dokumenten oder Texten aus dem Web 2.0, führt zu einer Vergrößerung des Potenzials und dient der Erweiterung des Geschäftsverständnisses der Verbesserung der Entscheidungsfindung. Der Beitrag erläutert dabei nicht nur Konzepte und Verfahren, die diese Analysen ermöglichen, sondern zeigt auch Fallbeispiele zur Demonstration ihrer Nützlichkeit.:1 Einführung 2 Business Intelligence 2.1 Definition 2.2 Ordnungsrahmen 2.3 Analyseorientierte BI und Data Mining 3 Text Mining 3.1 Berührungspunkte mit anderen Disziplinen 3.2 Definition 3.3 Prozessmodell nach HIPPNER & RENTZMANN (2006a) 3.3.1 Aufgabendefinition 3.3.2 Dokumentselektion 3.3.3 Dokumentaufbereitung 3.3.4 Text-Mining-Methoden 3.3.5 Interpretation / Evaluation 3.3.6 Anwendung 4 Potenziale der Textanalyse 4.1 Erweiterung des CRM 4.2 Alternative zur Marktforschung 5 Fazit und Ausblick Literaturverzeichnis
147

Entwicklung eines generischen Vorgehensmodells für Text Mining

Schieber, Andreas, Hilbert, Andreas 29 April 2014 (has links)
Vor dem Hintergrund des steigenden Interesses von computergestützter Textanalyse in Forschung und Praxis entwickelt dieser Beitrag auf Basis aktueller Literatur ein generisches Vorgehensmodell für Text-Mining-Prozesse. Das Ziel des Beitrags ist, die dabei anfallenden, umfangreichen Aktivitäten zu strukturieren und dadurch die Komplexität von Text-Mining-Vorhaben zu reduzieren. Das Forschungsziel stützt sich auf die Tatsache, dass im Rahmen einer im Vorfeld durchgeführten, systematischen Literatur-Review keine detaillierten, anwendungsneutralen Vorgehensmodelle für Text Mining identifiziert werden konnten. Aufbauend auf den Erkenntnissen der Literatur-Review enthält das resultierende Modell daher sowohl induktiv begründete Komponenten aus spezifischen Ansätzen als auch aus literaturbasierten Anforderungen deduktiv abgeleitete Bestandteile. Die Evaluation des Artefakts belegt die Nützlichkeit des Vorgehensmodells im Vergleich mit dem bisherigen Forschungsstand.:1 Einführung 1.1 Motivation 1.2 Forschungsziel und Methodik 1.2.1 Systematische Literatur-Review 1.2.2 Design-Science-Research-Ansatz 1.3 Aufbau des Beitrags 2 Stand der Forschung 2.1 Begriffsverständnis 2.2 Merkmale von Vorgehensmodellen für Text Mining 2.3 Aktivitäten im Text-Mining-Prozess 2.4 Zusammenfassung 3 Anforderungen an ein generisches Vorgehensmodell 3.1 Strukturelle Anforderungen 3.2 Funktionelle Anforderungen 3.3 Zusammenfassung 4 Entwicklung des Modells 4.1 Aufgabendefinition 4.2 Dokumentenselektion und -untersuchung 4.3 Dokumentenaufbereitung 4.3.1 Linguistische Aufbereitung 4.3.2 Technische Aufbereitung 4.4 Text-Mining-Verfahren 4.5 Ergebnisevaluation 4.6 Anwendung 4.7 Zusammenfassung 4.7.1 Gesamtmodell 4.7.2 Feedbackschleifen 5 Evaluation 5.1 Evaluationsdesign 5.2 Messung und Auswertung 6 Fazit und Ausblick Literaturverzeichnis Anhang A1 Anwendungsneutrale Vorgehensmodelle A2 Auswirkungen von Grund- und Stammformenreduktion auf die Interpretierbarkeit von Texten A3 Gesamtmodell
148

Unsupervised Natural Language Processing for Knowledge Extraction from Domain-specific Textual Resources

Hänig, Christian 17 April 2013 (has links)
This thesis aims to develop a Relation Extraction algorithm to extract knowledge out of automotive data. While most approaches to Relation Extraction are only evaluated on newspaper data dealing with general relations from the business world their applicability to other data sets is not well studied. Part I of this thesis deals with theoretical foundations of Information Extraction algorithms. Text mining cannot be seen as the simple application of data mining methods to textual data. Instead, sophisticated methods have to be employed to accurately extract knowledge from text which then can be mined using statistical methods from the field of data mining. Information Extraction itself can be divided into two subtasks: Entity Detection and Relation Extraction. The detection of entities is very domain-dependent due to terminology, abbreviations and general language use within the given domain. Thus, this task has to be solved for each domain employing thesauri or another type of lexicon. Supervised approaches to Named Entity Recognition will not achieve reasonable results unless they have been trained for the given type of data. The task of Relation Extraction can be basically approached by pattern-based and kernel-based algorithms. The latter achieve state-of-the-art results on newspaper data and point out the importance of linguistic features. In order to analyze relations contained in textual data, syntactic features like part-of-speech tags and syntactic parses are essential. Chapter 4 presents machine learning approaches and linguistic foundations being essential for syntactic annotation of textual data and Relation Extraction. Chapter 6 analyzes the performance of state-of-the-art algorithms of POS tagging, syntactic parsing and Relation Extraction on automotive data. The findings are: supervised methods trained on newspaper corpora do not achieve accurate results when being applied on automotive data. This is grounded in various reasons. Besides low-quality text, the nature of automotive relations states the main challenge. Automotive relation types of interest (e. g. component – symptom) are rather arbitrary compared to well-studied relation types like is-a or is-head-of. In order to achieve acceptable results, algorithms have to be trained directly on this kind of data. As the manual annotation of data for each language and data type is too costly and inflexible, unsupervised methods are the ones to rely on. Part II deals with the development of dedicated algorithms for all three essential tasks. Unsupervised POS tagging (Chapter 7) is a well-studied task and algorithms achieving accurate tagging exist. All of them do not disambiguate high frequency words, only out-of-lexicon words are disambiguated. Most high frequency words bear syntactic information and thus, it is very important to differentiate between their different functions. Especially domain languages contain ambiguous and high frequent words bearing semantic information (e. g. pump). In order to improve POS tagging, an algorithm for disambiguation is developed and used to enhance an existing state-of-the-art tagger. This approach is based on context clustering which is used to detect a word type’s different syntactic functions. Evaluation shows that tagging accuracy is raised significantly. An approach to unsupervised syntactic parsing (Chapter 8) is developed in order to suffice the requirements of Relation Extraction. These requirements include high precision results on nominal and prepositional phrases as they contain the entities being relevant for Relation Extraction. Furthermore, accurate shallow parsing is more desirable than deep binary parsing as it facilitates Relation Extraction more than deep parsing. Endocentric and exocentric constructions can be distinguished and improve proper phrase labeling. unsuParse is based on preferred positions of word types within phrases to detect phrase candidates. Iterating the detection of simple phrases successively induces deeper structures. The proposed algorithm fulfills all demanded criteria and achieves competitive results on standard evaluation setups. Syntactic Relation Extraction (Chapter 9) is an approach exploiting syntactic statistics and text characteristics to extract relations between previously annotated entities. The approach is based on entity distributions given in a corpus and thus, provides a possibility to extend text mining processes to new data in an unsupervised manner. Evaluation on two different languages and two different text types of the automotive domain shows that it achieves accurate results on repair order data. Results are less accurate on internet data, but the task of sentiment analysis and extraction of the opinion target can be mastered. Thus, the incorporation of internet data is possible and important as it provides useful insight into the customer\''s thoughts. To conclude, this thesis presents a complete unsupervised workflow for Relation Extraction – except for the highly domain-dependent Entity Detection task – improving performance of each of the involved subtasks compared to state-of-the-art approaches. Furthermore, this work applies Natural Language Processing methods and Relation Extraction approaches to real world data unveiling challenges that do not occur in high quality newspaper corpora.
149

Du dossier résident informatisé à la recherche en santé publique : Application des méthodes de surveillance en temps réel à des données médico-sociales de la personne âgée et exploration de données de cohorte pour la santé publique. / From a nursing home electronic resident data warehouse to public health research : Applying public health surveillance systems methods to a real time long term care database and building a resident cohort study.

Delespierre, Tiba 19 June 2018 (has links)
La France connaît un vieillissement de sa population sans précédent. La part des séniors s’accroît et notre société se doit de repenser son organisation pour tenir compte de ce changement et mieux connaître cette population.De nombreuses cohortes de personnes âgées existent déjà à travers le monde dont quatre en France et, bien que la part de cette population vivant dans des structures d’hébergement collectif (EHPAD, cliniques de soins de suite) augmente, la connaissance de ces seniors reste lacunaire.Aujourd’hui les groupes privés de maisons de retraite et d’établissements sanitaires comme Korian ou Orpéa s’équipent de grandes bases de données relationnelles permettant d’avoir de l’information en temps réel sur leurs patients/résidents. Depuis 2010 les dossiers de tous les résidents Korian sont dématérialisés et accessibles par requêtes. Ils comprennent à la fois des données médico-sociales structurées décrivant les résidents et leurs traitements et pathologies, mais aussi des données textuelles explicitant leur prise en charge au quotidien et saisies par le personnel soignant.Au fil du temps et alors que le dossier résident informatisé (DRI) avait surtout été conçu comme une application de gestion de base de données, il est apparu comme une nécessité d’exploiter cette mine d’informations et de construire un outil d’aide à la décision destiné à améliorer l’efficacité des soins. L’Institut du Bien Vieillir IBV devenu entretemps la Fondation Korian pour le Bien Vieillir a alors choisi, dans le cadre d’un partenariat Public/Privé de financer un travail de recherche destiné à mieux comprendre le potentiel informatif de ces données, d’évaluer leur fiabilité et leur capacité à apporter des réponses en santé publique. Ce travail de recherche et plus particulièrement cette thèse a alors été pensée en plusieurs étapes.- D’abord l’analyse de contenu du data warehouse DRI, l’objectif étant de construire une base de données recherche, avec un versant social et un autre de santé. Ce fut le sujet du premier article.- Ensuite, par extraction directe des informations socio-démographiques des résidents dès leur entrée, de leurs hospitalisations et décès puis, par un processus itératif d’extractions d’informations textuelles de la table des transmissions et l’utilisation de la méthode Delphi, nous avons généré vingt-quatre syndromes, ajouté les hospitalisations et les décès et construit une base de données syndromique, la Base du Bien Vieillir (BBV) . Ce système d’informations d’un nouveau type a permis la constitution d’une cohorte de santé publique à partir de la population des résidents de la BBV et l’organisation d’un suivi longitudinal syndromique de celle-ci. La BBV a également été évaluée scientifiquement dans un cadre de surveillance et de recherche en santé publique au travers d’une analyse de l’existant : contenu, périodicité, qualité des données. La cohorte construite a ainsi permis la constitution d’un outil de surveillance. Cet échantillon de population a été suivi en temps réel au moyen des fréquences quotidiennes d’apparitions des 26 syndromes des résidents. La méthodologie d’évaluation était celle des systèmes de surveillance sanitaire proposée par le CDC d’Atlanta et a été utilisée pour les syndromes grippaux et les gastro entérites aiguës. Ce fut l’objet du second article.- Enfin la construction d’un nouvel outil de santé publique : la distribution de chacun des syndromes dans le temps (dates de transmissions) et l’espace (les EHPAD de transmissions) a ouvert le champ de la recherche à de nouvelles méthodes d’exploration des données et permis d’étudier plusieurs problématiques liées à la personne âgée : chutes répétées, cancer, vaccinations et fin de vie. / French population is rapidly aging. Senior citizens ratio is increasing and our society needs to rethink its organization, taking into account this change, better knowing this fast growing population group.Even if numerous cohorts of elderly people already exist worldly with four in France and, even as they live in growing numbers in nursing homes and out-patient treatment clinics, knowledge of this population segment is still missing.Today several health and medico-social structures groups as Korian and Orpéa invest in big relational data bases enabling them to get real-time information about their patients/residents. Since 2010 all Korian residents’ files are dematerialized and accessible by requests. They contain at the same time, structured medico-social data describing the residents as well as their treatments and pathologies, but also free-textual data detailing their daily care by the medical staff.Through time and as the computerized resident file (DRI) was mainly conceived as a data base management application, it appeared essential to mine these data and build a decision-making tool intended to improve the care efficiency. The Ageing Well Institute becoming meanwhile the Korian Ageing Well Foundation chose then, working in a private/public partnership, to finance a research work intented to better understand these datas’ informative potential, to assess their reliability and response to public health threats. This research work and this thesis were then designed in several steps:- First, a content analysis of the data warehouse DRI, the objective being to build a research data base, with a social side and a health side. This was the first paper subject.- Then, by direct extraction of the residents’ socio-demographic information at nursing home (NH) entry, adding hospitalizations and deaths, and finally, by an iterative textual extraction process of the transmissions data and by using the Delphi method, we created twenty-four syndromes, added hospitalizations and deaths and built a syndromic data base, the Ageing Well data base. This information system of a new kind, allowed the constitution of a public health cohort for elderly people from the BBV residents’population and its syndromic longitudinal follow-up. The BBV was also scientifically assessed for surveillance and public health research through present situation analysis: content, periodicity and data quality. This cohort then gave us the opportunity to build a surveillance tool and follow the residents’ population in real-time by watching their 26 daily frequency syndromic distributions. The methodology for that assessment, Atlanta CDCs’ health surveillance systems method, was used for flu and acute gastro enteritis syndroms and was the second paper subject.- Finally, the building of a new public health tool: each syndrom’s distribution through time (transmissions dates) and space (transmissions NH ids) opened the research field to new data exploration methods. I used these to study different health problems afflicting senior citizens: frequent falls, cancer, vaccinations and the end of life.
150

(Intelligentes) Text Mining in der Marktforschung

Stützer, Cathleen M., Wachenfeld-Schell, Alexandra, Oglesby, Stefan 24 November 2021 (has links)
Die Extraktion von Informationen aus Texten – insbesondere aus unstrukturierten Textdaten wie Foren, Bewertungsportalen bzw. aus offenen Angaben – stellen heute eine besondere Herausforderung für Marktforscher und Marktforscherinnen dar. Hierzu wird zum einen neues methodisches Know-how gebraucht, um mit den komplexen Datenbeständen sowohl bei der Erhebung wie auch bei der Bewertung dieser umzugehen. Zum anderen müssen im Kontext der digitalen Beforschung von neuen Customer Insights sowohl technische als auch organisationale Infrastrukturen geschaffen werden, um u. a. Geschäftsmodelle in Abläufen und Arbeitsprozessen von Unternehmen, Institutionen und Organisationen etablieren zu können. Die Beiträge des Bandes besprechen nicht nur vielfältigste Methoden und Verfahren zur automatischen Textextraktion, sondern zeigen hierbei sowohl die Relevanz als auch die Herausforderungen für die Online-Marktforschung auf, die mit dem Einsatz solch innovativer Ansätze und Verfahren verbunden sind.:C. M. Stützer, A. Wachenfeld-Schell & S. Oglesby: Digitale Transformation der Marktforschung A. Lang & M. Egger, Insius UG: Wie Marktforscher durch kooperatives Natural Language Processing bei der qualitativen Inhaltsanalyse profitieren können M. Heurich & S. Štajner, Symanto Research: Durch Technologie zu mehr Empathie in der Kundenansprache – Wie Text Analytics helfen kann, die Stimme des digitalen Verbrauchers zu verstehen G. Heisenberg, TH Köln & T. Hees, Questback GmbH: Text Mining-Verfahren zur Analyse offener Antworten in Online-Befragungen im Bereich der Markt- und Medienforschung T. Reuter, Cogia Intelligence GmbH: Automatische semantische Analysen für die Online-Marktforschung P. de Buren, Caplena GmbH: Offenen Nennungen gekonnt analysieren

Page generated in 0.117 seconds