• Refine Query
  • Source
  • Publication year
  • to
  • Language
  • 55
  • 14
  • 12
  • 10
  • 6
  • 4
  • 3
  • 3
  • 3
  • 2
  • 2
  • 1
  • 1
  • 1
  • 1
  • Tagged with
  • 125
  • 125
  • 49
  • 22
  • 22
  • 20
  • 17
  • 16
  • 16
  • 15
  • 13
  • 13
  • 12
  • 11
  • 11
  • About
  • The Global ETD Search service is a free service for researchers to find electronic theses and dissertations. This service is provided by the Networked Digital Library of Theses and Dissertations.
    Our metadata is collected from universities around the world. If you manage a university/consortium/country archive and want to be added, details can be found on the NDLTD website.
81

Context-aware semantic analysis of video metadata

Steinmetz, Nadine January 2013 (has links)
Im Vergleich zu einer stichwortbasierten Suche ermöglicht die semantische Suche ein präziseres und anspruchsvolleres Durchsuchen von (Web)-Dokumenten, weil durch die explizite Semantik Mehrdeutigkeiten von natürlicher Sprache vermieden und semantische Beziehungen in das Suchergebnis einbezogen werden können. Eine semantische, Entitäten-basierte Suche geht von einer Anfrage mit festgelegter Bedeutung aus und liefert nur Dokumente, die mit dieser Entität annotiert sind als Suchergebnis. Die wichtigste Voraussetzung für eine Entitäten-zentrierte Suche stellt die Annotation der Dokumente im Archiv mit Entitäten und Kategorien dar. Textuelle Informationen werden analysiert und mit den entsprechenden Entitäten und Kategorien versehen, um den Inhalt semantisch erschließen zu können. Eine manuelle Annotation erfordert Domänenwissen und ist sehr zeitaufwendig. Die semantische Annotation von Videodokumenten erfordert besondere Aufmerksamkeit, da inhaltsbasierte Metadaten von Videos aus verschiedenen Quellen stammen, verschiedene Eigenschaften und Zuverlässigkeiten besitzen und daher nicht wie Fließtext behandelt werden können. Die vorliegende Arbeit stellt einen semantischen Analyseprozess für Video-Metadaten vor. Die Eigenschaften der verschiedenen Metadatentypen werden analysiert und ein Konfidenzwert ermittelt. Dieser Wert spiegelt die Korrektheit und die wahrscheinliche Mehrdeutigkeit eines Metadatums wieder. Beginnend mit dem Metadatum mit dem höchsten Konfidenzwert wird der Analyseprozess innerhalb eines Kontexts in absteigender Reihenfolge des Konfidenzwerts durchgeführt. Die bereits analysierten Metadaten dienen als Referenzpunkt für die weiteren Analysen. So kann eine möglichst korrekte Analyse der heterogen strukturierten Daten eines Kontexts sichergestellt werden. Am Ende der Analyse eines Metadatums wird die für den Kontext relevanteste Entität aus einer Liste von Kandidaten identifiziert - das Metadatum wird disambiguiert. Hierfür wurden verschiedene Disambiguierungsalgorithmen entwickelt, die Beschreibungstexte und semantische Beziehungen der Entitätenkandidaten zum gegebenen Kontext in Betracht ziehen. Der Kontext für die Disambiguierung wird für jedes Metadatum anhand der Eigenschaften und Konfidenzwerte zusammengestellt. Der vorgestellte Analyseprozess ist an zwei Hypothesen angelehnt: Um die Analyseergebnisse verbessern zu können, sollten die Metadaten eines Kontexts in absteigender Reihenfolge ihres Konfidenzwertes verarbeitet werden und die Kontextgrenzen von Videometadaten sollten durch Segmentgrenzen definiert werden, um möglichst Kontexte mit kohärentem Inhalt zu erhalten. Durch ausführliche Evaluationen konnten die gestellten Hypothesen bestätigt werden. Der Analyseprozess wurden gegen mehrere State-of-the-Art Methoden verglichen und erzielt verbesserte Ergebnisse in Bezug auf Recall und Precision, besonders für Metadaten, die aus weniger zuverlässigen Quellen stammen. Der Analyseprozess ist Teil eines Videoanalyse-Frameworks und wurde bereits erfolgreich in verschiedenen Projekten eingesetzt. / The Semantic Web provides information contained in the World Wide Web as machine-readable facts. In comparison to a keyword-based inquiry, semantic search enables a more sophisticated exploration of web documents. By clarifying the meaning behind entities, search results are more precise and the semantics simultaneously enable an exploration of semantic relationships. However, unlike keyword searches, a semantic entity-focused search requires that web documents are annotated with semantic representations of common words and named entities. Manual semantic annotation of (web) documents is time-consuming; in response, automatic annotation services have emerged in recent years. These annotation services take continuous text as input, detect important key terms and named entities and annotate them with semantic entities contained in widely used semantic knowledge bases, such as Freebase or DBpedia. Metadata of video documents require special attention. Semantic analysis approaches for continuous text cannot be applied, because information of a context in video documents originates from multiple sources possessing different reliabilities and characteristics. This thesis presents a semantic analysis approach consisting of a context model and a disambiguation algorithm for video metadata. The context model takes into account the characteristics of video metadata and derives a confidence value for each metadata item. The confidence value represents the level of correctness and ambiguity of the textual information of the metadata item. The lower the ambiguity and the higher the prospective correctness, the higher the confidence value. The metadata items derived from the video metadata are analyzed in a specific order from high to low confidence level. Previously analyzed metadata are used as reference points in the context for subsequent disambiguation. The contextually most relevant entity is identified by means of descriptive texts and semantic relationships to the context. The context is created dynamically for each metadata item, taking into account the confidence value and other characteristics. The proposed semantic analysis follows two hypotheses: metadata items of a context should be processed in descendent order of their confidence value, and the metadata that pertains to a context should be limited by content-based segmentation boundaries. The evaluation results support the proposed hypotheses and show increased recall and precision for annotated entities, especially for metadata that originates from sources with low reliability. The algorithms have been evaluated against several state-of-the-art annotation approaches. The presented semantic analysis process is integrated into a video analysis framework and has been successfully applied in several projects for the purpose of semantic video exploration of videos.
82

Μελέτη και συγκριτική αξιολόγηση μεθόδων δόμησης περιεχομένου ιστοτόπων : εφαρμογή σε ειδησεογραφικούς ιστοτόπους

Στογιάννος, Νικόλαος-Αλέξανδρος 20 April 2011 (has links)
Η κατάλληλη οργάνωση του περιεχομένου ενός ιστοτόπου, έτσι ώστε να αυξάνεται η ευρεσιμότητα των πληροφοριών και να διευκολύνεται η επιτυχής ολοκλήρωση των τυπικών εργασιών των χρηστών, αποτελεί έναν από τους πρωταρχικούς στόχους των σχεδιαστών ιστοτόπων. Οι υπάρχουσες τεχνικές του πεδίου Αλληλεπίδρασης-Ανθρώπου Υπολογιστή που συνεισφέρουν στην επίτευξη αυτού του στόχου συχνά αγνοούνται εξαιτίας των απαιτήσεών τους σε χρονικούς και οικονομικούς πόρους. Ειδικότερα για ειδησεογραφικούς ιστοτόπους, τόσο το μέγεθος τους όσο και η καθημερινή προσθήκη και τροποποίηση των παρεχόμενων πληροφοριών, καθιστούν αναγκαία τη χρήση αποδοτικότερων τεχνικών για την οργάνωση του περιεχομένου τους. Στην εργασία αυτή διερευνούμε την αποτελεσματικότητα μίας μεθόδου, επονομαζόμενης AutoCardSorter, που έχει προταθεί στη βιβλιογραφία για την ημιαυτόματη κατηγοριοποίηση ιστοσελίδων, βάσει των σημασιολογικών συσχετίσεων του περιεχομένου τους, στο πλαίσιο οργάνωσης των πληροφοριών ειδησεογραφικών ιστοτόπων. Για το σκοπό αυτό διενεργήθηκαν πέντε συνολικά μελέτες, στις οποίες πραγματοποιήθηκε τόσο ποσοτική όσο και ποιοτική σύγκριση των κατηγοριοποιήσεων που προέκυψαν από συμμετέχοντες σε αντίστοιχες μελέτες ταξινόμησης καρτών ανοικτού και κλειστού τύπου, με τα αποτελέσματα της τεχνικής AutoCardSorter. Από την ανάλυση των αποτελεσμάτων προέκυψε ότι η AutoCardSorter παρήγαγε ομαδοποιήσεις άρθρων που βρίσκονται σε μεγάλη συμφωνία με αυτές των συμμετεχόντων στις μελέτες, αλλά με σημαντικά αποδοτικότερο τρόπο, επιβεβαιώνοντας προηγούμενες παρόμοιες μελέτες σε ιστοτόπους άλλων θεματικών κατηγοριών. Επιπρόσθετα, οι μελέτες έδειξαν ότι μία ελαφρώς τροποποιημένη εκδοχή της AutoCardSorter τοποθετεί νέα άρθρα σε προϋπάρχουσες κατηγορίες με αρκετά μικρότερο ποσοστό συμφωνίας συγκριτικά με τον τρόπο που επέλεξαν οι συμμετέχοντες. Η εργασία ολοκληρώνεται με την παρουσίαση κατευθύνσεων για την βελτίωση της αποτελεσματικότητας της AutoCardSorter, τόσο στο πλαίσιο οργάνωσης του περιεχομένου ειδησεογραφικών ιστοτόπων όσο και γενικότερα. / The proper structure of a website's content, so as to increase the findability of the information provided and to ease the typical user task-making, is one of the primary goals of website designers. The existing methods from the field of HCI that assist designers in this, are often neglected due to their high cost and human resources demanded. Even more so on News Sites, their size and the daily content updating call for improved and more efficient techniques. In this thesis we investigate the efficiency of a novel method, called AutoCardSorter, that has been suggested in bibliography for the semi-automatic content categorisation based on the semantic similarity of each webpage-content. To accomplish this we conducted five comparative studies in which the method was compared, to the primary alternatives of the classic Card Sorting method (open, closed). The analysis of the results showed that AutoCardSorter suggested article categories with high relavance to the ones suggested from a group of human subjects participating in the CardSort studies, although in a much more efficient way. This confirms the results of similar previous studies on websites of other themes (eg. travel, education). Moreover, the studies showed that a modified version of the method places articles under pre-existing categories with significant less relavance to the categorisation suggested by the participants. The thesis is concluded with the proposal of different ways to improve the proposed method's efficiency, both in the content of News Sites and in general.
83

Determinantes do sucesso de campanhas de equity e de reward crowdfunding

Felipe, Israel José dos Santos 10 October 2017 (has links)
Submitted by Israel Felipe (israeljfelipe@gmail.com) on 2017-10-23T14:12:53Z No. of bitstreams: 1 Versão final com correções - Tese Israel Felipe FGV.pdf: 2231561 bytes, checksum: 6286843055e43fd7589fe1360de3296a (MD5) / Rejected by Maria Tereza Fernandes Conselmo (maria.conselmo@fgv.br), reason: Israel, O resumo e o abstract tem que ser em uma única pagina. Retirar a informação que esta acima do resumo e abstract, não pode conter aquela informação. Após feito esses ajuste submeter novamente a biblioteca. Tereza SRA on 2017-10-24T15:48:57Z (GMT) / Submitted by Israel Felipe (israeljfelipe@gmail.com) on 2017-10-24T16:15:56Z No. of bitstreams: 2 Versão final com correções - Tese Israel Felipe FGV.pdf: 2231561 bytes, checksum: 6286843055e43fd7589fe1360de3296a (MD5) Versão final com correções - Tese Israel Felipe FGV.pdf: 2228406 bytes, checksum: 2004a64bf28cf51094c363977af1f844 (MD5) / Approved for entry into archive by Maria Tereza Fernandes Conselmo (maria.conselmo@fgv.br) on 2017-10-24T16:40:50Z (GMT) No. of bitstreams: 2 Versão final com correções - Tese Israel Felipe FGV.pdf: 2231561 bytes, checksum: 6286843055e43fd7589fe1360de3296a (MD5) Versão final com correções - Tese Israel Felipe FGV.pdf: 2228406 bytes, checksum: 2004a64bf28cf51094c363977af1f844 (MD5) / Made available in DSpace on 2017-10-24T19:23:24Z (GMT). No. of bitstreams: 2 Versão final com correções - Tese Israel Felipe FGV.pdf: 2231561 bytes, checksum: 6286843055e43fd7589fe1360de3296a (MD5) Versão final com correções - Tese Israel Felipe FGV.pdf: 2228406 bytes, checksum: 2004a64bf28cf51094c363977af1f844 (MD5) Previous issue date: 2017-10-10 / In recent years, the advancement of financial technologies, 'Fintechs', has profoundly modified the way of doing business in several countries. A good example of this is the crowdfunding market, which has benefited from the 'wisdom of the crowds' to enable individuals and small businesses to raise funds necessary for the development of their business. Barriers access to finance, especially collections of consolidated financial information and business-related collateral, have limited the use of financial resources to a broad range of ventures. As a result, crowdfunding has attracted the attention of a large part of the finance and entrepreneurship communities, which point to a vast economic, financial and innovative potential for this modality of financing. However, some studies have reported that crowdfunding may be sensitive to information asymmetry, similar to what occurs with traditional funding channels. Therefore, this research intends to investigate the process of capital allocation in the crowdfunding market, based on approaches that can provide information capable of smoothing the informational friction in the collective funding process. The first approach is the use of semantic analysis of mass media news, which can be applied on the behavior of investors and thus reveal information that serves as a parameter for choosing the best timing for the investment decision and for strategies reduction of business exposure to factors external to the financing process. The second approach consists of the geography of investments, that is, the observation of the geographical characteristics of the origins of the contributions made in the enterprises. Thus, economic and demographic information can be used to select the location of the projects, as well as the geographical orientation of the financing campaign. To fulfill the objective of this research, four trials were developed. The first essay points out a current research agenda around crowdfunding and presents theoretical and empirical discussions on this topic. The second essay identifies the determinants of the success of funding for reward crowdfunding projects, through consideration of headquarters and project attributes. Based on more than 4,200 projects from 417 Brazilian cities, the main results obtained through logit regression and survival analysis suggest that art projects, those that intend larger sums of resources, and those that offer a greater number of rewards are, probably, less successful. On the other hand, projects accounting for more support and those developed in regions with a higher concentration of per capita household income tend to be more successful in their financing campaigns. The third essay investigates the effect of the geographical characteristics of the origin of the contributions and the textual sentiment of the news about the capture of reward crowdfunding. Based on more than 350,000 contribution operations, allocated in more than 2,600 crowdfunding projects, based in 390 Brazilian cities, the research findings indicate that the operational attributes of the projects may lead to increases in the value of the contributions. With respect to geography, the distance between entrepreneur / investor can reduce the tendency of the investor to deposit larger amounts in the ventures. However, the concentration of household income per capita has shown that regions with higher income stocks can contribute more to collective initiatives. Both pessimistic and optimistic news could influence the behavior of contributions in ventures. That is, through a pessimistic scenario, investors tend to deposit smaller amounts in the ventures. On the other hand, in optimistic scenarios, they are inclined to make bigger investments. Finally, the fourth essay analyzes the value of equity investments in equity crowdfunding ventures, through semantic analysis of the media and the geography of financial contributions. Based on 736 investments made in 22 equity crowdfunding ventures located in nine Brazilian cities, the results of the study indicate that the attributes of the ventures may impact the value of the capital allocation. As far as investor behavior on the news is concerned, it is noted that mass news releases with a greater number of positive words can encourage investors to apply larger financial amounts in the ventures. In contrast, the geographical distance between entrepreneur-investor may negatively affect the value of these investments. It is hoped that the empirical results presented here can provide theoretical contributions and managerial implications for academia, professionals, platform managers and public policymakers. / Nos últimos anos, o avanço das tecnologias financeiras, 'Fintechs', tem modificado profundamente o modo de fazer negócios em diversos países. Um bom exemplo disso é o mercado de crowdfunding, o qual tem se beneficiado da ‘sabedoria das multidões’ para permitir que indivíduos e pequenas empresas consigam captar financiamentos necessários para o desenvolvimento de seu negócio. As barreiras de acesso ao financiamento, especialmente as cobranças de informações financeiras consolidadas e de garantias vinculadas aos negócios, têm limitado a utilização de recursos financeiros para um amplo conjunto de empreendimentos. Dessa forma, o crowdfunding, por conseguir flexibilizar exigências excessivas para concessão do crédito, tem despertado a atenção de boa parte das comunidades de finanças e empreendedorismo, as quais apontam um vasto potencial econômico, financeiro e inovador para essa modalidade de financiamento. Entretanto, alguns estudos têm relatado que o crowdfunding pode ser sensível à assimetria de informações, de modo análogo ao que ocorre com os canais tradicionais de financiamento. Assim sendo, esta pesquisa pretende investigar o processo de alocação de capital no mercado de crowdfunding, a partir de abordagens que possam fornecer informações capazes de suavizar o atrito informacional no processo de funding coletivo. A primeira abordagem é a utilização da análise semântica de notícias de mídia de massa, a qual pode ser aplicada sobre o comportamento dos investidores e, assim, revelar informações que sirvam de parâmetro para a escolha do melhor timing para a decisão de investimento e para estratégias de redução de exposição do negócio a fatores externos ao processo de financiamento. A segunda abordagem consiste na geografia dos investimentos, ou seja, na observação das características geográficas das origens das contribuições realizadas nos empreendimentos. Assim, informações econômicas e demográficas poderão servir para a escolha da localização dos empreendimentos, bem como para a orientação geográfica da campanha de financiamento. Para cumprimento do objetivo desta pesquisa, foram desenvolvidos quatro ensaios. O primeiro ensaio aponta uma agenda de pesquisa atual ao redor do crowdfunding e apresenta discussões teóricas e empíricas sobre essa temática. O segundo ensaio identifica os determinantes do sucesso da captação de financiamento para projetos de reward crowdfunding, mediante a consideração da sede e de atributos dos projetos. Com base em mais de 4.200 projetos, oriundos de 417 cidades brasileiras, os principais resultados obtidos via regressão logit e survival analysis sugerem que projetos de arte, aqueles que intencionam maiores somas de recursos, e os que oferecem um maior número de recompensas, são, provavelmente, menos exitosos. Por outro lado, os projetos que contabilizam um maior número de apoios e aqueles desenvolvidos em regiões com maior concentração de renda per capita domiciliar tendem a obter sucesso mais rapidamente em suas campanhas de financiamento. O terceiro ensaio investiga o efeito das características geográficas da origem das contribuições e do sentimento textual das notícias sobre a captação de reward crowdfunding. Baseando-se em mais de 350 mil operações de contribuição, alocadas em mais de 2,6 mil projetos de crowdfunding, sediados em 390 cidades brasileiras, os achados da pesquisa indicam que os atributos operacionais dos projetos podem provocar acréscimos no valor das contribuições. Com relação à geografia, a distância entre empreendedor-investidor pode reduzir a propensão do investidor em depositar maiores quantias nos empreendimentos. Contudo, a concentração de renda domiciliar per capita revelou que regiões com maiores estoques de renda podem contribuir mais com as iniciativas coletivas. Tanto as notícias pessimistas quanto as otimistas foram capazes de influenciar o comportamento das contribuições nos empreendimentos. Isto é, mediante cenário pessimista, os investidores tendem a depositar menores quantias nos empreendimentos. Em contrapartida, em cenários otimistas, eles se inclinam a efetuar maiores investimentos. Por último, o quarto ensaio analisa o valor dos investimentos alocados em empreendimentos de equity crowdfunding, por meio de análise semântica da mídia e da geografia dos aportes financeiros. Com base em 736 investimentos realizados sobre 22 empreendimentos de equity crowdfunding, localizados em nove cidades brasileiras, os resultados do estudo apontam que os atributos dos empreendimentos podem impactar no valor da alocação de capital. No que diz respeito ao comportamento do investidor frente às notícias, nota-se que notícias de massa divulgadas com uma maior quantidade de palavras positivas podem incentivar os investidores a aplicarem maiores quantias financeiras nos empreendimentos. Em contraste, a distância geográfica entre empreendedor-investidor pode atuar negativamente sobre o valor desses investimentos. Espera-se que os resultados empíricos aqui apresentados possam fornecer contribuições teóricas e implicações gerenciais para a academia, profissionais, gestores das plataformas e formadores de políticas públicas.
84

Le Psaume 90 et les fragilités humaines : analyse sémantique et lecture contextuelle / Psalm 90 and the human weakness : semantic analysis and contextual reading

Abelava, Kepezo Robert 12 June 2012 (has links)
Le motif de la fragilité humaine dans le Psaume dit « de Moïse » est appréhendé dans une vision positive. Elle relativise l’apparente souffrance humaine, et suggère la reconnaissance du destin de l’homme, comme une voie de sagesse. Si cette réalité apparaît d’une façon visible dans ce Psaume, c’est grâce à la stratégie que les rédacteurs du Psautier ont exploité. Cette stratégie apparait à travers la disposition des versets dans le Psaume 90 d’une part, d’autre part, à travers la place du Psaume 90 dans le quatrième livre du Psautier. En effet, préoccupé par la question du retour de YHWH qu’il sollicite dès le début du Psaume, le psalmiste sollicite le motif de la fragilité humaine, comme un trait de différence entre YHWH et l’homme, dans le but de persuader ce dernier à reconnaître la grandeur de YHWH, comme maître du temps et de l’histoire. Dès lors, les limites du temps qui s’imposent à la nature de l’homme, ne doivent plus être appréhendées comme un mal, ni comme le résultat d’une quelconque punition de l’homme, mais plutôt comme une caractéristique de la condition humaine. En ce sens, le Psaume 90 considère la fragilité humaine comme un code- cryptogramme de la sagesse et comme un motif, qui offre un accès significatif, pour véhiculer une perception tout à fait originale de la foi à l’époque Post-Exilique. Dans ce contexte en effet, le psalmiste plonge son lecteur dans une méditation où se révèle le sentiment de culpabilité du psalmiste et de sa communauté, à la suite de la catastrophe de l’Exil, et de la chute de la monarchie qu’il vit comme la conséquence de la colère de YHWH. En réponse à cette lamentation, la reprise du motif de la fragilité humaine des Ps 88-89 dans le Psaume 90 et dans le quatrième livre du Psautier, requiert une double importance à savoir, amener l’homme à assumer les échecs, et les souffrances dûes à la chute de la monarchie et à la destruction du temple d’une part, et d’autre part, persuader ce dernier à compter encore et toujours sur la dsx de YHWH.D’une façon générale, les réponses du Psaume 90 sur les fragilités humaines remettent la question de la foi en Dieu au cœur de la vie humaine. Ce Psaume 90 n’élude nullement la réelle souffrance que l’homme éprouve, ni n’encourage une résignation coupable de sa part, mais propose une nouvelle re-définition de l’homme. L’homme peut affirmer sa foi en Dieu, comme maître des temps et de l’histoire,sans être perturbé par les limites que lui impose son état d’être mortel. Ce réalisme de la vie humaine apparaît, comme une condition indispensable pour réaliser l’équilibre entre le réalisme de la vie et l’espérance en Dieu. / The motif of human weakness, in the Psalm known as the Psalm of Moses is apprehended in a positive vision. It relativizes the apparent human suffering, and suggests the recognition of the destiny of man, as a way to wisdom. If this reality appears visibly in this Psalm, it’s thanks to the strategy used by the writers of the Psalter. This strategy appears through the layout of the verses in Psalm 90 on the one hand, and through the place of Psalm 90 in the fourth book of the Psalter on the other hand. Indeed, concerned by the question of the return of YHWH that he seeks from the beginning of the Psalm, the psalmist considers the reason for human frailty, as a sign of the difference between YHWH and man to persuade the latter to recognize the greatness of YHWH, master of time and history. Therefore, the time limits imposed on the nature of the human being, should be understood neither as an evil, nor as the result of any punishment for him, but rather as a feature of the human condition. In this sense, Psalm 90 considers human weakness as a cipher code of wisdom and as a ground, which offers a significant access to convey a quite original perception of the faith at the Post - Exilique time. In this context, indeed, the psalmist plunges his reader into a meditation where the guilt of the psalmist and his community, following the disaster of the Exile and the fall of the monarchy which he saw as a consequence of the wrath of YHWH, is revealed. In response to this lament, the resumption of the motif of the human weakn ess of the Ps 88-89 in Psalm 90 and in the fourth book of the Psalter, requires a double importance that consists, on the one hand, in leading man to assume the failures and the suffering due to the fall of the monarchy and the destruction of the temple, and on the other hand in persuading the latter to rely on YHWH’s dsx again and again.Generally, the responses of Psalm 90 about human weakness put the question of faith in God at the heart of the human life. Psalm 90 doesn’t at all, either elude the real suffering that man experiences, or encourages a guilty resignation on his part, but proposes a new re - definition of man. Man can assert his faith in God the master of time and history, without being disturbed by the limits imposed on him by his state of being mortal. This realism of human life appears as an essential condition to achieve the balance between the realism of life and hope in God.
85

Freie Prädikative in der Satzanalyse im Deutschen : Eine syntaktisch-semantische Analyse

Petersson, Tommy January 2009 (has links)
This survey focuses on the omissible constituents of a sentence, which not only relate to the verb but also describe the subject or object. In a sentence analysis, they cannot be determined by the syntactic structure alone, but need an additional semantic approach. These units are commonly used, but so far have been insufficiently described in grammars and research literature, although the understanding of the function of these units is necessary in order to be able to make a complete sentence analysis. Further, there is no consensus regarding terminology and classification. The term depictive secondary predication is commonly used in English, and the terms prädikatives Attribut, freies Prädikativ and Koprädikativ are the most widely used terms in German. In this study, the term freies Prädikativ (FP) is used. The classification of these constituents in the literature varies between FP, adverbials and in some cases attributes. Based on the information available in German grammars and mainly German research literature, a model was developed, which makes it possible to distinguish FP from adverbials and attributes. A corpus containing German novels and newspapers was analysed in the light of this model. The controlled sentences of grammar books and research literature almost always use adjectives as examples of FP. This study shows that FP can be found in all types of phrases like adjective phrases, participle constructions, als- phrases and prepositional phrases. In addition, it is shown that FP occur with all types of verbs.
86

Uma abordagem de desenvolvimento baseada em modelos de arquitetura organizacional de TI: da semântica ao desenvolvimento de sistemas

Santos Junior, Paulo Sergio dos 16 December 2009 (has links)
Made available in DSpace on 2016-12-23T14:33:46Z (GMT). No. of bitstreams: 1 Dissertacao Paulo Sergio dos Santos Junior.pdf: 7370393 bytes, checksum: ca652e878138a12bfd37e280315d8fc8 (MD5) Previous issue date: 2009-12-16 / The management of organizations is a highly complex activity, since it requires the use of knowledge from several knowledge domains (including business process, information technology and infrastructure). In order to analyze how these domains are interrelated an Enterprise Architecture is essential. The main benefits of using Enterprise Architecture are: (i) to capture the essence and evolution of a business and its information systems; and (ii) to manage the alignment between the business and information systems in a cost-effective manner, possibly by revealing how business processes and information systems are interrelated. To address the alignment between business processes and information systems, several Model-Driven Development approaches enable designers to derive process-oriented systems directly from the business process models through automatic transformation. However, most of these approaches do not enable designers to explore the semantic richness of many Enterprise Architectural models (using only the control flow of business process models), and further define rather inflexible transformations. Moreover, many of these approaches to not clearly separate the development process into platform-independent and platform-specific steps, polluting business process models with platform concerns. This work proposes a novel Model-Driven Development approach to address the aforementioned issues. This approach enable the designers to (i) profit from the semantics of Enterprise Architecture models throughout the system development process; (ii) to apply parameterization in pre-defined transformations and; (iii) to clearly divide the development process into platform-independent and platform-specific steps / O gerenciamento das organizações é uma tarefa que envolve um nível de complexidade significante, uma vez que agrega diversos domínios de conhecimento (incluindo processos de negócios, tecnologias da informação e infraestrutura). Para que seja possível analisar como esses fatores estão interconectados entre si e como a priorização de um deles pode ocasionar a postergação de outro, a utilização de uma Arquitetura Organizacional de TI (Enterprise Architecture) torna-se necessária. Através de uma Arquitetura Organizacional de TI é possível: (i) capturar a essência e as evoluções do negócio e dos sistemas de informação presentes na organização; e (ii) realizar de maneira mais eficaz e menos onerosa o alinhamento entre tecnologia da informação e os processos de negócios executados a uma ou mais organizações. Diversas abordagens de Desenvolvimento Orientado a Modelos utilizam os modelos que representam as Arquiteturas Organizacionais de TI para o desenvolvimento de sistemas computacionais. Porém, a grande maioria das abordagens não permite que o projetista do sistema (i) explore a riqueza semântica dos diferentes modelos da Arquitetura Organizacional de TI (além dos modelos de processos de negócio), (ii) parametrize transformações pré-definidas nessas abordagens para adequar as transformações ao sistema sendo desenvolvido; e (iii) divida o processo de desenvolvimento em etapas independente e dependente de plataforma. Este trabalho propõe uma nova abordagem de Desenvolvimento Orientado a Modelos que visa mitigar estas limitações do estado-da-arte, através de transformações parametrizadas que exploram a riqueza semântica dos modelos das Arquiteturas Organizacionais de TI. Adicionalmente, a abordagem propõe uma divisão clara das etapas independente e dependente de plataforma
87

Analyse sémantique de nuages de points 3D dans le milieu urbain : sol, façades, objets urbains et accessibilité / Semantic analysis of 3D point clouds from urban environments : ground, facades, urban objects and accessibility

Serna Morales, Andrés Felipe 16 December 2014 (has links)
Les plus grandes villes au monde disposent de plans 2D très détaillés des rues et des espaces publics. Ces plans contiennent des informations relatives aux routes, trottoirs, façades et objets urbains tels que, entre autres, les lampadaires, les panneaux de signalisation, les poteaux, et les arbres.De nos jours, certaines autorités locales, agences nationales de cartographie et sociétés privées commencent à adjoindre à leurs cartes de villes des informations en 3D, des choix de navigation et d'accessibilité.En comparaison des premiers systèmes de scanning en 3D d'il y a 30 ans, les scanners laser actuels sont moins chers, plus rapides et fournissent des nuages de points 3D plus précis et plus denses.L'analyse de ces données est difficile et laborieuse, et les méthodes semi-automatiques actuelles risquent de ne pas être suffisamment précises ni robustes. C'est en ce sens que des méthodes automatiques pour l'analyse urbaine sémantique en 3D sont nécessaires.Cette thèse constitue une contribution au domaine de l'analyse sémantique de nuages de points en 3D dans le cadre d'un environnement urbain.Nos méthodes sont basées sur les images d'élévation et elles illustrent l'efficacité de la morphologie mathématique pour développer une chaîne complète de traitement en 3D, incluant 6 étapes principales:i)~filtrage et pré-traitement;ii)~segmentation du sol et analyse d'accessibilité;iii)~segmentation des façades;iv)~détection d'objets;v)~segmentation d'objets;vi)~classification d'objets.De plus, nous avons travaillé sur l'intégration de nos résultats dans une chaîne de production à grande échelle.Ainsi, ceux-ci ont été incorporés en tant que ``shapefiles'' aux Systèmes d'Information Géographique et exportés en tant que nuages de points 3D pour la visualisation et la modélisation.Nos méthodes ont été testées d'un point de vue qualitatif et quantitatif sur plusieurs bases de données issues de l'état de l'art et du projet TerraMobilita.Nos résultats ont montré que nos méthodes s'avèrent précises, rapides et surpassent les travaux décrits par la littérature sur ces mêmes bases.Dans la conclusion, nous abordons également les perspectives de développement futur. / Most important cities in the world have very detailed 2D urban plans of streets and public spaces.These plans contain information about roads, sidewalks, facades and urban objects such as lampposts, traffic signs, bollards, trees, among others.Nowadays, several local authorities, national mapping agencies and private companies have began to consider justifiable including 3D information, navigation options and accessibility issues into urban maps.Compared to the first 3D scanning systems 30 years ago, current laser scanners are cheaper, faster and provide more accurate and denser 3D point clouds.Urban analysis from these data is difficult and tedious, and existing semi-automatic methods may not be sufficiently precise nor robust.In that sense, automatic methods for 3D urban semantic analysis are required.This thesis contributes to the field of semantic analysis of 3D point clouds from urban environments.Our methods are based on elevation images and illustrate how mathematical morphology can be exploited to develop a complete 3D processing chain including six main steps:i)~filtering and preprocessing;ii)~ground segmentation and accessibility analysis;iii)~facade segmentation,iv)~object detection;v)~object segmentation;and, vi)~object classification.Additionally, we have worked on the integration of our results into a large-scale production chain. In that sense, our results have been exported as 3D point clouds for visualization and modeling purposes and integrated as shapefiles into Geographical Information Systems (GIS).Our methods have been qualitative and quantitative tested in several databases from the state of the art and from TerraMobilita project.Our results show that our methods are accurate, fast and outperform other works reported in the literature on the same databases.Conclusions and perspectives for future work are discussed as well.
88

Découverte et analyse des communautés implicites par une approche sémantique en ligne : l'outil WebTribe / Discovery and analysis of implicit communities using an online semantic approach : the WebTribe tool

Leprovost, Damien 30 November 2012 (has links)
Avec l’essor du Web 2.0 et des technologies collaboratives qui y sont rattachées,le Web est aujourd’hui devenu une vaste plate-forme d’échanges entre internautes.La majeure partie des sites Web sont actuellement soit dédiés aux interactionssociales de leurs utilisateurs, soit proposent des outils pour développer ces interactions.Nos travaux portent sur la compréhension de ces échanges, ainsi que desstructures communautaires qui en découlent, au moyen d’une approche sémantique.Pour répondre aux besoins de compréhension propres aux analystes de siteWeb et autres gestionnaires de communautés, nous analysons ces structures communautairespour en extraire des caractéristiques essentielles comme leurs centresthématiques et contributeurs centraux. Notre analyse sémantique s’appuie notammentsur des ontologies légères de référence pour définir plusieurs nouvelles métriques,comme la centralité sémantique temporelle et la probabilité de propagationsémantique. Nous employons une approche « en ligne » afin de suivre l’activitéutilisateur en temps réel, au sein de notre outil d’analyse communautaire Web-Tribe. Nous avons implémenté et testé nos méthodes sur des données extraites desystèmes réels de communication sociale sur le Web / With the rise of Web 2.0 and collaborative technologies that are attached to,the Web has now become a broad platform of exchanges between users. The majorityof websites is now dedicated to social interactions of their users, or offerstools to develop these interactions. Our work focuses on the understanding of theseexchanges, as well as emerging community structures arising, through a semanticapproach. To meet the needs of web analysts, we analyze these community structuresto identify their essential characteristics as their thematic centers and centralcontributors. Our semantic analysis is mainly based on reference light ontologiesto define several new metrics such as the temporal semantic centrality and thesemantic propagation probability. We employ an online approach to monitor useractivity in real time in our community analysis tool WebTribe. We have implementedand tested our methods on real data from social communication systemson the Web
89

Especificação, instanciação e experimentação de um arcabouço para criação automática de ligações hipertexto entre informações homogêneas / Specification, instantion and experimentation of a framework intended to support the task of automatic creation of hypertext links between homogeneous repositories

Alessandra Alaniz Macedo 02 July 2004 (has links)
Com a evolução da informática, diferentes meios de comunicação passaram a explorar a Web como um meio de divulgação de suas informações. Diferentes fontes de informações, diferentes estilos de escrita e a curiosidade nata do ser humano despertam o interesse de leitores por conhecer mais de um relato sobre um mesmo tema. Para que a leitura de diferentes relatos com conteúdo similar seja possível, leitores precisam procurar, ler e analisar informações fornecidas por diferentes fontes de informação. Essa atividade, além de exigir grande investimento de tempo, sobrecarrega cognitivamente usuários. Faz parte das pesquisas da área de Hipermídia investigar mecanismos que apóiem usuários no processo de identificação de informações em repositórios homogêneos, sejam eles disponibilizados na Web ou não. No contexto desta tese, repositórios com informações de conteúdo homogêneo são aqueles cujas informações tratam do mesmo assunto. Esta tese tem por objetivo investigar a especificação, a instanciação e a experimentação de um arcabouço para apoiar a tarefa de criação automática de ligações hipertexto entre repositórios homogêneos. O arcabouço proposto, denominado CARe (Criação Automática de Relacionamentos), é representado por um conjunto de classes que realizam a coleta de informações a serem relacionadas e que processam essas informações para a geração de índices. Esses índices são relacionados e utilizados na criação automática de ligações hipertexto entre a informação original. A definição do arcabouço se deu após uma fase de análise de domínio na qual foram identificados requisitos e construídos componentes de software. Nessa fase, vários protótipos também foram construídos de modo iterativo / With the evolution of the Internet, distinct communication media have focused on the Web as a channel of information publishing. An immediate consequence is an abundance of sources of information and writing styles in the Web. This effect, combining with the inherent curiosity of human beings, has led Web users to look for more than a single article about a same subject. To gain access to separate on a same subject, readers need to search, read and analyze information provided by different sources of information. Besides consuming a great amount of time, that activity imposes a cognitive overhead to users. Several hypermedia researches have investigated mechanisms for supporting users during the process of identifying information on homogeneous repositories, available or not on the Web. In this thesis, homogeneous repositories are those containing information that describes a same subject. This thesis aims at investigating the specification and the construction of a framework intended to support the task of automatic creation of hypertext links between homogeneous repositories. The framework proposed, called CARe (Automatic Creation of Relationships), is composed of a set of classes, methods and relationships that gather information to be related, and also process that information for generating an index. Those indexes are related and used in the automatic creation of hypertext links among distinct excerpts of original information. The framework was defined based on a phase of domain analysis in which requirements were identified and software components were built. In that same phase several prototypes were developed in an iterative prototyping
90

Text mining Twitter social media for Covid-19 : Comparing latent semantic analysis and latent Dirichlet allocation

Sheikha, Hassan January 2020 (has links)
In this thesis, the Twitter social media is data mined for information about the covid-19 outbreak during the month of March, starting from the 3’rd and ending on the 31’st. 100,000 tweets were collected from Harvard’s opensource data and recreated using Hydrate. This data is analyzed further using different Natural Language Processing (NLP) methodologies, such as termfrequency inverse document frequency (TF-IDF), lemmatizing, tokenizing, Latent Semantic Analysis (LSA) and Latent Dirichlet Allocation (LDA). Furthermore, the results of the LSA and LDA algorithms is reduced dimensional data that will be clustered using clustering algorithms HDBSCAN and K-Means for later comparison. Different methodologies are used to determine the optimal parameters for the algorithms. This is all done in the python programing language, as there are libraries for supporting this research, the most important being scikit-learn. The frequent words of each cluster will then be displayed and compared with factual data regarding the outbreak to discover if there are any correlations. The factual data is collected by World Health Organization (WHO) and is then visualized in graphs in ourworldindata.org. Correlations with the results are also looked for in news articles to find any significant moments to see if that affected the top words in the clustered data. The news articles with good timelines used for correlating incidents are that of NBC News and New York Times. The results show no direct correlations with the data reported by WHO, however looking into the timelines reported by news sources some correlation can be seen with the clustered data. Also, the combination of LDA and HDBSCAN yielded the most desireable results in comparison to the other combinations of the dimnension reductions and clustering. This was much due to the use of GridSearchCV on LDA to determine the ideal parameters for the LDA models on each dataset as well as how well HDBSCAN clusters its data in comparison to K-Means.

Page generated in 0.0361 seconds