• Refine Query
  • Source
  • Publication year
  • to
  • Language
  • 200
  • 70
  • 23
  • 22
  • 21
  • 8
  • 5
  • 3
  • 3
  • 3
  • 3
  • 2
  • 2
  • 2
  • 1
  • Tagged with
  • 442
  • 442
  • 442
  • 177
  • 145
  • 99
  • 86
  • 73
  • 72
  • 58
  • 55
  • 55
  • 54
  • 49
  • 48
  • About
  • The Global ETD Search service is a free service for researchers to find electronic theses and dissertations. This service is provided by the Networked Digital Library of Theses and Dissertations.
    Our metadata is collected from universities around the world. If you manage a university/consortium/country archive and want to be added, details can be found on the NDLTD website.
241

Improving Filtering of Email Phishing Attacks by Using Three-Way Text Classifiers

Trevino, Alberto 13 March 2012 (has links) (PDF)
The Internet has been plagued with endless spam for over 15 years. However, in the last five years spam has morphed from an annoying advertising tool to a social engineering attack vector. Much of today's unwanted email tries to deceive users into replying with passwords, bank account information, or to visit malicious sites which steal login credentials and spread malware. These email-based attacks are known as phishing attacks. Much has been published about these attacks which try to appear real not only to users and subsequently, spam filters. Several sources indicate traditional content filters have a hard time detecting phishing attacks because the emails lack the traditional features and characteristics of spam messages. This thesis tests the hypothesis that by separating the messages into three categories (ham, spam and phish) content filters will yield better filtering performance. Even though experimentation showed three-way classification did not improve performance, several additional premises were tested, including the validity of the claim that phishing emails are too much like legitimate emails and the ability of Naive Bayes classifiers to properly classify emails.
242

Maskininlärning för att förutspå churn baserat på diskontinuerlig beteendedata / Machine learning to predict churn based on discontinuous behavioral data

Öbom, Anton, Bratteby, Adrian January 2017 (has links)
This report is about examining the fields of machine learning and digital marketing, using machine learning as a tool to predict churn in a new domain of companies that do not track their customers extensively, i.e where behaviour data is discontinuous.  To predict churn relatively simple out of the box models, such as support vector machines and random forests, are used to achieve an acceptable outcome. To be on par with the models used for churn prediction in subscription based services, this report concludes that more research has to be done using more effective evaluation metrics. Finally it is presented how these discoveries can be commercialized and the business related benefits of using churn prediction for the employer Sellpy. / Denna rapport handlar om att utforska fälten maskininlärning och digital marknadsföring, genom att använda maskininlärning som ett redskap för att förutspå churn i en typ av företag med diskontinuerlig beteendedata. För att förutspå churn finns relativt simpla "out of the box"-modeller, som support vector machines och random forests, som används för att nå acceptabla resultat. För att nå liknande resultat som i arbeten där churn utförs på kontinuerlig beteendedata konstaterar denna rapport att framtida arbeten forska på vilka utvärderingsmetriker som är mest lämpade. I rapporten presenteras också hur dessa upptäckter kan kommersialiseras och hur företaget Sellpy kan tjäna på att förutspå churn.
243

Tillämpning av maskininlärning för att införa automatisk adaptiv uppvärmning genom en studie på KTH Live-In Labs lägenheter / Using machine learning to implement adaptive heating; A study on KTH Live-In Labs apartments

Åsenius, Ingrid January 2020 (has links)
The purpose of this study is to investigate if it is possible to decrease Sweden's energy consumption through adaptive heating that uses climate data to detect occupancy in apartments using machine learning. The application of the study has been made using environmental data from one of KTH Live-In Labs apartments. The data was first used to investigate the possibility to detect occupancy through machine learning and was then used as input in an adaptive heating model to investigate potential benefits on the energy consumption and costs of heating. The result of the study show that occupancy can be detected using environmental data but not with 100% accuracy. It also shows that the features that have greatest impact in detecting occupancy is light and carbon dioxide and that the best performing machine learning algorithm, for the used dataset, is the Decision Tree algorithm. The potential energy savings through adaptive heating was estimated to be up to 10,1%. In the final part of the paper, it is discussed how a value creating service can be created around adaptive heating and its possibility to reach the market. / Syftet med den här rapporten är att undersöka om det är möjligt att sänka Sveriges energikonsumtion genom att i lägenheter införa adaptiv uppvärmning som baserar sig på närvaroklassificering av klimatdata. Klimatdatan som använts i studien är tagen från en av KTH Live-In Labs lägenheter. Datan användes först för att undersöka om det var möjligt att detektera närvaro  genom maskininlärning och sedan som input i en modell för adaptiv uppvärmning. I modellen för adaptiv uppvärmning undersöktes de potentiella besparingarna i energibehov och uppvärmningskostnader. Resultaten visar att de bästa featuresen för att klassificera närvaro är ljus och koldioxid. Den maskininlärningsalgoritm som presterade bäst på datasetet var Decision Tree algoritmen. Den potentiella energibesparingen genom införandet av adaptiv uppvärmning uppskattas vara upp till 10,1%. I rapportens sista del diskuteras det hur en värdeskapande tjänst kan skapas kring adaptiv uppvärmning samt dess potential att nå marknaden.
244

Efficient Techniques For Relevance Feedback Processing In Content-based Image Retrieval

Liu, Danzhou 01 January 2009 (has links)
In content-based image retrieval (CBIR) systems, there are two general types of search: target search and category search. Unlike queries in traditional database systems, users in most cases cannot specify an ideal query to retrieve the desired results for either target search or category search in multimedia database systems, and have to rely on iterative feedback to refine their query. Efficient evaluation of such iterative queries can be a challenge, especially when the multimedia database contains a large number of entries, and the search needs many iterations, and when the underlying distance measure is computationally expensive. The overall processing costs, including CPU and disk I/O, are further emphasized if there are numerous concurrent accesses. To address these limitations involved in relevance feedback processing, we propose a generic framework, including a query model, index structures, and query optimization techniques. Specifically, this thesis has five main contributions as follows. The first contribution is an efficient target search technique. We propose four target search methods: naive random scan (NRS), local neighboring movement (LNM), neighboring divide-and-conquer (NDC), and global divide-and-conquer (GDC) methods. All these methods are built around a common strategy: they do not retrieve checked images (i.e., shrink the search space). Furthermore, NDC and GDC exploit Voronoi diagrams to aggressively prune the search space and move towards target images. We theoretically and experimentally prove that the convergence speeds of GDC and NDC are much faster than those of NRS and recent methods. The second contribution is a method to reduce the number of expensive distance computation when answering k-NN queries with non-metric distance measures. We propose an efficient distance mapping function that transfers non-metric measures into metric, and still preserves the original distance orderings. Then existing metric index structures (e.g., M-tree) can be used to reduce the computational cost by exploiting the triangular inequality property. The third contribution is an incremental query processing technique for Support Vector Machines (SVMs). SVMs have been widely used in multimedia retrieval to learn a concept in order to find the best matches. SVMs, however, suffer from the scalability problem associated with larger database sizes. To address this limitation, we propose an efficient query evaluation technique by employing incremental update. The proposed technique also takes advantage of a tuned index structure to efficiently prune irrelevant data. As a result, only a small portion of the data set needs to be accessed for query processing. This index structure also provides an inexpensive means to process the set of candidates to evaluate the final query result. This technique can work with different kernel functions and kernel parameters. The fourth contribution is a method to avoid local optimum traps. Existing CBIR systems, designed around query refinement based on relevance feedback, suffer from local optimum traps that may severely impair the overall retrieval performance. We therefore propose a simulated annealing-based approach to address this important issue. When a stuck-at-a-local-optimum occurs, we employ a neighborhood search technique (i.e., simulated annealing) to continue the search for additional matching images, thus escaping from the local optimum. We also propose an index structure to speed up such neighborhood search. Finally, the fifth contribution is a generic framework to support concurrent accesses. We develop new storage and query processing techniques to exploit sequential access and leverage inter-query concurrency to share computation. Our experimental results, based on the Corel dataset, indicate that the proposed optimization can significantly reduce average response time while achieving better precision and recall, and is scalable to support a large user community. This latter performance characteristic is largely neglected in existing systems making them less suitable for large-scale deployment. With the growing interest in Internet-scale image search applications, our framework offers an effective solution to the scalability problem.
245

IMAGE CAPTIONING FOR REMOTE SENSING IMAGE ANALYSIS

Hoxha, Genc 09 August 2022 (has links)
Image Captioning (IC) aims to generate a coherent and comprehensive textual description that summarizes the complex content of an image. It is a combination of computer vision and natural language processing techniques to encode the visual features of an image and translate them into a sentence. In the context of remote sensing (RS) analysis, IC has been emerging as a new research area of high interest since it not only recognizes the objects within an image but also describes their attributes and relationships. In this thesis, we propose several IC methods for RS image analysis. We focus on the design of different approaches that take into consideration the peculiarity of RS images (e.g. spectral, temporal and spatial properties) and study the benefits of IC in challenging RS applications. In particular, we focus our attention on developing a new decoder which is based on support vector machines. Compared to the traditional decoders that are based on deep learning, the proposed decoder is particularly interesting for those situations in which only a few training samples are available to alleviate the problem of overfitting. The peculiarity of the proposed decoder is its simplicity and efficiency. It is composed of only one hyperparameter, does not require expensive power units and is very fast in terms of training and testing time making it suitable for real life applications. Despite the efforts made in developing reliable and accurate IC systems, the task is far for being solved. The generated descriptions are affected by several errors related to the attributes and the objects present in an RS scene. Once an error occurs, it is propagated through the recurrent layers of the decoders leading to inaccurate descriptions. To cope with this issue, we propose two post-processing techniques with the aim of improving the generated sentences by detecting and correcting the potential errors. They are based on Hidden Markov Model and Viterbi algorithm. The former aims to generate a set of possible states while the latter aims at finding the optimal sequence of states. The proposed post-processing techniques can be injected to any IC system at test time to improve the quality of the generated sentences. While all the captioning systems developed in the RS community are devoted to single and RGB images, we propose two captioning systems that can be applied to multitemporal and multispectral RS images. The proposed captioning systems are able at describing the changes occurred in a given geographical through time. We refer to this new paradigm of analysing multitemporal and multispectral images as change captioning (CC). To test the proposed CC systems, we construct two novel datasets composed of bitemporal RS images. The first one is composed of very high-resolution RGB images while the second one of medium resolution multispectral satellite images. To advance the task of CC, the constructed datasets are publically available in the following link: https://disi.unitn.it/~melgani/datasets.html. Finally, we analyse the potential of IC for content based image retrieval (CBIR) and show its applicability and advantages compared to the traditional techniques. Specifically, we focus our attention on developing a CBIR systems that represents an image with generated descriptions and uses sentence similarity to search and retrieve relevant RS images. Compare to traditional CBIR systems, the proposed system is able to search and retrieve images using either an image or a sentence as a query making it more comfortable for the end-users. The achieved results show the promising potentialities of our proposed methods compared to the baselines and state-of-the art methods.
246

Current based condition monitoring of electromechanical systems. Model-free drive system current monitoring: faults detection and diagnosis through statistical features extraction and support vector machines classification.

Bin Hasan, M.M.A. January 2012 (has links)
A non-invasive, on-line method for detection of mechanical (rotor, bearings eccentricity) and stator winding faults in a 3-phase induction motors from observation of motor line current supply input. The main aim is to avoid the consequence of unexpected failure of critical equipment which results in extended process shutdown, costly machinery repair, and health and safety problems. This thesis looks into the possibility of utilizing machine learning techniques in the field of condition monitoring of electromechanical systems. Induction motors are chosen as an example for such application. Electrical motors play a vital role in our everyday life. Induction motors are kept in operation through monitoring its condition in a continuous manner in order to minimise their off times. The author proposes a model free sensor-less monitoring system, where the only monitored signal is the input to the induction motor. The thesis considers different methods available in literature for condition monitoring of induction motors and adopts a simple solution that is based on monitoring of the motor current. The method proposed use the feature extraction and Support Vector Machines (SVM) to set the limits for healthy and faulty data based on the statistical methods. After an extensive overview of the related literature and studies, the motor which is the virtual sensor in the drive system is analysed by considering its construction and principle of operation. The mathematical model of the motor is used for analysing the system. This is followed by laboratory testing of healthy motors and comparing their output signals with those of the same motors after being intentionally failed, concluding with the development of a full monitoring system. Finally, a monitoring system is proposed that can detect the presence of a fault in the monitored machine and diagnose the fault type and severity / Ministry of Higher Education, Libya; Switchgear & Instruments Ltd.
247

A Comparison of Signal Processing and Classification Methods for Brain-Computer Interface

Renfrew, Mark E. January 2009 (has links)
No description available.
248

A Probabilistic Technique For Open Set Recognition Using Support Vector Machines

Scherreik, Matthew January 2014 (has links)
No description available.
249

A SNP Microarray Analysis Pipeline Using Machine Learning Techniques

Evans, Daniel T. January 2010 (has links)
No description available.
250

Exploration of Acoustic Features for Automatic Vowel Discrimination in Spontaneous Speech

Tyson, Na'im R. 26 June 2012 (has links)
No description available.

Page generated in 0.0433 seconds