• Refine Query
  • Source
  • Publication year
  • to
  • Language
  • 397
  • 64
  • 43
  • 26
  • 6
  • 4
  • 4
  • 3
  • 2
  • 2
  • 2
  • 2
  • 1
  • 1
  • 1
  • Tagged with
  • 626
  • 626
  • 284
  • 222
  • 213
  • 150
  • 138
  • 131
  • 101
  • 95
  • 93
  • 88
  • 80
  • 78
  • 78
  • About
  • The Global ETD Search service is a free service for researchers to find electronic theses and dissertations. This service is provided by the Networked Digital Library of Theses and Dissertations.
    Our metadata is collected from universities around the world. If you manage a university/consortium/country archive and want to be added, details can be found on the NDLTD website.
401

Apprentissage et noyau pour les interfaces cerveau-machine / Study of kernel machines towards brain-computer interfaces

Tian, Xilan 07 May 2012 (has links)
Les Interfaces Cerveau-Machine (ICM) ont été appliquées avec succès aussi bien dans le domaine clinique que pour l'amélioration de la vie quotidienne de patients avec des handicaps. En tant que composante essentielle, le module de traitement du signal détermine nettement la performance d'un système ICM. Nous nous consacrons à améliorer les stratégies de traitement du signal du point de vue de l'apprentissage de la machine. Tout d'abord, nous avons développé un algorithme basé sur les SVM transductifs couplés aux noyaux multiples afin d'intégrer différentes vues des données (vue statistique ou vue géométrique) dans le processus d'apprentissage. Deuxièmement, nous avons proposé une version enligne de l'apprentissage multi-noyaux dans le cas supervisé. Les résultats expérimentaux montrent de meilleures performances par rapport aux approches classiques. De plus, l'algorithme proposé permet de sélectionner automatiquement les canaux de signaux EEG utiles grâce à l'apprentissage multi-noyaux.Dans la dernière partie, nous nous sommes attaqués à l'amélioration du module de traitement du signal au-delà des algorithmes d'apprentissage automatique eux-mêmes. En analysant les données ICM hors-ligne, nous avons d'abord confirmé qu'un modèle de classification simple peut également obtenir des performances satisfaisantes en effectuant une sélection de caractéristiques (et/ou de canaux). Nous avons ensuite conçu un système émotionnel ICM par en tenant compte de l'état émotionnel de l'utilisateur. Sur la base des données de l'EEG obtenus avec différents états émotionnels, c'est-à -dire, positives, négatives et neutres émotions, nous avons finalement prouvé que l'émotion affectait les performances ICM en utilisant des tests statistiques. Cette partie de la thèse propose des bases pour réaliser des ICM plus adaptées aux utilisateurs. / Brain-computer Interface (BCI) has achieved numerous successful applications in both clinicaldomain and daily life amelioration. As an essential component, signal processing determines markedly the performance of a BCI system. In this thesis, we dedicate to improve the signal processing strategy from perspective of machine learning strategy. Firstly, we proposed TSVM-MKL to explore the inputs from multiple views, namely, from statistical view and geometrical view; Secondly, we proposed an online MKL to reduce the computational burden involved in most MKL algorithm. The proposed algorithms achieve a better classifcation performance compared with the classical signal kernel machines, and realize an automatical channel selection due to the advantages of MKL algorithm. In the last part, we attempt to improve the signal processing beyond the machine learning algorithms themselves. We first confirmed that simple classifier model can also achieve satisfying performance by careful feature (and/or channel) selection in off-line BCI data analysis. We then implement another approach to improve the BCI signal processing by taking account for the user's emotional state during the signal acquisition procedure. Based on the reliable EEG data obtained from different emotional states, namely, positive, negative and neutral emotions, we perform strict evaluation using statistical tests to confirm that the emotion does affect BCI performance. This part of work provides important basis for realizing user-friendly BCIs.
402

Data-Driven Engine Fault Classification and Severity Estimation Using Residuals and Data

Lundgren, Andreas January 2020 (has links)
Recent technological advances in the automotive industry have made vehicularsystems increasingly complex in terms of both hardware and software. As thecomplexity of the systems increase, so does the complexity of efficient monitoringof these system. With increasing computational power the field of diagnosticsis becoming evermore focused on software solutions for detecting and classifyinganomalies in the supervised systems. Model-based methods utilize knowledgeabout the physical system to device nominal models of the system to detect deviations,while data-driven methods uses historical data to come to conclusionsabout the present state of the system in question. This study proposes a combinedmodel-based and data-driven diagnostic framework for fault classification,severity estimation and novelty detection. An algorithm is presented which uses a system model to generate a candidate setof residuals for the system. A subset of the residuals are then selected for eachfault using L1-regularized logistic regression. The time series training data fromthe selected residuals is labelled with fault and severity. It is then compressedusing a Gaussian parametric representation, and data from different fault modesare modelled using 1-class support vector machines. The classification of datais performed by utilizing the support vector machine description of the data inthe residual space, and the fault severity is estimated as a convex optimizationproblem of minimizing the Kullback-Leibler divergence (kld) between the newdata and training data of different fault modes and severities. The algorithm is tested with data collected from a commercial Volvo car enginein an engine test cell and the results are presented in this report. Initial testsindicate the potential of the kld for fault severity estimation and that noveltydetection performance is closely tied to the residual selection process.
403

Semisupervizované hluboké učení v označování sekvencí / Semi-supervised deep learning in sequence labeling

Páll, Juraj Eduard January 2019 (has links)
Sequence labeling is a type of machine learning problem that involves as- signing a label to each sequence member. Deep learning has shown good per- formance for this problem. However, one disadvantage of this approach is its requirement of having a large amount of labeled data. Semi-supervised learning mitigates this problem by using cheaper unlabeled data together with labeled data. Currently, usage of semi-supervised deep learning for sequence labeling is limited. Therefore, the focus of this thesis is on the application of semi-super- vised deep learning in sequence labeling. Existing semi-supervised deep learning approaches are examined, and approaches for sequence labeling are proposed. The proposed approaches were implemented and experimentally evaluated on named-entity recognition and part-of-speech tagging tasks.
404

Evaluation of In-Silico Labeling for Live Cell Imaging

Sörman Paulsson, Elsa January 2021 (has links)
Today new drugs are tested on cell cultures in wells to minimize time, cost, andanimal testing. The cells are studied using microscopy in different ways and fluorescentprobes are used to study finer details than the light microscopy can observe.This is an invasive method, so instead of molecular analysis, imaging can be used.In this project, phase-contrast microscopy images of cells together with fluorescentmicroscopy images were used. We use Machine Learning to predict the fluorescentimages from the light microscopy images using a strategy called In-Silico Labeling.A Convolutional Neural Network called U-Net was trained and showed good resultson two different datasets. Pixel-wise regression, pixel-wise classification, andimage classification with one cell in each image was tested. The image classificationwas the most difficult part due to difficulties assigning good quality labels tosingle cells. Pixel-wise regression showed the best result.
405

Hyperparameters impact in a convolutional neural network

Bylund, Andreas, Erikssen, Anton, Mazalica, Drazen January 2020 (has links)
Machine learning and image recognition is a big and growing subject in today's society. Therefore the aim of this thesis is to compare convolutional neural networks with different hyperparameter settings and see how the hyperparameters affect the networks test accuracy in identifying images of traffic signs. The reason why traffic signs are chosen as objects to evaluate hyperparameters is due to the author's previous experience in the domain. The object itself that is used for image recognition does not matter. Any dataset with images can be used to see the hyperparameters affect. Grid search is used to create a large amount of models with different width and depth, learning rate and momentum. Convolution layers, activation functions and batch size are all tested separately. These experiments make it possible to evaluate how the hyperparameters affect the networks in their performance of recognizing images of traffic signs. The models are created using Keras API and then trained and tested on the dataset Traffic Signs Preprocessed. The results show that hyperparameters affect test accuracy, some affect more than others. Configuring learning rate and momentum can in some cases result in disastrous results if they are set too high or too low. Activation function also show to be a crucial hyperparameter where it in some cases produce terrible results.
406

Neural Network Classification Approach to Clutter Removal for UTM-Enabling Low-Altitude Radar Surveillance

Emshoff, Brandon Lane January 2021 (has links)
No description available.
407

Training a computer vision model using semi-supervised learning and applying post-training quantizations

Vedin, Albernn January 2022 (has links)
Electrical scooters have gained a lot of attention and popularity among commuters all around the world since they entered the market. After all, electrical scooters have shown to be efficient and cost-effective mode of transportation for commuters and travelers. As of today electrical scooters have firmly established themselves in the micromobility industry, with an increasing global demand.  Although, as the industry is booming so are the accidents as well as getting into dangerous situations of riding electrical scooters. There is a growing concern regarding the safety of the scooters where more and more people are getting injured.   This research focuses on training a computer vision model using semi-supervised learning to help detect traffic rule violations and also prevent collisions for people using electrical scooters. However, applying a computer vision model on an embedded system can be challenging due to the limited capabilities of the hardware. This is where the model can enable post-training quantizations. This thesis examines which post-training quantization has the best performance and if it can perform better compared to the non-quantized model. There are three post-training quantizations that are being applied to the model, dynamic range, full integer and float16 post-training quantizations. The results showed that the non-quantized model achieved a mean average precision (mAP) of 0.03894 with a mean average training and validation loss of 22.10 and 28.11. The non-quantized model was compared with the three post-training quantizations in terms of mAP where the dynamic range post-training quantization achieve the best performance with a mAP of 0.03933.
408

Data Quality Evaluation and Improvement for Machine Learning

Chen, Haihua 05 1900 (has links)
In this research the focus is on data-centric AI with a specific concentration on data quality evaluation and improvement for machine learning. We first present a practical framework for data quality evaluation and improvement, using a legal domain as a case study and build a corpus for legal argument mining. We first created an initial corpus with 4,937 instances that were manually labeled. We define five data quality evaluation dimensions: comprehensiveness, correctness, variety, class imbalance, and duplication, and conducted a quantitative evaluation on these dimensions for the legal dataset and two existing datasets in the medical domain for medical concept normalization. The first group of experiments showed that class imbalance and insufficient training data are the two major data quality issues that negatively impacted the quality of the system that was built on the legal corpus. The second group of experiments showed that the overlap between the test datasets and the training datasets, which we defined as "duplication," is the major data quality issue for the two medical corpora. We explore several widely used machine learning methods for data quality improvement. Compared to pseudo-labeling, co-training, and expectation-maximization (EM), generative adversarial network (GAN) is more effective for automated data augmentation, especially when a small portion of labeled data and a large amount of unlabeled data is available. The data validation process, the performance improvement strategy, and the machine learning framework for data evaluation and improvement discussed in this dissertation can be used by machine learning researchers and practitioners to build high-performance machine learning systems. All the materials including the data, code, and results will be released at: https://github.com/haihua0913/dissertation-dqei.
409

Transfer learning techniques in time series analysis

Sablons de Gélis, Robinson January 2021 (has links)
Deep learning works best with vast andd well-distributed data collections. However, collecting and annotating large data sets can be very time-consuming and expensive. Moreover, deep learning is specific to domain knowledge, even with data and computation. E.g., models trained to classify animals would probably underperform when they classify vehicles. Although techniques such as domain adaptation and transfer learning have been popularised recently, tasks in cross-domain knowledge transfer have also taken off. However, most of these works are limited to computer vision. In the domain of time series, this is relatively underexplored. This thesis explores methods to use time series data from one domain to classify data generated from another domain via transfer learning. It focuses on using accelerometer data from running recordings to improve the classification performance on jumping data based on the apparent similarity of individual recordings. Thus, transfer learning and domain adaptation techniques were used to use the learning acquired through deep model training on running sequences. This thesis has performed four experiments to test this domain similarity. The first one consists of transforming time series with the continuous wavelet transform to get both time and frequency information. The model is then pre-trained within a contrastive learning framework. However, the continuous wavelet transformation (CWT) did not improve the classification results. The following two experiments consisted of pre-training the models with self-supervised learning. The first one with a contrastive pretext-task improved the classification results, and the resilience to data decrease. The second one with a forward forecasting pretext-task improved the results when all the data was available but was very sensitive to data decrease. Finally, the domain adaptation was tested and showed interesting performances on the classification task. Although some of the employed techniques did not show improvement, pre-training using contrastive learning on the running dataset has shown great improvement to classify the jumping dataset. / Djupinlärning fungerar bäst med stora och väl distribuerade datasamlingar. Det kan dock vara mycket tidskrävande och dyrt att samla in och kommentera stora datamängder. Även med alla data och beräkningar är djupinlärning specifik för domänkunskap. Exempelvis skulle modeller som tränats för att klassificera djur förmodligen underprestera när de klassificerar fordon. Även om tekniker som domänanpassning och överföringsinlärning har populariserats på senare tid, har även uppgifter inom kunskapsöverföring mellan olika domäner tagit fart. De flesta av dessa arbeten är dock begränsade till datorseende. Inom tidsseriernas område är detta relativt outforskat. I den här avhandlingen undersöks metoder för att använda tidsseriedata från en domän för att klassificera data från en annan domän med hjälp av djupinlärning. Fokus ligger på att använda accelerometerdata från löpning för att förbättra klassificeringen av hoppdata, baserat på den uppenbara likheten mellan löpning och hoppning. Således användes tekniker för överföringsinlärning och domänanpassning för att använda den inlärning som förvärvats genom träning av djupa modeller på löpsekvenser. I den här avhandlingen har fyra experiment utförts för att testa denna domänlikhet. Det första består av att omvandla tidsserier med den kontinuerliga wavelettransformen för att få fram både tids- och frekvensinformation. Modellen förtränas sedan inom en ram för kontrastiv inlärning. Användningen av CWT förbättrade dock inte klassificeringsresultaten. De följande två experimenten bestod av att förträna modellerna med självövervakad inlärning. Det första försöket med en kontrasterande förtextuppgift förbättrade klassificeringsresultaten och motståndskraften mot dataförlust. Det andra försöket med en prognostiserande förtextuppgift förbättrade resultaten när alla data var tillgängliga, men var mycket känslig för dataförlust. Slutligen testades domänanpassningen och visade intressanta resultat i klassificeringsuppgiften. Även om några av de använda teknikerna inte visade någon förbättring, har förträning med hjälp av kontrastinlärning på löpande dataset visat sig ge stora förbättringar när det gäller klassificering av hoppdata.
410

Identifying electrons with deep learning methods

Kahya, Emre Onur 12 1900 (has links)
Cette thèse porte sur les techniques de l’apprentissage machine et leur application à un problème important de la physique des particules expérimentale: l’identification des électrons de signal résultant des collisions proton-proton au Grand collisionneur de hadrons. Au chapitre 1, nous fournissons des informations sur le Grand collisionneur de hadrons et expliquons pourquoi il a été construit. Nous présentons ensuite plus de détails sur ATLAS, l’un des plus importants détecteurs du Grand collisionneur de hadrons. Ensuite, nous expliquons en quoi consiste la tâche d’identification des électrons ainsi que l’importance de bien la mener à terme. Enfin, nous présentons des informations détaillées sur l’ensemble de données que nous utilisons pour résoudre cette tâche d’identification des électrons. Au chapitre 2, nous donnons une brève introduction des principes fondamentaux de l’apprentissage machine. Après avoir défini et introduit les différents types de tâche d’apprentissage, nous discutons des diverses façons de représenter les données d’entrée. Ensuite, nous présentons ce qu’il faut apprendre de ces données et comment y parvenir. Enfin, nous examinons les problèmes qui pourraient se présenter en régime de “sur-apprentissage”. Au chapitres 3, nous motivons le choix de l’architecture choisie pour résoudre notre tâche, en particulier pour les sections où des images séquentielles sont utilisées comme entrées. Nous présentons ensuite les résultats de nos expériences et montrons que notre modèle fonctionne beaucoup mieux que les algorithmes présentement utilisés par la collaboration ATLAS. Enfin, nous discutons des futures orientations afin d’améliorer davantage nos résultats. Au chapitre 4, nous abordons les deux concepts que sont la généralisation hors distribution et la planéité de la surface associée à la fonction de coût. Nous prétendons que les algorithmes qui font converger la fonction coût vers minimum couvrant une région large et plate sont également ceux qui offrent le plus grand potentiel de généralisation pour les tâches hors distribution. Nous présentons les résultats de l’application de ces deux algorithmes à notre ensemble de données et montrons que cela soutient cette affirmation. Nous terminons avec nos conclusions. / This thesis is about applying the tools of Machine Learning to an important problem of experimental particle physics: identifying signal electrons after proton-proton collisions at the Large Hadron Collider. In Chapters 1, we provide some information about the Large Hadron Collider and explain why it was built. We give further details about one of the biggest detectors in the Large Hadron Collider, the ATLAS. Then we define what electron identification task is, as well as the importance of solving it. Finally, we give detailed information about our dataset that we use to solve the electron identification task. In Chapters 2, we give a brief introduction to fundamental principles of machine learning. Starting with the definition and types of different learning tasks, we discuss various ways to represent inputs. Then we present what to learn from the inputs as well as how to do it. And finally, we look at the problems that would arise if we “overdo” learning. In Chapters 3, we motivate the choice of the architecture to solve our task, especially for the parts that have sequential images as inputs. We then present the results of our experiments and show that our model performs much better than the existing algorithms that the ATLAS collaboration currently uses. Finally, we discuss future directions to further improve our results. In Chapter 4, we discuss two concepts: out of distribution generalization and flatness of loss surface. We claim that the algorithms, that brings a model into a wide flat minimum of its training loss surface, would generalize better for out of distribution tasks. We give the results of implementing two such algorithms to our dataset and show that it supports our claim. Finally, we end with our conclusions.

Page generated in 0.096 seconds