• Refine Query
  • Source
  • Publication year
  • to
  • Language
  • 1
  • Tagged with
  • 1
  • 1
  • 1
  • 1
  • 1
  • 1
  • 1
  • 1
  • 1
  • 1
  • 1
  • 1
  • About
  • The Global ETD Search service is a free service for researchers to find electronic theses and dissertations. This service is provided by the Networked Digital Library of Theses and Dissertations.
    Our metadata is collected from universities around the world. If you manage a university/consortium/country archive and want to be added, details can be found on the NDLTD website.
1

Domain Adaptation with a Classifier Trained by Robust Pseudo-Labels

Zhou, Yunke 07 January 2022 (has links)
With the rapid growth of computing power, approaches based on deep learning algorithms have achieved remarkable results in solving computer vision classification problems. These performance improvements are achieved by assuming the source and target data are collected from the same probability distribution. However, this assumption is usually too strict to be satisfied in many real-world applications, such as big data analysis, natural language processing, and computer vision classification problems. Because of distribution discrepancies between these domains, directly training the model on the source domain cannot be expected to generate satisfactory results on the target domain. Therefore, the problem of minimizing these data distribution discrepancies is the main challenge with which modern machine learning is now faced. To address this problem, domain adaptation (DA) aims to identify domain-invariant features between two different but related domains. This thesis proposes a state-of-the-art DA approach that overcomes the limitations of traditional DA methods. To capture fine-grained information for each category, I deploy centroid-to-centroid alignment to perform domain adaptation. An Exponential Moving Average strategy (EMA) is used to ensure we can form robust source and target centroids. A Gaussian-uniform mixture model is trained using an Expectation-Maximization (EM) algorithm to infer the robustness of the target pseudo-labels. With the help of target pseudo-labels, I propose two novel types of classifiers: (1) a target-oriented classifier (TO); and (2) a centroid-oriented classifier (CO). Extensive experiments show that these two classifiers exhibit superior performance on a variety of DA benchmarks when compared to standard baseline methods. / Master of Science / Approaches based on deep learning algorithms have achieved remarkable results in solving computer vision classification problems. These performance improvements are achieved by assuming the source and target data are collected from the same probability distribution; however, in many real-world applications, such as big data analysis, natural language processing, and computer vision classification problems, this assumption is usually too strict to be satisfied. For example, these two domains may have the same types of classes, but the objects in each category of these different domains can vary in shape, color, background, or even illumination. Because the probability distributions are slightly mismatched, directly training the model on one domain cannot achieve a satisfactory result on the other domain. To address this problem, domain adaptation (DA) aims to extract common features on both domains to transfer knowledge from one domain to another. In this thesis, I propose a state-of-the-art DA approach that overcomes the limitation of the traditional DA methods. To capture the low-level information of each category, I deploy centroid-to-centroid alignment to perform domain adaptation. An Exponential Moving Average (EMA) strategy is used to ensure the generation of robust centroids. A Gaussian-Uniform Mixture model is trained by using the Expectation-Maximization (EM) algorithm to infer the robustness of the target sample pseudo-labels. With the help of robust target pseudo-labels, I propose two novel types of classifiers: (1) a target-oriented classifier (TO); and (2) a centroid-oriented classifier (CO). Extensive experiments show that the proposed method outperforms traditional baseline methods on various DA benchmarks.

Page generated in 0.1457 seconds