Spelling suggestions: "subject:"[een] REGULARIZATION"" "subject:"[enn] REGULARIZATION""
241 |
Regularization Methods for Ill-posed ProblemsNeuman, Arthur James, III 15 June 2010 (has links)
No description available.
|
242 |
Space-Frequency Regularization for Qualitative Inverse ScatteringAlqadah, Hatim F. January 2011 (has links)
No description available.
|
243 |
A Geometric Singular Perturbation Theory Approach to Viscous Singular Shocks Profiles for Systems of Conservation LawsHsu, Ting-Hao 14 October 2015 (has links)
No description available.
|
244 |
PARAMETER CHOICES FOR THE SPLIT BREGMAN METHOD APPLIED TO SIGNAL RESTORATIONHashemi, Seyyed Amirreza 20 October 2016 (has links)
No description available.
|
245 |
Regularized Fine-tuning Strategies for Neural Language Models : Application of entropy regularization on GPT-2Hong, Jae Eun January 2022 (has links)
Deep neural language models like GPT-2 is undoubtedly strong at text generation, but often requires special decoding strategies to prevent producing degenerate output - namely repetition. The use of maximum likelihood training objective results in a peaked probability distribution, leading to the over-confidence of neural networks. In this thesis, we explore entropy regularization for a neural language model that can easily smooth peaked output distribution during the fine-tuning process employing GPT-2. We first define the models in three ways: (1) Out of-the box model without fine-tuning process, (2) Fine-tuned model without entropy regularization, and (3) Fine-tuned model with entropy regularization. To investigate the effect of domains on the model, we also divide the dataset into three ways: (1) fine-tuned on heterogeneous dataset, tested on heterogeneous dataset, (2) fine-tuned on homogeneous dataset, tested on homogeneous dataset, and (3) fine-tuned on heterogeneous dataset, tested on homogeneous dataset. In terms of entropy regularization, we experiment controlling the entropy strength parameter (𝛽) in the range of [0.5, 1.0, 2.0, 4.0, 6.0] and annealing the parameter during fine-tuning process. Our findings prove that the entropy-based regularization during fine-tuning process improve the text generation models by significantly reducing the repetition rate without tuning the decoding strategies. As a result of comparing the probabilities of human-generated sentence tokens, it was observed that entropy regularization compensates for the shortcomings of the deterministic decoding method (Beam search) that mostly selects few high-probability words. Various studies have explored entropy regularization in the cold-start training process of neural networks. However, there are not many studies covering the effect of the fine-tuning stage of text generation tasks when employing large scale pre-trained language models. Our findings present strong evidence that one can achieve significant improvement in text generation by way of utilizing entropy regularization, a highly cost-effective approach, during the fine-tuning process.
|
246 |
MRI Velocity Quantification Implementation and Evaluation of Elementary Functions for the Cell Broadband EngineLi, Wei 27 June 2007 (has links)
<p> Magnetic Resonance Imaging (MRI) velocity quantification is addressed in part I of this thesis. In simple MR imaging, data is collected and tissue densities are displayed as images. Moving tissue creates signals which appear as artifacts in the images. In velocity imaging, more data is collected and phase differences are used to quantify the velocity of tissue components. The problem is described and a novel formulation of a regularized, nonlinear inverse problem is proposed. Both Tikhonov and Total Variation Regularization are discussed. Results of numerical simulations show that significant noise reduction is possible.</p> <p> The method is firstly verified with MATLAB. A number of experiments are carried out with different regularization parameters, different magnetic fields and different noise levels. The experiments show that the stronger the complex noise is, the stronger the magnetic field requires for estimating the velocity. The regularization parameter also plays an important role in the experiments. Given the noise level and with an appropriate value of regularization parameter, the estimated velocity converges to ideal velocity very quickly. A proof-of-concept implementation on the Cell BE processor is described, quantifying the performance potential of this platform.</p> <p> The second part of this thesis concerns the evaluation of an elementary function library. Since CBE SPU is designed for compute intensive applications, the well developed Math functions can help developer program and save time to take care other details. Dr. Anand's research group in McMaster developed 28 math functions for CBE SPU. The test tools for accuracy and performance were developed on CBE. The functions were tuned while testing. The functions are either competitive or an addition to the existing SDK1.1 SPU math functions.</p> / Thesis / Master of Applied Science (MASc)
|
247 |
[en] A THEORY BASED, DATA DRIVEN SELECTION FOR THE REGULARIZATION PARAMETER FOR LASSO / [pt] SELECIONANDO O PARÂMETRO DE REGULARIZAÇÃO PARA O LASSO: BASEADO NA TEORIA E NOS DADOSDANIEL MARTINS COUTINHO 25 March 2021 (has links)
[pt] O presente trabalho apresenta uma nova forma de selecionar o parâmetro
de regularização do LASSO e do adaLASSO. Ela é baseada na teoria e
incorpora a estimativa da variância do ruído. Nós mostramos propriedades
teóricas e simulações Monte Carlo que o nosso procedimento é capaz de lidar
com mais variáveis no conjunto ativo do que outras opções populares para a
escolha do parâmetro de regularização. / [en] We provide a new way to select the regularization parameter for the
LASSO and adaLASSO. It is based on the theory and incorporates an estimate
of the variance of the noise. We show theoretical properties of the procedure
and Monte Carlo simulations showing that it is able to handle more variables
in the active set than other popular options for the regularization parameter.
|
248 |
Studies on Subword-based Low-Resource Neural Machine Translation: Segmentation, Encoding, and Decoding / サブワードに基づく低資源ニューラル機械翻訳に関する研究:分割、符号化、及び復号化Haiyue, Song 25 March 2024 (has links)
京都大学 / 新制・課程博士 / 博士(情報学) / 甲第25423号 / 情博第861号 / 新制||情||144(附属図書館) / 京都大学大学院情報学研究科知能情報学専攻 / (主査)特定教授 黒橋 禎夫, 教授 河原 達也, 教授 西野 恒 / 学位規則第4条第1項該当 / Doctor of Informatics / Kyoto University / DFAM
|
249 |
<b>DIFFUSION MODELS OVER DYNAMIC NETWORKS USINGTRANSFORMERS</b>Aniruddha Mukherjee (20414015) 13 December 2024 (has links)
<p dir="ltr">In my thesis, I propose a Graph Regularized-Attention-Based Diffusion Transformer (GRAD-T) model, which uses kernel temporal attention and a regularized sparse graph method to analyze model general diffusion processes over networks. The proposed model uses the spatiotemporal nature of data generated from diffusion processes over networks to examine phenomena that vary across different locations and time, such as disease outbreaks, climate patterns, ecological changes, information flows, news contagion, transportation flows or information and sentiment contagion over social networks. The kernel attention models the temporal dependence of diffusion processes within locations, and the regularized spatial attention mechanism accounts for the spatial diffusion process. The proposed regularization using a combination of penalized matrix estimation and a resampling approach helps in modeling high-dimensional data from large graphical networks, and identify the dominant diffusion pathways. I use the model to predict how emotions spread across sparse networks. I applied the model to a unique dataset of COVID-19 tweets that I curated, spanning April to July 2020 across various U.S. locations. I used model parameters (attention measures) to create indices for comparing emotion diffusion potential within and between nodes. Our findings show that negative emotions like fear, anger, and disgust demonstrate substantial potential for temporal and spatial diffusion. Using the dataset and the proposed method we demonstrate that different types of emotions exhibit different patters of temporal and spatial diffusion. I show that the proposed model improves prediction accuracy of emotion diffusion over social medial networks over standard models such as LSTM and CNN methods. Our key contribution is the regularized graph transformer using a penalty and a resampling approach to enhance the robustness, interpretability, and scalability of sparse graph learning.</p>
|
250 |
Boosting for Learning From Imbalanced, Multiclass Data SetsAbouelenien, Mohamed 12 1900 (has links)
In many real-world applications, it is common to have uneven number of examples among multiple classes. The data imbalance, however, usually complicates the learning process, especially for the minority classes, and results in deteriorated performance. Boosting methods were proposed to handle the imbalance problem. These methods need elongated training time and require diversity among the classifiers of the ensemble to achieve improved performance. Additionally, extending the boosting method to handle multi-class data sets is not straightforward. Examples of applications that suffer from imbalanced multi-class data can be found in face recognition, where tens of classes exist, and in capsule endoscopy, which suffers massive imbalance between the classes. This dissertation introduces RegBoost, a new boosting framework to address the imbalanced, multi-class problems. This method applies a weighted stratified sampling technique and incorporates a regularization term that accommodates multi-class data sets and automatically determines the error bound of each base classifier. The regularization parameter penalizes the classifier when it misclassifies instances that were correctly classified in the previous iteration. The parameter additionally reduces the bias towards majority classes. Experiments are conducted using 12 diverse data sets with moderate to high imbalance ratios. The results demonstrate superior performance of the proposed method compared to several state-of-the-art algorithms for imbalanced, multi-class classification problems. More importantly, the sensitivity improvement of the minority classes using RegBoost is accompanied with the improvement of the overall accuracy for all classes. With unpredictability regularization, a diverse group of classifiers are created and the maximum accuracy improvement reaches above 24%. Using stratified undersampling, RegBoost exhibits the best efficiency. The reduction in computational cost is significant reaching above 50%. As the volume of training data increase, the gain of efficiency with the proposed method becomes more significant.
|
Page generated in 0.0298 seconds