Domain Decomposed Learning Algorithms and Applications / Jingwei Li.

In machine learning (ML), dimensionality reduction (DR) is a common method to reduce the computational cost, and among the linear DR techniques, principal component analysis (PCA) and linear discriminant analysis (LDA) are widely used for unsupervised and supervised learning. In this thesis, we main...

Full description

Saved in:
Bibliographic Details
Online Access: Connect to online resource
Main Author: Li, Jingwei (Author)
Format: Thesis Electronic eBook
Language:English
Published: Ann Arbor : ProQuest Dissertations & Theses, 2023
Subjects:

MARC

LEADER 00000nam a22000003i 4500
001 in00000015394
006 m d
007 cr un
008 230821s2023 miu|||||sm |||||||eng d
005 20230828203114.9
020 |a 9798379532482 
035 |a (MiAaPQD)AAI30421494 
035 |a AAI30421494 
040 |a MiAaPQD  |b eng  |c MiAaPQD  |e rda 
100 1 |a Li, Jingwei,  |e author. 
245 1 0 |a Domain Decomposed Learning Algorithms and Applications /  |c Jingwei Li. 
264 1 |a Ann Arbor :   |b ProQuest Dissertations & Theses,   |c 2023 
300 |a 1 electronic resource (105 pages) 
336 |a text  |b txt  |2 rdacontent 
337 |a computer  |b c  |2 rdamedia 
338 |a online resource  |b cr  |2 rdacarrier 
500 |a Source: Dissertations Abstracts International, Volume: 84-11, Section: B. 
500 |a Advisors: Cai, Xiao-Chuan Committee members: Brown, Jed; Morrison, Rebecca; Schnabel, Robert; Huang, Yu-Jui. 
502 |b Ph.D.  |c University of Colorado at Boulder  |d 2023. 
506 |a This item is not available from ProQuest Dissertations & Theses. 
520 |a In machine learning (ML), dimensionality reduction (DR) is a common method to reduce the computational cost, and among the linear DR techniques, principal component analysis (PCA) and linear discriminant analysis (LDA) are widely used for unsupervised and supervised learning. In this thesis, we mainly propose some new variants of linear reduction methods and an iterative deflation method that addresses some of the limitations of traditional algorithms. For PCA, the reconstruction error is large even when a large number of eigenmodes is used in some applications. In this thesis, we show that an unexpected error source is the pollution effect of a summation operation in the objective function of the PCA algorithm. The summation operator brings together unrelated parts of the data into the same optimization and the result is the reduction of the accuracy of the overall algorithm. To overcome the problem, we introduce a domain decomposed PCA (ddPCA) that improves the accuracy and also increases the parallelism of the algorithm. To demonstrate the accuracy and parallel efficiency of the proposed algorithm, we consider two applications including a face recognition problem, and a brain tumor detection problem using two- and three-dimensional MRI images. For LDA, we propose a domain decomposed method and an iterative deflated method to improve classification accuracy. In the domain decomposed LDA, we decompose the given dataset into subsets and applied LDA separately to each subset for the training phase of the algorithm. In the testing phase, we project the samples into multiple subspaces, contrary to one as in the traditional LDA. From the multiple low-dimensional projections, we determine the class or classes that the sample belongs to. In the iteratively deflated method, the traditional LDA method serves as the initial iteration from which we select separable classes to be deflated from the training dataset, and the remaining samples in the dataset are then used for the next iteration. As the process goes on we generate a sequence of projection matrices that are used to determine which class or classes a sample belongs to using certain classification criteria. With the proper choices of the quantile radii in the separable criteria for the training and testing phases, we show that the proposed method is much more accurate than the traditional LDA. To test these two proposed methods, we consider the CIFAR-10/100 datasets and a gene expression dataset of cancer patients. The results show that the new approaches outperform the traditional LDA by a large margin. 
546 |a English 
590 |a School code: 0051 
650 4 |a Computer science. 
650 4 |a Mathematics. 
650 4 |a Bioinformatics. 
653 |a Bioinfomatics 
653 |a Domain decomposition 
653 |a Linear discriminant analysis 
653 |a Machine learning 
653 |a Principle component analysis 
653 |a Statistics 
655 |a Theses  |x CU Boulder  |x Computer Science.  |2 local. 
700 1 |a Cai, Xiao-Chuan  |e degree supervisor. 
773 0 |t Dissertations Abstracts International  |g 84-11B. 
791 |a Ph.D. 
792 |a 2023 
856 4 0 |z Connect to online resource  |u https://colorado.idm.oclc.org/login?url=http://gateway.proquest.com/openurl?url_ver=Z39.88-2004&rft_val_fmt=info:ofi/fmt:kev:mtx:dissertation&res_dat=xri:pqm&rft_dat=xri:pqdiss:30421494 
999 f f |s 5460cb92-d598-4cea-93c8-8863511704cc  |i 53fc4d62-0d6f-42d5-b264-f05a3b1f78fb 
952 f f |p Can circulate  |a University of Colorado Boulder  |b Online  |c Online  |d Online  |i web