Data Sparse Algorithms and Mathematical Theory for Large-scale Machine Learning Problems

Data Sparse Algorithms and Mathematical Theory for Large-scale Machine Learning Problems
Author :
Publisher :
Total Pages :
Release :
ISBN-10 : OCLC:1050978856
ISBN-13 :
Rating : 4/5 (56 Downloads)

Book Synopsis Data Sparse Algorithms and Mathematical Theory for Large-scale Machine Learning Problems by : Ruoxi Wang

Download or read book Data Sparse Algorithms and Mathematical Theory for Large-scale Machine Learning Problems written by Ruoxi Wang and published by . This book was released on 2018 with total page pages. Available in PDF, EPUB and Kindle. Book excerpt: This dissertation presents scalable algorithms for high-dimensional large-scale datasets in machine learning applications. The ability to generate data at the scale of millions and even billions has increased rapidly, posing computational challenges to most machine learning algorithms. I propose fast kernel-matrix-based algorithms that avoid intensive kernel matrix operations and neural-network-based algorithms that efficiently learn feature interactions. My contributions include: 1) A structured low-rank approximation method--the Block Basis Factorization (BBF)--that reduces the training time and memory for kernel methods from quadratic to linear and enjoys better accuracy than state-of-art kernel approximation algorithms. 2) Mathematical theories for the ranks of RBF kernel matrices generated from high-dimensional datasets. 3) A parallel black-box fast multipole method (FMM) software library--PBBFMM3D--that evaluates particle interactions in 3D. 4) A neural network--the Deep & Cross Network (DCN)--for web-scale data predictions that requires no exhaustive feature searching nor manual feature engineering and efficiently learns bounded-degree feature interactions combined with complex deep representations. Chapter 2 presents BBF, which accelerates kernel methods by factorizing an n by n kernel matrix into a sparse representation with O(n) nonzero entries as compared to O(n^2). By identifying the low-rank properties of certain blocks, BBF extends the domain of applicability of low-rank approximation methods to the cases where traditional low-rank approximations are inefficient. By leveraging the knowledge from numerical linear algebra and randomized algorithms, the factorization can be constructed in O(n) time complexity while being accurate and stable. Our empirical results demonstrate the stability and superiority over the state-of-art kernel approximation algorithms. Chapter 3 presents a theoretical analysis of the RBF kernel matrix rank. Our three main results are as follows. First, we study the kernel rank, which for a fixed precision grows algebraically with the data dimension (in the worst case), and where the power is related to the accuracy. Second, we derive precise error bounds for the low-rank approximation in the L_infty norm in terms of the function smoothness and the domain diameters. And third, we analyze a group pattern in the magnitude of the singular values of the RBF kernel matrix. We explain this pattern by a grouping of the expansion terms in the kernel's low-rank representation. Empirical results verify the theoretical results. Chapter 4 presents PBBFMM3D, which is a parallel implementation of the fast multipole method (FMM) for evaluating pair-wise particle interactions (matrix-vector product) in three dimensions. PBBFMM3D applies to all non-oscillatory smooth kernel functions and only requires the kernel evaluations at data points. It has O(N) complexity as opposed to O(N^2) complexity from a direct computation. We discuss several algorithmic improvements and performance optimizations, such as shared memory parallelism using OpenMP. We present convergence and scalability results, as well as applications including particle potential evaluations, which frequently occur in PDE-related simulations, and covariance matrix computations that are essential parts in parameter estimation techniques, e.g., Kriging and Kalman filtering. Chapter 5 presents DCN, which is designed for datasets with dense and sparse combined features and enables automatic and efficient feature learning. Feature engineering is the key to the success of prediction models; however, the process often requires manual feature engineering or exhaustive searching. DCN combines a deep neural network that learns complex but implicit feature interactions, with a novel cross network that is more efficient in learning certain explicit bounded-degree feature interactions. Our experimental results have demonstrated its superiority over the state-of-art algorithms on the click-through-rate prediction dataset and dense classification dataset, in terms of both model accuracy and memory usage.


Data Sparse Algorithms and Mathematical Theory for Large-scale Machine Learning Problems Related Books

Data Sparse Algorithms and Mathematical Theory for Large-scale Machine Learning Problems
Language: en
Pages:
Authors: Ruoxi Wang
Categories:
Type: BOOK - Published: 2018 - Publisher:

DOWNLOAD EBOOK

This dissertation presents scalable algorithms for high-dimensional large-scale datasets in machine learning applications. The ability to generate data at the s
Mathematical Theories of Machine Learning - Theory and Applications
Language: en
Pages: 133
Authors: Bin Shi
Categories: Technology & Engineering
Type: BOOK - Published: 2019-06-12 - Publisher: Springer

DOWNLOAD EBOOK

This book studies mathematical theories of machine learning. The first part of the book explores the optimality and adaptivity of choosing step sizes of gradien
Large-Scale Convex Optimization
Language: en
Pages: 320
Authors: Ernest K. Ryu
Categories: Mathematics
Type: BOOK - Published: 2022-12-01 - Publisher: Cambridge University Press

DOWNLOAD EBOOK

Starting from where a first course in convex optimization leaves off, this text presents a unified analysis of first-order optimization methods – including pa
Stochastic Optimization for Large-scale Machine Learning
Language: en
Pages: 177
Authors: Vinod Kumar Chauhan
Categories: Computers
Type: BOOK - Published: 2021-11-18 - Publisher: CRC Press

DOWNLOAD EBOOK

Advancements in the technology and availability of data sources have led to the `Big Data' era. Working with large data offers the potential to uncover more fin
Sparse Learning Under Regularization Framework
Language: en
Pages: 152
Authors: Haiqin Yang
Categories:
Type: BOOK - Published: 2011-04 - Publisher: LAP Lambert Academic Publishing

DOWNLOAD EBOOK

Regularization is a dominant theme in machine learning and statistics due to its prominent ability in providing an intuitive and principled tool for learning fr