Information theoretic approaches to statistical learning problems

Placeholder Show Content

Abstract/Contents

Abstract
Machine learning algorithms have achieved super-human scores in several benchmark tasks from various areas such as computer vision, natural language processing, and speech processing. These algorithms, however, have been observed to lack sufficient robustness and generalizability necessary for their safe application to high-risk tasks such as designing self-driving cars and healthcare systems. A state-of-the-art learning algorithm demonstrates great generalization behavior over one particular domain of data, while the same algorithm can generalize poorly in other practical domains or under a slight domain shift such as adding minor additive noise to its input. Resolving such issues and improving robustness and generalization properties of learning algorithms require a thorough theoretical analysis of statistical learning settings, raising the need for novel theoretical approaches to analyze machine learning problems. This doctoral dissertation introduces several new theoretical frameworks for analyzing generalization and robustness in statistical learning settings. The proposed frameworks are inspired by existing tools and ideas from the following fields of study: information theory, convex analysis, Fourier analysis, and optimal transport theory. More specifically, this dissertation is divided into three parts. The first part focuses on developing information theoretic approaches to supervised learning problems. We consider and optimize well-known information measures such as entropy, mutual information, and maximal correlation to design robust supervised learning and feature selection algorithms. The second part discusses novel information theoretic frameworks for generative adversarial nets (GANs) and group-structured adversarial learning problems. We use convex analysis tools as well as optimal transport costs to develop those frameworks. The third part is dedicated to understanding generalization in various machine learning contexts such as classic kernel learning and modern adversarial deep learning. We introduce novel Fourier-based and spectral approaches to better understand generalization in different learning contexts and propose several regularization tools for improving the generalization performance in those learning settings.

Description

Type of resource text
Form electronic resource; remote; computer; online resource
Extent 1 online resource.
Place California
Place [Stanford, California]
Publisher [Stanford University]
Copyright date 2019; ©2019
Publication date 2019; 2019
Issuance monographic
Language English

Creators/Contributors

Author Farnia, Farzan
Degree supervisor Tse, David
Thesis advisor Tse, David
Thesis advisor Boyd, Stephen P
Thesis advisor Montanari, Andrea
Thesis advisor Weissman, Tsachy
Degree committee member Boyd, Stephen P
Degree committee member Montanari, Andrea
Degree committee member Weissman, Tsachy
Associated with Stanford University, Department of Electrical Engineering.

Subjects

Genre Theses
Genre Text

Bibliographic information

Statement of responsibility Farzan Farnia.
Note Submitted to the Department of Electrical Engineering.
Thesis Thesis Ph.D. Stanford University 2019.
Location electronic resource

Access conditions

Copyright
© 2019 by Farzan Farnia
License
This work is licensed under a Creative Commons Attribution 3.0 Unported license (CC BY).

Also listed in

Loading usage metrics...