Monday, May 21 |
07:30 - 08:45 |
Breakfast (Restaurant at your assigned hotel) |
08:45 - 09:00 |
Introduction and Welcome (Conference Room San Felipe) |
09:00 - 09:45 |
Xiaoming Huo: Computationally and Statistically Efficient distributed Inference with Theoretical Guarantees ↓ In many contemporary data-analysis settings, it is expensive and/or infeasible to assume that the entire data set is available at a central location. In recent works of computational mathematics and machine learning, great strides have been made in distributed optimization and distributed learning (i.e., machine learning). On the other hand, classical statistical methodology, theory, and computation are typically based on the assumption that the entire data are available at a central location - this is a significant shortcoming in classical statistical knowledge. The statistical methodology and theory for distributed inference have been actively developed. This talk will discuss one distributed statistical method that is computationally efficient, requiring minimal communication, and have comparable statistical properties. Theoretical guarantees of this distributed statistical estimator is presented. (Conference Room San Felipe) |
09:45 - 10:30 |
Stanislav Minsker: Distributed Statistical Estimation and Rates of Convergence in Normal Approximation ↓ In this talk, we will present algorithms for distributed statistical estimation that can take advantage of the divide-and-conquer approach.
We show that one of the key benefits attained by an appropriate divide-and-conquer strategy is robustness, an important characteristic of large distributed systems.
Moreover, we introduce a class of algorithms that are based on the properties of the spatial median, establish connections between performance of these distributed algorithms and rates of convergence in normal approximation, and provide tight deviations guarantees for resulting estimators in the form of exponential concentration inequalities.
Techniques are illustrated with several examples; in particular, we obtain new results for the median-of-means estimator, as well as provide performance guarantees for robust distributed maximum likelihood estimation.
The talk is based on a joint work with Nate Strawn. (Conference Room San Felipe) |
10:30 - 11:00 |
Coffee Break (Conference Room San Felipe) |
11:00 - 11:45 |
Peter Song: Meta Estimation of Normal Mean Parameter: Seven Perspectives of Data Integration ↓ Data integration has recently drawn considerable attention in the statistical literature. At this talk we will present a synergic treatment on the estimation of mean parameter of a normal distribution
from seven different schools of statistics, which sheds light on the future development of data integration analytics. They include best linear unbiased estimation (BLUE), maximum likelihood estimation (MLE), Bayesian estimation, empirical Bayesian estimation (EBE), Fisher's fiducial estimation, generalized methods of moments (GMM) estimation, and empirical likelihood estimation (ELE). Their properties of scalability and distributed inference will be discussed and compared analytically and numerically. (Conference Room San Felipe) |
11:45 - 12:30 |
Ding-Xuan Zhou: Theory of Deep Convolutional Neural Networks and Distributed Learning ↓ Deep learning has been widely applied and brought breakthroughs in speech recognition,
computer vision, and many other domains. The involved deep neural network architectures and
computational issues have been well studied in machine learning. But there lacks a theoreti-
cal foundation for understanding the approximation or generalization ability of deep learning
methods with network architectures such as deep convolutional neural networks with convo-
lutional structures. This talk describes a mathematical theory of deep convolutional neural
networks (CNNs). In particular, we show the universality of a deep CNN, meaning that it can
be used to approximate any continuous function to an arbitrary accuracy when the depth of
the neural network is large enough. Our quantitative estimate, given tightly in terms of the
number of free parameters to be computed, verifies the efficiency of deep CNNs in dealing with
large dimensional data. Some related distributed learning algorithms will also be discussed. (Conference Room San Felipe) |
12:30 - 13:15 |
Bochao Jia: Double-Parallel Monte Carlo for Bayesian Analysis of Big Data ↓ This paper proposes a simple, practical and efficient MCMC algorithm for Bayesian analysis of big data. The proposed algorithm suggests to divide the big dataset into some smaller subsets and provides a simple method to aggregate the subset posteriors to approximate the full data posterior. To further speed up computation, the proposed algorithm employs the population stochastic approximation Monte Carlo (Pop-SAMC) algorithm, a parallel MCMC algorithm, to simulate from each subset posterior. Since this algorithm consists of two levels of parallel, data parallel and simulation parallel, it is coined as “Double Parallel Monte Carlo”. The validity of the proposed algorithm is justified mathematically and numerically. (Conference Room San Felipe) |
13:20 - 13:30 |
Group Photo (Hotel Hacienda Los Laureles) |
13:30 - 15:00 |
Lunch (Restaurant Hotel Hacienda Los Laureles) |
15:15 - 16:00 |
Jin Zhou: Variance Component Testing and Selection for a Longitudinal Microbiome Study ↓ High-throughput sequencing technology has enabled population-based studies of the role of the human microbiome in disease etiology and exposure response. Due to the high cost of sequencing technology such studies usually have limited sample sizes. We study the association of microbiome composition and clinical phenotypes by testing the nullity of variance components. When the null model has more than one variance parameters and sample sizes are limited, such as in longitudinal metagenomics studies, testing zero variance components remains an open challenge. In this talk, I first introduce a series of efficient exact tests (score test, likelihood ratio test, and restricted likelihood ratio test) of testing zero variance components in presence of multiple variance components. Our approach does not rely on the asymptotic theory thus significantly boosts the power in small samples. Furthermore, to further conquer limited sample size and high dimensional features of metagenomics data, we introduce a variance component selection scheme with lasso penalization. We propose an minorization-maximization (MM) algorithm for the difficult optimization problem. Extensive simulations demonstrate the superiority of our methods vs existing methods. Finally, we apply our method to a longitudinal microbiome study of HIV infected patients (Conference Room San Felipe) |
16:00 - 16:30 |
Coffee Break (Conference Room San Felipe) |
16:30 - 17:15 |
working group time (Conference Room San Felipe) |
17:15 - 19:00 |
working group time (Conference Room San Felipe) |
19:00 - 21:00 |
Dinner (Restaurant Hotel Hacienda Los Laureles) |