`

Timezone: »

 
Poster
Efficient Truncated Linear Regression with Unknown Noise Variance
Constantinos Daskalakis · Patroklos Stefanou · Rui Yao · Emmanouil Zampetakis

Tue Dec 07 04:30 PM -- 06:00 PM (PST) @ None #None
Truncated linear regression is a classical challenge in Statistics, wherein a label, $y = w^T x + \varepsilon$, and its corresponding feature vector, $x \in \mathbb{R}^k$, are only observed if the label falls in some subset $S \subseteq \mathbb{R}$; otherwise the existence of the pair $(x, y)$ is hidden from observation. Linear regression with truncated observations has remained a challenge, in its general form, since the early works of [Tobin'58, Amemiya '73]. When the distribution of the error is normal with known variance, recent work of [Daskalakis et al. '19] provides computationally and statistically efficient estimators of the linear model, $w$. In this paper, we provide the first computationally and statistically efficient estimators for truncated linear regression when the noise variance is unknown, estimating both the linear model and the variance of the noise. Our estimator is based on an efficient implementation of Projected Stochastic Gradient Descent on the negative log-likelihood of the truncated sample. Importantly, we show that the error of our estimates is asymptotically normal, and we use this to provide explicit confidence regions for our estimates.

Author Information

Constantinos Daskalakis (MIT)
Patroklos Stefanou (Massachusetts Institute of Technology)
Rui Yao (Massachusetts Institute of Technology)
Emmanouil Zampetakis (University of California Berkeley)

More from the Same Authors