Papers
Topics
Authors
Recent
Search
2000 character limit reached

Bivariate Matrix-valued Linear Regression (BMLR): Finite-sample performance under Identifiability and Sparsity Assumptions

Published 23 Dec 2024 in math.ST, stat.ML, and stat.TH | (2412.17749v2)

Abstract: This study explores the estimation of parameters in a matrix-valued linear regression model, where the $T$ responses $(Y_t){t=1}T \in \mathbb{R}{n \times p}$ and predictors $(X_t){t=1}T \in \mathbb{R}{m \times q}$ satisfy the relationship $Y_t = A* X_t B* + E_t$ for all $t = 1, \ldots, T$. In this model, $A* \in \mathbb{R}+{n \times m}$ has $L_1$-normalized rows, $B* \in \mathbb{R}{q \times p}$, and $(E_t){t=1}T$ are independent noise matrices following a matrix Gaussian distribution. The primary objective is to estimate the unknown parameters $A*$ and $B*$ efficiently. We propose explicit optimization-free estimators and establish non-asymptotic convergence rates to quantify their performance. Additionally, we extend our analysis to scenarios where $A*$ and $B*$ exhibit sparse structures. To support our theoretical findings, we conduct numerical simulations that confirm the behavior of the estimators, particularly with respect to the impact of the dimensions $n, m, p, q$, and the sample size $T$ on finite-sample performances. We complete the simulations by investigating the denoising performances of our estimators on noisy real-world images.

Authors (1)

Summary

No one has generated a summary of this paper yet.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 1 tweet with 0 likes about this paper.