论文标题
张量结构的素描,以限制最小二乘
Tensor-structured sketching for constrained least squares
论文作者
论文摘要
在许多应用中,限制了最小二乘问题。在实践中,他们的记忆和计算成本涉及高维输入数据的昂贵。我们采用所谓的“草图”策略,通过随机素描矩阵将最小二乘问题的问题投射到一个较低的“草图维度”的空间上。素描的关键思想是在保持近似准确性的同时,尽可能地减少问题的维度。 张量结构通常存在于最小二乘的数据矩阵中,包括线性化的反问题和张量分解。在这项工作中,我们利用了一类逐行的张力量子矩阵作为素描矩阵的素描矩阵,以素描设计的限制优化,以素描设计与张量结构的兼容性。我们就误差标准和概率故障率方面提供了草图维度的理论保证。在无约束的线性回归的背景下,我们获得了草图维度的最佳估计。对于具有一般约束集的优化问题,我们表明草图维度取决于表征基础问题几何形状的统计复杂性。我们的理论在一些具体示例中得到了证明,包括无约束的线性回归和稀疏的恢复问题。
Constrained least squares problems arise in many applications. Their memory and computation costs are expensive in practice involving high-dimensional input data. We employ the so-called "sketching" strategy to project the least squares problem onto a space of a much lower "sketching dimension" via a random sketching matrix. The key idea of sketching is to reduce the dimension of the problem as much as possible while maintaining the approximation accuracy. Tensor structure is often present in the data matrices of least squares, including linearized inverse problems and tensor decompositions. In this work, we utilize a general class of row-wise tensorized sub-Gaussian matrices as sketching matrices in constrained optimizations for the sketching design's compatibility with tensor structures. We provide theoretical guarantees on the sketching dimension in terms of error criterion and probability failure rate. In the context of unconstrained linear regressions, we obtain an optimal estimate for the sketching dimension. For optimization problems with general constraint sets, we show that the sketching dimension depends on a statistical complexity that characterizes the geometry of the underlying problems. Our theories are demonstrated in a few concrete examples, including unconstrained linear regression and sparse recovery problems.