Différences
Ci-dessous, les différences entre deux révisions de la page.
| Les deux révisions précédentes Révision précédente | |||
| mega:seminaire [2026/05/06 12:01] – Raphaël BUTEZ | mega:seminaire [2026/05/06 14:27] (Version actuelle) – Raphaël BUTEZ | ||
|---|---|---|---|
| Ligne 30: | Ligne 30: | ||
| Abstract: This work presents a comprehensive understanding of the estimation of a planted low-rank signal from a general spiked tensor model near the computational threshold. Relying on standard tools from the theory of large random matrices, we characterize the large-dimensional spectral behavior of the unfoldings of the data tensor and exhibit relevant signal-to-noise ratios governing the detectability of the principal directions of the signal. These results allow to accurately predict the reconstruction performance of truncated multilinear SVD (MLSVD) in the non-trivial regime. This is particularly important since it serves as an initialization of the higher-order orthogonal iteration (HOOI) scheme, whose convergence to the best low-multilinear-rank approximation depends entirely on its initialization. We give a sufficient condition for the convergence of HOOI and show that the number of iterations before convergence tends to 1 in the large-dimensional limit. | Abstract: This work presents a comprehensive understanding of the estimation of a planted low-rank signal from a general spiked tensor model near the computational threshold. Relying on standard tools from the theory of large random matrices, we characterize the large-dimensional spectral behavior of the unfoldings of the data tensor and exhibit relevant signal-to-noise ratios governing the detectability of the principal directions of the signal. These results allow to accurately predict the reconstruction performance of truncated multilinear SVD (MLSVD) in the non-trivial regime. This is particularly important since it serves as an initialization of the higher-order orthogonal iteration (HOOI) scheme, whose convergence to the best low-multilinear-rank approximation depends entirely on its initialization. We give a sufficient condition for the convergence of HOOI and show that the number of iterations before convergence tends to 1 in the large-dimensional limit. | ||
| - | * 16h00-16h30: | + | * 16h00-16h30: |
| - | * 16h30-17h00: | + | Abstract: In this work, we aim to characterize the effect that modern data augmentation schemes have on the generalization error of large deep learning models. |
| + | * 16h30-17h00: | ||
| + | Abstract: | ||
| + | Understanding how information propagates in very deep neural networks is essential for designing architectures that remain trainable as depth increases. In our previous work, A New Initialisation to Control Gradients in Sinusoidal Neural Networks (ICLR 2026), we introduced an initialisation scheme for SIREN networks that controls both gradient variance and the Fourier spectrum of the network output. This led us to study the large-depth limit of neural networks more generally and to address the central question of this work: how do correlations, | ||
| + | To answer this question, we develop a theoretical framework in the thermodynamic sequential limit, corresponding to the infinite-width mean-field regime. This framework provides a unified description of these propagation mechanisms by combining mean-field analysis with tools from free probability, | ||
| + | This framework allowed us to identify an initialisation strategy based on orthogonal weights that applies to a broad class of activation functions. At this initialisation, | ||
| + | |||
| + | Reference: | ||
| + | [1] J. Pennington, S. Schoenholz, and S. Ganguli. “Resurrecting the Sigmoid in Deep Learning through Dynamical Isometry.” | ||