summaryrefslogtreecommitdiff
diff options
context:
space:
mode:
authorJaron Kent-Dobias <jaron@kent-dobias.com>2025-03-11 10:54:52 -0300
committerJaron Kent-Dobias <jaron@kent-dobias.com>2025-03-11 10:54:52 -0300
commit155a3fd57c53268b44ca2a22763b35992ef1f821 (patch)
tree1a8f91914b2d2f6671966cc0a32e8feb8e864d6f
parentb0d446e02af5645d4979b56d44f1b84fb7e7daf0 (diff)
downloadSciPostPhys_18_158-155a3fd57c53268b44ca2a22763b35992ef1f821.tar.gz
SciPostPhys_18_158-155a3fd57c53268b44ca2a22763b35992ef1f821.tar.bz2
SciPostPhys_18_158-155a3fd57c53268b44ca2a22763b35992ef1f821.zip
More bibliography tweaks
-rw-r--r--topology.bib8
-rw-r--r--topology.tex8
2 files changed, 8 insertions, 8 deletions
diff --git a/topology.bib b/topology.bib
index 74cb7e0..bf5b2e8 100644
--- a/topology.bib
+++ b/topology.bib
@@ -675,19 +675,19 @@
eprinttype = {arxiv}
}
-@inproceedings{Mannelli_2019_Who,
+@inproceedings{SaraoMannelli_2019_Who,
author = {Sarao Mannelli, Stefano and Biroli, Giulio and Cammarota, Chiara and Krzakala, Florent and Zdeborová, Lenka},
- title = {Who is Afraid of Big Bad Minima? Analysis of gradient-flow in spiked matrix-tensor models},
+ title = {Who is Afraid of Big Bad Minima? {A}nalysis of gradient-flow in spiked matrix-tensor models},
publisher = {Curran Associates, Inc.},
year = {2019},
volume = {32},
pages = {},
url = {https://proceedings.neurips.cc/paper_files/paper/2019/file/fbad540b2f3b5638a9be9aa6a4d8e450-Paper.pdf},
booktitle = {Advances in Neural Information Processing Systems},
- editor = {Wallach, H. and Larochelle, H. and Beygelzimer, A. and Alché-Buc, F. d' and Fox, E. and Garnett, R.}
+ editor = {Wallach, H. and Larochelle, H. and Beygelzimer, A. and d'Alché-Buc, F. and Fox, E. and Garnett, R.}
}
-@inproceedings{Mannelli_2019_Passed,
+@inproceedings{SaraoMannelli_2019_Passed,
author = {Sarao Mannelli, Stefano and Krzakala, Florent and Urbani, Pierfrancesco and Zdeborová, Lenka},
title = {Passed \& Spurious: Descent Algorithms and Local Minima in Spiked Matrix-Tensor Models},
publisher = {PMLR},
diff --git a/topology.tex b/topology.tex
index 0edd478..fcf99c7 100644
--- a/topology.tex
+++ b/topology.tex
@@ -793,20 +793,20 @@ Finally, a common extension of the spherical spin glasses is to add a
deterministic piece to the energy, sometimes called a signal or a spike. Recent
work argued that gradient descent can avoid being trapped in marginal minima
and reach the vicinity of the signal if the set of trapping marginal minima has
-been destabilized by the presence of the signal \cite{Mannelli_2019_Passed,
-Mannelli_2019_Who}. The authors of Ref.~\cite{Mannelli_2019_Who}
+been destabilized by the presence of the signal \cite{SaraoMannelli_2019_Passed,
+SaraoMannelli_2019_Who}. The authors of Ref.~\cite{SaraoMannelli_2019_Who}
conjecture based on \textsc{dmft} data for $2+3$ mixed spherical spin glasses that the
trapping marginal minima are those at the traditional threshold energy
$E_\text{th}$. However, Ref.~\cite{Folena_2023_On} demonstrated that in mixed
$p+s$ spherical spin glasses with small $p$ and $s$, the difference between
$E_\text{th}$ and the true trapping energy is difficult to resolve with the
current precision of \textsc{dmft} integration schemes. Therefore,
-the authors of Ref.~\cite{Mannelli_2019_Who} may have incorrectly conflated the
+the authors of Ref.~\cite{SaraoMannelli_2019_Who} may have incorrectly conflated the
threshold with the trapping marginal minima, and that the correct set of
marginal minima that must be destabilized to reach a signal might be the same set
that trap dynamics in the signal-free model. This paper conjectures that the important trapping minima
are those at the shattering energy. Comparing the predictions of
-Ref.~\cite{Mannelli_2019_Who} to \textsc{dmft} simulations of a model with
+Ref.~\cite{SaraoMannelli_2019_Who} to \textsc{dmft} simulations of a model with
better separation between $p$ and $s$ would help resolve this issue.
\section{Conclusion}