A copy of this work was available on the public web and has been preserved in the Wayback Machine. The capture dates from 2020; you can also visit the original URL.
The file type is application/pdf
.
Filters
Learning Rate Annealing Can Provably Help Generalization, Even for Convex Problems
[article]
2020
arXiv
pre-print
We give a toy convex problem where learning rate annealing (large initial learning rate, followed by small learning rate) can lead gradient descent to minima with provably better generalization than using ...
In this note, we show that this phenomenon can exist even for convex learning problems -- in particular, linear regression in 2 dimensions. ...
Acknowledgements We thank John Schulman for a discussion around learning rates that led to wondering if this can occur in convex problems. ...
arXiv:2005.07360v1
fatcat:plkqqrko6rerjerfavquoqkfpu
Prospects and challenges of quantum finance
[article]
2020
arXiv
pre-print
as quantum annealing heuristics for portfolio optimization. ...
We consider quantum speedups for Monte Carlo methods, portfolio optimization, and machine learning. ...
Acknowledgments We thank Kay Giesecke, Rajiv Krishnakumar, Ashley Montanaro, Nikitas Stamatopoulos, and Will Zeng for helpful discussions and comments on this manuscript. ...
arXiv:2011.06492v1
fatcat:mqzj2a2pzzaz5pdcllxgkr73oq
Adaptive Gradient Methods with Local Guarantees
[article]
2022
arXiv
pre-print
In this paper we study the problem of learning a local preconditioner, that can change as the data is changing along the optimization trajectory. ...
Without the need to manually tune a learning rate schedule, our method can, in a single run, achieve comparable and stable task accuracy as a fine-tuned optimizer. ...
We demonstrate the effectiveness and robustness of SAMUEL in experiments, where we show that SAMUEL can automatically adapt to the optimal learning rate and achieve comparable task accuracy as a fine-tuned ...
arXiv:2203.01400v2
fatcat:5wg5u5ctuja37lqmfibh34ybh4
Message-passing for graph-structured linear programs
2008
Proceedings of the 25th international conference on Machine learning - ICML '08
A large body of past work has focused on the first-order tree-based LP relaxation for the MAP problem in Markov random fields. ...
We establish various convergence guarantees for our algorithms, illustrate their performance, and also present rounding schemes with provable optimality guarantees. ...
We thank the anonymous reviewers for helpful comments. ...
doi:10.1145/1390156.1390257
dblp:conf/icml/RavikumarAW08
fatcat:gzalmwvudzdvvdk4kleb5gm4zm
A Survey of Quantum Computing for Finance
[article]
2022
arXiv
pre-print
learning, showing how these solutions, adapted to work on a quantum computer, can help solve more efficiently and accurately problems such as derivative pricing, risk analysis, portfolio optimization, ...
We hope this article will not only serve as a reference for academic researchers and industry practitioners but also inspire new ideas for future research. ...
In general, however, the speedup for each task can vary greatly or may even be currently unknown (Section 4). ...
arXiv:2201.02773v3
fatcat:aqcl6blbyvbljg627ot6zxtaj4
Construction of non-convex polynomial loss functions for training a binary classifier with quantum annealing
[article]
2014
arXiv
pre-print
These loss functions may also be useful for classical approaches as they compile to regularized risk expressions which can be evaluated in constant time with respect to the number of training examples. ...
To take advantage of a potential quantum advantage, one needs to be able to map the problem of interest to the native hardware with reasonably low overhead. ...
However, there is evidence that quantum resources such as tunneling and entanglement are generic computational resources which may help to solve problem instances which would be otherwise intractable for ...
arXiv:1406.4203v1
fatcat:nmri7mwwp5asrnymdmdpurdd24
Efficient Full-Matrix Adaptive Regularization
[article]
2020
arXiv
pre-print
We also provide a novel theoretical analysis for adaptive regularization in non-convex optimization settings. ...
Due to the large number of parameters of machine learning problems, full-matrix preconditioning methods are prohibitively expensive. ...
Acknowledgments We are grateful to Yoram Singer, Tomer Koren, Nadav Cohen, and Sanjeev Arora for helpful discussions. ...
arXiv:1806.02958v2
fatcat:vyzeqvt7bbedrn2tyfdjhsat5a
Generalization and Exploration via Randomized Value Functions
[article]
2016
arXiv
pre-print
generalization. ...
We propose randomized least-squares value iteration (RLSVI) -- a new reinforcement learning algorithm designed to explore and generalize efficiently via linearly parameterized value functions. ...
A recommendation engine We will now show that efficient exploration and generalization can be helpful in a simple model of customer interaction. ...
arXiv:1402.0635v3
fatcat:aoqndidaz5gnbkg65punxvf4ge
Scaling-up Distributed Processing of Data Streams for Machine Learning
[article]
2020
arXiv
pre-print
Further, it reviews guarantees underlying these methods, which show there exist regimes in which systems can learn from distributed, streaming data at order-optimal rates. ...
In particular, it focuses on methods that solve: (i) distributed stochastic convex problems, and (ii) distributed principal component analysis, which is a nonconvex problem with geometric structure that ...
(Note that some of these methods have provable convergence issues, even for convex problems [79] .) ...
arXiv:2005.08854v2
fatcat:y6fvajvq2naajeqs6lo3trrgwy
Quantum Computing at the Frontiers of Biological Sciences
[article]
2019
arXiv
pre-print
However, challenges arise as we push the limits of scale and complexity in biological problems. ...
view towards quantum computation and quantum information science, where algorithms have demonstrated potential polynomial and exponential computational speedups in certain applications, such as machine learning ...
The former may be a candidate for an exact quantum solution for small-scale problems, while both may benefit from approximate quantum annealing approaches (an annealing-based approach to NMF is found in ...
arXiv:1911.07127v1
fatcat:k2agx5yysjgi3m3ryhicptzauq
Quantum machine learning: a classical perspective
2018
Proceedings of the Royal Society A
learning problems. ...
Learning in the presence of noise and certain computationally hard problems in machine learning are identified as promising directions for the field. ...
Acknowledgements We thank Scott Aaronson, David Barber, Marcello Benedetti, Fernando Brandão, Dan Brown, Carlos González-Guillén, Joshua Lockhart, and Alessandro Rudi for helpful comments on the manuscript ...
doi:10.1098/rspa.2017.0551
pmid:29434508
pmcid:PMC5806018
fatcat:zlfvny7iyzb47di2cndbvvrglu
The sharp, the flat and the shallow: Can weakly interacting agents learn to escape bad minima?
[article]
2019
arXiv
pre-print
An open problem in machine learning is whether flat minima generalize better and how to compute such minima efficiently. This is a very challenging problem. ...
Our primary focus is on the design of algorithms for machine learning applications; however the underlying mathematical framework is suitable for the understanding of large scale systems of agent based ...
If instead β is gradually increased using a so called annealing schedule, then adding noise to the normal gradient flow can help the dynamics in (1) provably converge to a global minimum of Φ (Geman and ...
arXiv:1905.04121v1
fatcat:yq7i6o3ok5dvfbykaed5ycwr4q
Training verified learners with learned verifiers
[article]
2018
arXiv
pre-print
., networks that provably satisfy some desired input-output properties. ...
also be scaled to produce the first known (to the best of our knowledge) verifiably robust networks for CIFAR-10. ...
Instead, PVT exploits the idea that the solution of this optimization problem can be learned, i.e., the mapping from a nominal training example to the optimal dual variables can be learned by the verifier ...
arXiv:1805.10265v2
fatcat:ratq2s4kdjh3jhesxt4w7444qe
A NASA Perspective on Quantum Computing: Opportunities and Challenges
[article]
2017
arXiv
pre-print
For most problems, however, it is currently unknown whether quantum algorithms can provide an advantage, and if so by how much, or how to design quantum algorithms that realize such advantages. ...
In the last couple of decades, the world has seen several stunning instances of quantum algorithms that provably outperform the best classical algorithms. ...
that can serve as a building block for deep learning architectures. ...
arXiv:1704.04836v1
fatcat:7itanvx3mzgrnfouu5wsx33v7i
Gradient Descent, Stochastic Optimization, and Other Tales
[article]
2022
arXiv
pre-print
Its stochastic version receives attention in recent years, and this is particularly true for optimizing deep neural networks. ...
Gradient descent is one of the most popular algorithms to perform optimization and by far the most common way to optimize machine learning tasks. ...
The toy example shows learning rate annealing schemes in general can help optimization methods "find" better local minima with better performance. ...
arXiv:2205.00832v1
fatcat:unridtvvi5b2jf6xbu2chlw7ce
« Previous
Showing results 1 — 15 out of 367 results