
arXiv: 1611.05545
Stochastic gradient descent in continuous time (SGDCT) provides a computationally efficient method for the statistical learning of continuous-time models, which are widely used in science, engineering, and finance. The SGDCT algorithm follows a (noisy) descent direction along a continuous stream of data. SGDCT performs an online parameter update in continuous time, with the parameter updates $θ_t$ satisfying a stochastic differential equation. We prove that $\lim_{t \rightarrow \infty} \nabla \bar g(θ_t) = 0$ where $\bar g$ is a natural objective function for the estimation of the continuous-time dynamics. The convergence proof leverages ergodicity by using an appropriate Poisson equation to help describe the evolution of the parameters for large times. SGDCT can also be used to solve continuous-time optimization problems, such as American options. For certain continuous-time problems, SGDCT has some promising advantages compared to a traditional stochastic gradient descent algorithm. As an example application, SGDCT is combined with a deep neural network to price high-dimensional American options (up to 100 dimensions).
FOS: Computer and information sciences, Probability (math.PR), Learning and adaptive systems in artificial intelligence, deep learning, Mathematics - Statistics Theory, Machine Learning (stat.ML), Martingales with continuous parameter, Statistics Theory (math.ST), stochastic differential equations, Stochastic ordinary differential equations (aspects of stochastic analysis), Neural nets and related approaches to inference from stochastic processes, statistical learning, machine learning, Derivative securities (option pricing, hedging, etc.), Statistics - Machine Learning, Optimization and Control (math.OC), stochastic gradient descent, FOS: Mathematics, American options, Mathematics - Optimization and Control, Mathematics - Probability
FOS: Computer and information sciences, Probability (math.PR), Learning and adaptive systems in artificial intelligence, deep learning, Mathematics - Statistics Theory, Machine Learning (stat.ML), Martingales with continuous parameter, Statistics Theory (math.ST), stochastic differential equations, Stochastic ordinary differential equations (aspects of stochastic analysis), Neural nets and related approaches to inference from stochastic processes, statistical learning, machine learning, Derivative securities (option pricing, hedging, etc.), Statistics - Machine Learning, Optimization and Control (math.OC), stochastic gradient descent, FOS: Mathematics, American options, Mathematics - Optimization and Control, Mathematics - Probability
| selected citations These citations are derived from selected sources. This is an alternative to the "Influence" indicator, which also reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | 36 | |
| popularity This indicator reflects the "current" impact/attention (the "hype") of an article in the research community at large, based on the underlying citation network. | Top 10% | |
| influence This indicator reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | Top 10% | |
| impulse This indicator reflects the initial momentum of an article directly after its publication, based on the underlying citation network. | Top 10% |
