In mathematics, the theory of optimal stopping or early stopping is concerned with the problem of choosing a time to take a particular action, in order to maximise an expected reward or minimise an expected cost. Optimal stopping problems can be found in areas of statistics, economics, and mathematical finance (related to the pricing of American options). A key example of an optimal stopping problem is the secretary problem. Optimal stopping problems can often be written in the form of a Bellman equation, and are therefore often solved using dynamic programming.
Stopping rule problems are associated with two objects:
A sequence of random variables , whose joint distribution is something assumed to be known
A sequence of 'reward' functions which depend on the observed values of the random variables in 1:
Given those objects, the problem is as follows:
You are observing the sequence of random variables, and at each step , you can choose to either stop observing or continue
If you stop observing at step , you will receive reward
You want to choose a stopping rule to maximize your expected reward (or equivalently, minimize your expected loss)
Consider a gain process defined on a filtered probability space and assume that is adapted to the filtration. The optimal stopping problem is to find the stopping time which maximizes the expected gain
where is called the value function. Here can take value .
A more specific formulation is as follows. We consider an adapted strong Markov process defined on a filtered probability space where denotes the probability measure where the stochastic process starts at . Given continuous functions , and , the optimal stopping problem is
This is sometimes called the MLS (which stand for Mayer, Lagrange, and supremum, respectively) formulation.
There are generally two approaches to solving optimal stopping problems. When the underlying process (or the gain process) is described by its unconditional finite-dimensional distributions, the appropriate solution technique is the martingale approach, so called because it uses martingale theory, the most important concept being the Snell envelope.
This page is automatically generated and may contain information that is not correct, complete, up-to-date, or relevant to your search query. The same applies to every other page on this website. Please make sure to verify the information with EPFL's official sources.
In probability theory, in particular in the study of stochastic processes, a stopping time (also Markov time, Markov moment, optional stopping time or optional time) is a specific type of “random time”: a random variable whose value is interpreted as the time at which a given stochastic process exhibits a certain behavior of interest. A stopping time is often defined by a stopping rule, a mechanism for deciding whether to continue or stop a process on the basis of the present position and past events, and which will almost always lead to a decision to stop at some finite time.
In microeconomics, search theory studies buyers or sellers who cannot instantly find a trading partner, and must therefore search for a partner prior to transacting. It involves determining the best approach to use when looking for a specific item or person in a sizable, uncharted environment. The goal of the theory is to determine the best search strategy, one that maximises the chance of finding the target while minimising search-related expenses. Search theory clarifies how buyers and sellers choose when to acknowledge a coordinating offer for a transaction.
A Bellman equation, named after Richard E. Bellman, is a necessary condition for optimality associated with the mathematical optimization method known as dynamic programming. It writes the "value" of a decision problem at a certain point in time in terms of the payoff from some initial choices and the "value" of the remaining decision problem that results from those initial choices. This breaks a dynamic optimization problem into a sequence of simpler subproblems, as Bellman's “principle of optimality" prescribes.
This course focuses on dynamic models of random phenomena, and in particular, the most popular classes of such models: Markov chains and Markov decision processes. We will also study applications in q
This doctoral course provides an introduction to optimal control covering fundamental theory, numerical implementation and problem formulation for applications.
Introduction à la théorie des martingales à temps discret, en particulier aux théorèmes de convergence et d'arrêt. Application aux processus de branchement. Introduction au mouvement brownien et étude
We present outlier-free isogeometric Galerkin discretizations of eigenvalue problems related to the biharmonic and the polyharmonic operator in the univariate setting. These are Galerkin discretizations in certain spline subspaces that provide accurate app ...
Lausanne2023
, , ,
This paper studies the problem of online performance optimization of constrained closed-loop control systems, where both the objective and the constraints are unknown black-box functions affected by exogenous time-varying contextual disturbances. A primal- ...
New York2023
, ,
A critical operational challenge in Mobility-on-demand systems is the problem of imbalance between vehicle supply and passenger demand. However, conventional model-based methods require accurate parametric system models with complex nonlinear dynamics that ...