When.com Web Search

Search results

  1. Results From The WOW.Com Content Network
  2. Bellman equation - Wikipedia

    en.wikipedia.org/wiki/Bellman_equation

    The Bellman equation was first applied to engineering control theory and to other topics in applied mathematics, and subsequently became an important tool in economic theory; though the basic concepts of dynamic programming are prefigured in John von Neumann and Oskar Morgenstern's Theory of Games and Economic Behavior and Abraham Wald's ...

  3. Dynamic programming - Wikipedia

    en.wikipedia.org/wiki/Dynamic_programming

    An accessible introduction to dynamic programming in economics. MATLAB code for the book Archived 2020-10-09 at the Wayback Machine. Bellman, Richard (1954), "The theory of dynamic programming", Bulletin of the American Mathematical Society, 60 (6): 503– 516, doi: 10.1090/S0002-9904-1954-09848-8, MR 0067459. Includes an extensive bibliography ...

  4. Hamilton–Jacobi–Bellman equation - Wikipedia

    en.wikipedia.org/wiki/Hamilton–Jacobi–Bellman...

    Its solution is the value function of the optimal control problem which, once known, can be used to obtain the optimal control by taking the maximizer (or minimizer) of the Hamiltonian involved in the HJB equation. [2] [3] The equation is a result of the theory of dynamic programming which was pioneered in the 1950s by Richard Bellman and ...

  5. Control theory - Wikipedia

    en.wikipedia.org/wiki/Control_theory

    Control theory is a field of control engineering and applied mathematics that deals with the control of dynamical ... Richard Bellman developed dynamic programming in ...

  6. Optimal control - Wikipedia

    en.wikipedia.org/wiki/Optimal_control

    Optimal control problem benchmark (Luus) with an integral objective, inequality, and differential constraint. Optimal control theory is a branch of control theory that deals with finding a control for a dynamical system over a period of time such that an objective function is optimized. [1]

  7. Pontryagin's maximum principle - Wikipedia

    en.wikipedia.org/wiki/Pontryagin's_maximum_Principle

    Widely regarded as a milestone in optimal control theory, the significance of the maximum principle lies in the fact that maximizing the Hamiltonian is much easier than the original infinite-dimensional control problem; rather than maximizing over a function space, the problem is converted to a pointwise optimization. [8]

  8. Stochastic dynamic programming - Wikipedia

    en.wikipedia.org/wiki/Stochastic_dynamic_programming

    Originally introduced by Richard E. Bellman in (Bellman 1957), stochastic dynamic programming is a technique for modelling and solving problems of decision making under uncertainty. Closely related to stochastic programming and dynamic programming , stochastic dynamic programming represents the problem under scrutiny in the form of a Bellman ...

  9. Recursive economics - Wikipedia

    en.wikipedia.org/wiki/Recursive_economics

    The recursive paradigm originated in control theory with the invention of dynamic programming by the American mathematician Richard E. Bellman in the 1950s. Bellman described possible applications of the method in a variety of fields, including Economics, in the introduction to his 1957 book. [1]