Optimal control is a branch of optimization that aims to find a control (a series of actions) in a dynamic system to optimize some function. For example, a common problem is
More generally, optimal control is defined with differential equations,
Note that this field is closely related to โ๏ธ Reinforcement Learning (where
Dynamic System
A dynamic system is defined by state transitions, which, in the context of reinforcement learning, can be deterministic or stochastic.
- Deterministic transitions follow
- Stochastic transitions follow
Moreover, the system may be open-loop or close-loop.
- Open-loop systems offer no feedback or new informationโweโre expected to output a sequence of actions
given just the initial state . Algorithms that deal with this case include ๐ฒ Cross Entropy Method and the Linear-Quadratic Regulator (LQR). Note that open loop control is often only feasible with deterministic transitions. - Closed-loop systems provide new information
after we take an action . In this case, we often derive a policy , falling closer into the realm of reinforcement learning.