\(\DeclarePairedDelimiterX{\Set}[2]{\{}{\}}{#1 \nonscript\;\delimsize\vert\nonscript\; #2}\) \( \DeclarePairedDelimiter{\set}{\{}{\}}\) \( \DeclarePairedDelimiter{\parens}{\left(}{\right)}\) \(\DeclarePairedDelimiterX{\innerproduct}[1]{\langle}{\rangle}{#1}\) \(\newcommand{\ip}[1]{\innerproduct{#1}}\) \(\newcommand{\bmat}[1]{\left[\hspace{2.0pt}\begin{matrix}#1\end{matrix}\hspace{2.0pt}\right]}\) \(\newcommand{\barray}[1]{\left[\hspace{2.0pt}\begin{matrix}#1\end{matrix}\hspace{2.0pt}\right]}\) \(\newcommand{\mat}[1]{\begin{matrix}#1\end{matrix}}\) \(\newcommand{\pmat}[1]{\begin{pmatrix}#1\end{pmatrix}}\) \(\newcommand{\mathword}[1]{\mathop{\textup{#1}}}\)
Needs:
Dynamic Optimization Problems
Real-Valued Random Variable Expectation
Needed by:
None.
Links:
Sheet PDF
Graph PDF

Stochastic Dynamical Systems

Why

We want to model uncertain outcomes in dynamical systems.1

Definition

Let $\mathcal{X} _0, \mathcal{X} _1, \dots , \mathcal{X} _{T}$ and $\mathcal{U} _0, \mathcal{U} _1, \dots , \mathcal{U} _{T-1}$ be sets. Let $(\Omega , \mathcal{A} , \mathbfsf{P} )$ be a probability space. Let $\mathcal{W} _{0}, \dots , \mathcal{W} _{T}$. Let $w_{t}: \Omega \to \mathcal{W} _{t}$ for $t = 0, \dots , T$ be random variables. For $t = 0$, $\dots $, $T-1$, let $f_{t}: \mathcal{X} _t \times \mathcal{U} _t \times \mathcal{W} _t \to \mathcal{X} _{t+1}$.

We call the sequence

\[ \mathcal{D} = ((\mathcal{X} _t)_{t = 0}^{T}), (\mathcal{U} _t)_{t=0}^{T-1}, (w_t)_{t=0}^{T-1}, (f_t)_{t=1}^{T-1}) \]

a stochastic discrete-time dynamical system. We call $w_t$ the noise variables.

Problem

Let $x_0: \Omega \to \mathcal{X} _0$ be a random variable. Define $x_1: \Omega \to \mathcal{X} _1$, $\dots $, $x_T: \Omega \to \mathcal{X} _t$ by

\[ x_{t+1} = f_t(x_t, u_t, w_t) \]

for $t = 0, \dots , T-1$. Roughly speaking, the state transition functions are nondeterministic. In other words, it is uncertain which state we will arrive in given our current state and action. The choice $u_t$ only determines the distribution of $x_{t+1}$. Here $x_0$ is (still) called the initial state and is a random variable, usually assumed independent of the $w_t$.

Let $g_t: \mathcal{X} _t \times \mathcal{U} _t \times \mathcal{W} _t \to \R \cup \set{\infty}$ for $t = 0, \dots , T-1$ and $g_{T}: \mathcal{X} _T \times \mathcal{W} _T \to \R \cup \set{\infty}$. We call $(x_0, \mathcal{D} , (g_t)_{t = 0}^{T})$ a stochastic dynamic optimization problem. As with dynamic optimization problems, we call $g_t$ the stage cost function and $g_T$ the terminal cost function. It is common for these to not depend on $w_T$ (in other words, to be deterministic). It is also common for these to take infinite values to encode constraints.

As before, a stochastic dynamic optimization problem is just an optimization problem. Define $U = \mathcal{U} _0 \times \mathcal{U} _1 \times \cdots \times \mathcal{U} _{T-1}$ and let $u \in U$. Define $C: \Omega \to \R $ by

\[ C = \sum_{t = 0}^{T-1} g_t(x_t, u_t, w_t) + g_T(x_T, w_T). \]

We call $C$ the total cost for actions $u$. It is a random variable.

Define $J: U \to \R \cup \set{\infty}$ by

\[ J = \E ( \sum_{t = 0}^{T-1} g_t(x_t, u_t, w_t) + g_T(x_T, w_T) ). \]

$J(u)$ is the expected total cost for inputs $u$.

The optimization problem is $(U, J)$. In other words, the objective is the mean total stage cost plus the terminal cost.

Other terminology

Stochastic dynamic optimization problems are frequently called stochastic control problems.


  1. Future editions will expand. ↩︎
Copyright © 2023 The Bourbaki Authors — All rights reserved — Version 13a6779cc About Show the old page view