Bellman Value Iteration (VI) (Optimal Policies for MDPs Optimal Policies for MDPs)

From Algorithm Wiki
Revision as of 11:53, 15 February 2023 by Admin (talk | contribs) (Created page with "== Time Complexity == $O({2}^n)$ == Space Complexity == $O(n)$ words (Only needs to store values (V) and policy (pi), both size O(n)) == Description == == Approximate? == Exact == Randomized? == No, deterministic == Model of Computation == Word/Real RAM == Year == 1957 == Reference == https://www.jstor.org/stable/24900506")
(diff) ← Older revision | Latest revision (diff) | Newer revision → (diff)
Jump to navigation Jump to search

Time Complexity

$O({2}^n)$

Space Complexity

$O(n)$ words

(Only needs to store values (V) and policy (pi), both size O(n))

Description

Approximate?

Exact

Randomized?

No, deterministic

Model of Computation

Word/Real RAM

Year

1957

Reference

https://www.jstor.org/stable/24900506