Dynamic Programming (with Management Applications) |
Contents
THE VALUE ITERATION ALGORITHM | 1 |
DETERMINISTIC APPLICATIONS | 43 |
FINITE STAGE MARKOV PROGRAMMING | 82 |
Copyright | |
3 other sections not shown
Other editions - View all
Common terms and phrases
action k actions and values bias quantity boats computational corresponds current policy denoted Determine deterministic discount factor discounted returns dynamic programming Figure function fundamental matrix gain optimal policies gain rate given linear simultaneous equations marketing example Markov decision problem Markov process Max r(n maximise Mean return mean total return minimises minmax month number of stages optimal action optimal path optimal plan optimal process optimal value optimality condition optimisation overhaul PARTITIONING PROBLEM plan is followed planning horizon policy evaluation operation policy improvement routine Prob random variable raw material recurrence relation relative bias values return associated return r(i semi-Markov semi-Markov process separability condition shortest path problem shown in Table Slotting Machine Example solution stage return Stage State Action stochastic stochastic matrix stock level suboptimality test summarised terminal values test quantity transition equation transition probability matrix transition return units value iteration algorithm value iteration method value table zero