Page 284 - A First Course In Stochastic Models
P. 284

REFERENCES                           277

                Hordijk, A. and Kallenberg, L.C.M. (1979) Linear programming and Markov decision
                  chains. Management Sci., 25, 352–362.
                Hordijk, A. and Kallenberg, L.C.M. (1984) Constrained undiscounted stochastic dynamic
                  programming. Math. Operat. Res., 9, 276–289.
                Howard, R.A. (1960) Dynamic Programming and Markov Processes. John Wiley & Sons,
                  Inc., New York.
                Kawai, H. (1983) An optimal ordering and replacement policy of a Markovian degradation
                  system under complete observation, part I. J. Operat. Res. Soc. Japan, 26, 279–290.
                Kolderman, J. and Volgenant, A. (1985) Optimal claiming in an automobile insurance system
                  with bonus-malus structure. J. Operat. Res. Soc., 36, 239–247.
                MacQueen, J. (1966) A modified dynamic programming method for Markovian decision
                  problems. J. Math. Appl. Math., 14, 38–43.
                Manne, A. (1960) Linear programming and sequential decisions. Management Sci., 6,
                  259–267.
                Norman, J.M. and Shearn, D.C.S. (1980) Optimal claiming on vehicle insurance revisited.
                  J. Operat. Res. Soc., 31, 181–186.
                Odoni, A. (1969) On finding the maximal gain for Markov decision processes. Operat. Res.,
                  17, 857–860.
                Popyack, J.L., Brown, R.L. and White, C.C. III (1979) Discrete versions of an algorithm
                  due to Varaiya. IEEE Trans. Automat. Contr., 24, 503–504.
                Puterman, M.L. (1994) Markov Decision Processes: Discrete Stochastic Dynamic Program-
                  ming. John Wiley & Sons, Inc., New York.
                Schweitzer, P.J. and Federgruen, A. (1979) Geometric convergence of value iteration in
                  multichain Markov decision problems. Adv. Appl. Prob., 11, 188–217.
                Sennott, L.I. (1999) Stochastic Dynamic Programming and the Control of Queueing Systems.
                  John Wiley & Sons, Inc., New York.
                Stengos, D. and Thomas, L.C. (1980) The blast furnaces problem. Eur. J. Operat. Res., 4,
                  330–336.
                Su, Y. and Deininger, R. (1972) Generalization of White’s method of successive approxi-
                  mations to periodic Markovian decision processes. Operat. Res., 20, 318–326.
                Tijms, H.C. and Van der Duyn Schouten, F.A. (1985) A Markov decision algorithm for
                  optimal inspections and revisions in a maintenance system with partial information. Eur.
                  J. Operat. Res., 21, 245–253.
                Van der Wal, J. (1980) The method of value oriented successive approximations for the
                  average reward Markov decision process. OR Spektrum, 1, 233–242.
                Veinott, A.F. Jr (1966) On finding optimal policies in discrete dynamic programming with
                  no discounting. Ann. Math. Statist., 37, 1284–1294.
                Wagner, H.M. (1975) Principles of Operations Research, 2nd edn. Prentice Hall, Englewood
                  Cliffs NJ.
                White, D.J. (1963) Dynamic programming, Markov chains and the method of successive
                  approximations. J. Math. Anal. Appl., 6, 373–376.
                White, D.J. (1985) Real applications of Markov decision processes. Interfaces, 15, no. 6,
                  73–78.
   279   280   281   282   283   284   285   286   287   288   289