Simulation-Based Parametric Optimization Techniques and Reinforcement Learning introduces the evolving area of simulation-based optimization. Since it became possible to analyze random systems using computers, scientists and engineers have sought the means to optimize systems using simulation models. Only recently, however, has this objective had success in practice. Cutting-edge work in computational operations research, including non-linear programming (simultaneous perturbation), dynamic programming (reinforcement learning), and game theory (learning automata) has made it possible to use simulation in conjunction with optimization techniques. As a result, this research has given simulation added dimensions and power that it did not have in the recent past.The book's objective is (1) It examines the mathematical governing principles of simulation-based optimization, thereby providing the reader with the ability to model relevant real-life problems using these techniques. (2) It outlines the computational technology underlying these methods. Taken together these two aspects demonstrate that the mathematical and computational methods discussed in this book do work. Broadly speaking, the book has two (1) parametric (static) optimization and (2) control (dynamic) optimization. Some of the book's special features *An accessible introduction to reinforcement learning and parametric-optimization techniques. *A step-by-step description of several algorithms of simulation-based optimization. *A clear and simple introduction to the methodology of neural networks. *A gentle introduction to convergence analysis of some of the methods enumerated above. *Computer programs for many algorithms of simulation-based optimization.This book is written for students and researchers in the fields of engineering (electrical, industrial and computer), computer science, operations research, management science, and applied mathematics.
The main motivation for writing this book was to provide an
accessible account of methods based on
Reinforcement Learning (closely related to what is now also called
Approximate Dynamic Programming) and
Meta-Heuristics (closely related to what is now also called
Stochastic Adaptive Search) for optimization in discrete-event systems via simulation. Reinforcement Learning (RL) is typically used for solving Markov decision problems (
MDPs), which are dynamic optimization problems where the underlying discrete-event stochastic system is driven by Markov chains, while Meta-Heuristics are used for solving static optimization problems where the underlying system is any discrete-event stochastic system (not necessarily driven by Markov chains).
This book provides a selected collection of topics, mostly focused on
model-free techniques, which are useful when one does not have access to the structure of the objective function (in static optimization) or the transition probability function (in dynamic optimization). My goal was neither to overwhelm the reader with mathematical details nor was it to cover every topic. Rather, the goal was to provide the reader with an overview of the fundamental concepts and at the same time provide the details required for solving real-world stochastic optimization problems via simulation-based techniques.
Some of the main topics covered are:
- Reinforcement learning techniques, mainly rooted in Q-Learning for discounted and average reward MDPs
- Static optimization techniques rooted in meta-heuristics (simulated annealing, genetic algorithms, and tabu search) for discrete solution spaces and simultaneous perturbation for continuous solution spaces
- Neural network algorithms useful for function approximation in response surface methods for static optimization and in reinforcement learning for MDPs with large state-action spaces
- A detailed background on dynamic programming (value and policy iteration)
- A special coverage of semi-MDPs (SMDPs) and average reward problems
- A discussion on convergence of a subset of methods enumerated above