Items related to Simulation-Based Optimization: Parametric Optimization...

Simulation-Based Optimization: Parametric Optimization Techniques and Reinforcement Learning (Operations Research/Computer Science Interfaces Series) - Hardcover

  • 3.80 out of 5 stars
    5 ratings by Goodreads
 
9781402074547: Simulation-Based Optimization: Parametric Optimization Techniques and Reinforcement Learning (Operations Research/Computer Science Interfaces Series)

Synopsis

Book by Gosavi, Abhijit

"synopsis" may belong to another edition of this title.

From the Author

The main motivation for writing this book was to provide an accessible account of methods based on Reinforcement Learning (closely related to what is now also called Approximate Dynamic Programming) and Meta-Heuristics (closely related to what is now also called Stochastic Adaptive Search) for optimization in discrete-event systems via simulation.  Reinforcement Learning (RL) is typically used for solving Markov decision problems (MDPs), which are dynamic optimization problems where the underlying discrete-event stochastic system is driven by Markov chains, while Meta-Heuristics are used for solving static optimization problems where the underlying system is any discrete-event stochastic system (not necessarily driven by Markov chains). 
This book provides a selected collection of topics, mostly focused on model-free techniques, which are useful when one does not have access to the structure of the objective function (in static optimization) or the transition probability function (in dynamic optimization).   My goal was neither to overwhelm the reader with mathematical details nor was it to cover every topic.  Rather, the goal was to provide the reader with an overview of the fundamental concepts and at the same time provide the details required for solving real-world stochastic optimization problems via simulation-based techniques.
Some of the main topics covered are:

  • Reinforcement learning techniques, mainly rooted in Q-Learning for discounted and average reward MDPs
  • Static optimization techniques rooted in meta-heuristics (simulated annealing, genetic algorithms, and tabu search) for discrete solution spaces and simultaneous perturbation for continuous solution spaces
  • Neural network algorithms useful for function approximation in response surface methods for static optimization and in reinforcement learning for MDPs with large state-action spaces
  • A detailed background on dynamic programming (value and policy iteration)
  • A special coverage of semi-MDPs (SMDPs) and average reward problems
  • A discussion on convergence of a subset of methods enumerated above

From the Back Cover

Simulation-Based Optimization: Parametric Optimization Techniques and Reinforcement Learning introduces the evolving area of static and dynamic simulation-based optimization. Covered in detail are model-free optimization techniques – especially designed for those discrete-event, stochastic systems which can be simulated but whose analytical models are difficult to find in closed mathematical forms.

Key features of this revised and improved Second Edition include:

· Extensive coverage, via step-by-step recipes, of powerful new algorithms for static simulation optimization, including simultaneous perturbation, backtracking adaptive search, and nested partitions, in addition to traditional methods, such as response surfaces, Nelder-Mead search, and meta-heuristics (simulated annealing, tabu search, and genetic algorithms)

· Detailed coverage of the Bellman equation framework for Markov Decision Processes (MDPs), along with dynamic programming (value and policy iteration) for  discounted, average, and total reward performance metrics

· An in-depth consideration of dynamic simulation optimization via temporal differences and Reinforcement Learning: Q-Learning, SARSA, and R-SMART algorithms, and policy search, via API, Q-P-Learning, actor-critics, and learning automata

· A special examination of neural-network-based function approximation for Reinforcement Learning, semi-Markov decision processes (SMDPs), finite-horizon problems, two time scales, case studies for industrial tasks, computer codes (placed online), and convergence proofs, via Banach fixed point theory and Ordinary Differential Equations

Themed around three areas in separate sets of chapters – Static Simulation Optimization, Reinforcement Learning, and Convergence Analysis – this book is written for researchers and students in the fields of engineering (industrial, systems, electrical, and computer), operations research, computer science, and applied mathematics.

"About this title" may belong to another edition of this title.

  • PublisherSpringer
  • Publication date2003
  • ISBN 10 1402074549
  • ISBN 13 9781402074547
  • BindingHardcover
  • LanguageEnglish
  • Number of pages554
  • Rating
    • 3.80 out of 5 stars
      5 ratings by Goodreads

Buy Used

Condition: As New
Like New
View this item

US$ 33.54 shipping from United Kingdom to U.S.A.

Destination, rates & speeds

Other Popular Editions of the Same Title

9781441953544: Simulation-Based Optimization: Parametric Optimization Techniques and Reinforcement Learning (Operations Research/Computer Science Interfaces Series)

Featured Edition

ISBN 10:  144195354X ISBN 13:  9781441953544
Publisher: Springer, 2010
Softcover

Search results for Simulation-Based Optimization: Parametric Optimization...

Stock Image

Gosavi, Abhijit
Published by Springer, 2003
ISBN 10: 1402074549 ISBN 13: 9781402074547
Used Hardcover

Seller: Mispah books, Redhill, SURRE, United Kingdom

Seller rating 4 out of 5 stars 4-star rating, Learn more about seller ratings

Hardcover. Condition: Like New. Like New. book. Seller Inventory # ERICA79614020745496

Contact seller

Buy Used

US$ 395.20
Convert currency
Shipping: US$ 33.54
From United Kingdom to U.S.A.
Destination, rates & speeds

Quantity: 1 available

Add to basket