Condition: New. 1st edition NO-PA16APR2015-KAP.
Seller: Ria Christie Collections, Uxbridge, United Kingdom
£ 33.16
Quantity: Over 20 available
Add to basketCondition: New. In English.
Seller: Revaluation Books, Exeter, United Kingdom
Paperback. Condition: Brand New. 129 pages. 9.50x7.50x0.25 inches. In Stock.
Language: English
Published by Springer International Publishing, 2017
ISBN 10: 3031004485 ISBN 13: 9783031004483
Seller: AHA-BUCH GmbH, Einbeck, Germany
Taschenbuch. Condition: Neu. Druck auf Anfrage Neuware - Printed after ordering - Many real-world decision problems have multiple objectives. For example, when choosing a medical treatment plan, we want to maximize the efficacy of the treatment, but also minimize the side effects. These objectives typically conflict, e.g., we can often increase the efficacy of the treatment, but at the cost of more severe side effects. In this book, we outline how to deal with multiple objectives in decision-theoretic planning and reinforcement learning algorithms. To illustrate this, we employ the popular problem classes of multi-objective Markov decision processes (MOMDPs) and multi-objective coordination graphs (MO-CoGs).First, we discuss different use cases for multi-objective decision making, and why they often necessitate explicitly multi-objective algorithms. We advocate a utility-based approach to multi-objective decision making, i.e., that what constitutes an optimal solution to a multi-objective decision problem should be derived from the availableinformation about user utility. We show how different assumptions about user utility and what types of policies are allowed lead to different solution concepts, which we outline in a taxonomy of multi-objective decision problems.Second, we show how to create new methods for multi-objective decision making using existing single-objective methods as a basis. Focusing on planning, we describe two ways to creating multi-objective algorithms: in the inner loop approach, the inner workings of a single-objective method are adapted to work with multi-objective solution concepts; in the outer loop approach, a wrapper is created around a single-objective method that solves the multi-objective problem as a series of single-objective problems. After discussing the creation of such methods for the planning setting, we discuss how these approaches apply to the learning setting.Next, we discuss three promising application domains for multi-objective decision making algorithms: energy, health, and infrastructure and transportation. Finally, we conclude by outlining important open problems and promising future directions.
Seller: Ria Christie Collections, Uxbridge, United Kingdom
£ 94.70
Quantity: Over 20 available
Add to basketCondition: New. In.
Seller: Ria Christie Collections, Uxbridge, United Kingdom
£ 96.15
Quantity: Over 20 available
Add to basketCondition: New. In.
Seller: Revaluation Books, Exeter, United Kingdom
Hardcover. Condition: Brand New. 133 pages. 9.00x6.25x0.25 inches. In Stock.
Language: English
Published by Springer Berlin Heidelberg, 2014
ISBN 10: 3642422314 ISBN 13: 9783642422317
Seller: AHA-BUCH GmbH, Einbeck, Germany
Taschenbuch. Condition: Neu. Druck auf Anfrage Neuware - Printed after ordering - This book presents new algorithms for reinforcement learning, a form of machine learning in which an autonomous agent seeks a control policy for a sequential decision task. Since current methods typically rely on manually designed solution representations, agents that automatically adapt their own representations have the potential to dramatically improve performance. This book introduces two novel approaches for automatically discovering high-performing representations.The first approach synthesizes temporal difference methods, the traditional approach to reinforcement learning, with evolutionary methods, which can learn representations for a broad class ofoptimization problems. This synthesis is accomplished by customizing evolutionarymethods to the on-line nature of reinforcement learning and using them to evolve representations for value function approximators.The second approach automatically learns representations based on piecewise-constant approximations of value functions. It begins with coarse representations and gradually refines them during learning, analyzing the current policy and value function to deduce the best refinements.This book also introduces a novel method for devising input representations. This method addresses the feature selection problem by extending an algorithm that evolves the topology and weights of neural networks such that it evolves their inputs too.In addition to introducing these new methods, this book presents extensive empirical results in multiple domains demonstrating that these techniques can substantially improve performance over methodswith manual representations.
Language: English
Published by Springer Berlin Heidelberg, 2010
ISBN 10: 3642139310 ISBN 13: 9783642139314
Seller: AHA-BUCH GmbH, Einbeck, Germany
Buch. Condition: Neu. Druck auf Anfrage Neuware - Printed after ordering - This book presents new algorithms for reinforcement learning, a form of machine learning in which an autonomous agent seeks a control policy for a sequential decision task. Since current methods typically rely on manually designed solution representations, agents that automatically adapt their own representations have the potential to dramatically improve performance. This book introduces two novel approaches for automatically discovering high-performing representations.The first approach synthesizes temporal difference methods, the traditional approach to reinforcement learning, with evolutionary methods, which can learn representations for a broad class ofoptimization problems. This synthesis is accomplished by customizing evolutionarymethods to the on-line nature of reinforcement learning and using them to evolve representations for value function approximators.The second approach automatically learns representations based on piecewise-constant approximations of value functions. It begins with coarse representations and gradually refines them during learning, analyzing the current policy and value function to deduce the best refinements.This book also introduces a novel method for devising input representations. This method addresses the feature selection problem by extending an algorithm that evolves the topology and weights of neural networks such that it evolves their inputs too.In addition to introducing these new methods, this book presents extensive empirical results in multiple domains demonstrating that these techniques can substantially improve performance over methodswith manual representations.
Condition: Sehr gut. Zustand: Sehr gut | Sprache: Englisch | Produktart: Bücher | This book presents new algorithms for reinforcement learning, a form of machine learning in which an autonomous agent seeks a control policy for a sequential decision task. Since current methods typically rely on manually designed solution representations, agents that automatically adapt their own representations have the potential to dramatically improve performance. This book introduces two novel approaches for automatically discovering high-performing representations. The first approach synthesizes temporal difference methods, the traditional approach to reinforcement learning, with evolutionary methods, which can learn representations for a broad class ofoptimization problems. This synthesis is accomplished by customizing evolutionarymethods to the on-line nature of reinforcement learning and using them to evolve representations for value function approximators.The second approach automatically learns representations based on piecewise-constant approximations of value functions. It begins with coarse representations and gradually refines them during learning, analyzing the current policy and value function to deduce the best refinements. This book also introduces a novel method for devising input representations. This method addresses the feature selection problem by extending an algorithm that evolves the topology and weights of neural networks such that it evolves their inputs too.In addition to introducing these new methods, this book presents extensive empirical results in multiple domains demonstrating that these techniques can substantially improve performance over methodswith manual representations.
Seller: Mispah books, Redhill, SURRE, United Kingdom
Hardcover. Condition: Like New. LIKE NEW. SHIPS FROM MULTIPLE LOCATIONS. book.
Condition: New. Print on Demand.
Language: English
Published by Springer International Publishing Apr 2017, 2017
ISBN 10: 3031004485 ISBN 13: 9783031004483
Seller: BuchWeltWeit Ludwig Meier e.K., Bergisch Gladbach, Germany
Taschenbuch. Condition: Neu. This item is printed on demand - it takes 3-4 days longer - Neuware -Many real-world decision problems have multiple objectives. For example, when choosing a medical treatment plan, we want to maximize the efficacy of the treatment, but also minimize the side effects. These objectives typically conflict, e.g., we can often increase the efficacy of the treatment, but at the cost of more severe side effects. In this book, we outline how to deal with multiple objectives in decision-theoretic planning and reinforcement learning algorithms. To illustrate this, we employ the popular problem classes of multi-objective Markov decision processes (MOMDPs) and multi-objective coordination graphs (MO-CoGs).First, we discuss different use cases for multi-objective decision making, and why they often necessitate explicitly multi-objective algorithms. We advocate a utility-based approach to multi-objective decision making, i.e., that what constitutes an optimal solution to a multi-objective decision problem should be derived from the available information about user utility. We show how different assumptions about user utility and what types of policies are allowed lead to different solution concepts, which we outline in a taxonomy of multi-objective decision problems.Second, we show how to create new methods for multi-objective decision making using existing single-objective methods as a basis. Focusing on planning, we describe two ways to creating multi-objective algorithms: in the inner loop approach, the inner workings of a single-objective method are adapted to work with multi-objective solution concepts; in the outer loop approach, a wrapper is created around a single-objective method that solves the multi-objective problem as a series of single-objective problems. After discussing the creation of such methods for the planning setting, we discuss how these approaches apply to the learning setting.Next, we discuss three promising application domains for multi-objective decision making algorithms: energy, health, and infrastructure and transportation. Finally, we conclude by outlining important open problems and promising future directions. 132 pp. Englisch.
Condition: New. PRINT ON DEMAND.
Language: English
Published by Springer, Berlin|Springer International Publishing|Morgan & Claypool|Springer, 2017
ISBN 10: 3031004485 ISBN 13: 9783031004483
Seller: moluna, Greven, Germany
Condition: New. Dieser Artikel ist ein Print on Demand Artikel und wird nach Ihrer Bestellung fuer Sie gedruckt. Many real-world decision problems have multiple objectives. For example, when choosing a medical treatment plan, we want to maximize the efficacy of the treatment, but also minimize the side effects. These objectives typically conflict, e.g., we can ofte.
Seller: Brook Bookstore On Demand, Napoli, NA, Italy
Condition: new. Questo è un articolo print on demand.
Seller: Brook Bookstore On Demand, Napoli, NA, Italy
Condition: new. Questo è un articolo print on demand.
Language: English
Published by Springer Berlin Heidelberg Okt 2010, 2010
ISBN 10: 3642139310 ISBN 13: 9783642139314
Seller: BuchWeltWeit Ludwig Meier e.K., Bergisch Gladbach, Germany
Buch. Condition: Neu. This item is printed on demand - it takes 3-4 days longer - Neuware -This book presents new algorithms for reinforcement learning, a form of machine learning in which an autonomous agent seeks a control policy for a sequential decision task. Since current methods typically rely on manually designed solution representations, agents that automatically adapt their own representations have the potential to dramatically improve performance. This book introduces two novel approaches for automatically discovering high-performing representations.The first approach synthesizes temporal difference methods, the traditional approach to reinforcement learning, with evolutionary methods, which can learn representations for a broad class ofoptimization problems. This synthesis is accomplished by customizing evolutionarymethods to the on-line nature of reinforcement learning and using them to evolve representations for value function approximators.The second approach automatically learns representations based on piecewise-constant approximations of value functions. It begins with coarse representations and gradually refines them during learning, analyzing the current policy and value function to deduce the best refinements.This book also introduces a novel method for devising input representations. This method addresses the feature selection problem by extending an algorithm that evolves the topology and weights of neural networks such that it evolves their inputs too.In addition to introducing these new methods, this book presents extensive empirical results in multiple domains demonstrating that these techniques can substantially improve performance over methodswith manual representations. 132 pp. Englisch.
Language: English
Published by Springer Berlin Heidelberg, 2014
ISBN 10: 3642422314 ISBN 13: 9783642422317
Seller: moluna, Greven, Germany
Condition: New. Dieser Artikel ist ein Print on Demand Artikel und wird nach Ihrer Bestellung fuer Sie gedruckt. Recent research in Adaptive Representations for Reinforcement LearningWritten by leading experts in this fieldState-of-the Art bookThis book presents new algorithms for reinforcement learning, a form of machine learning in which an autonomous agent seek.
Language: English
Published by Springer Berlin Heidelberg, 2010
ISBN 10: 3642139310 ISBN 13: 9783642139314
Seller: moluna, Greven, Germany
Gebunden. Condition: New. Dieser Artikel ist ein Print on Demand Artikel und wird nach Ihrer Bestellung fuer Sie gedruckt. Recent research in Adaptive Representations for Reinforcement LearningWritten by leading experts in this fieldState-of-the Art bookThis book presents new algorithms for reinforcement learning, a form of machine learning in which an autonomous agent seek.
Language: English
Published by Springer Berlin Heidelberg Nov 2014, 2014
ISBN 10: 3642422314 ISBN 13: 9783642422317
Seller: BuchWeltWeit Ludwig Meier e.K., Bergisch Gladbach, Germany
Taschenbuch. Condition: Neu. This item is printed on demand - it takes 3-4 days longer - Neuware -This book presents new algorithms for reinforcement learning, a form of machine learning in which an autonomous agent seeks a control policy for a sequential decision task. Since current methods typically rely on manually designed solution representations, agents that automatically adapt their own representations have the potential to dramatically improve performance. This book introduces two novel approaches for automatically discovering high-performing representations.The first approach synthesizes temporal difference methods, the traditional approach to reinforcement learning, with evolutionary methods, which can learn representations for a broad class ofoptimization problems. This synthesis is accomplished by customizing evolutionarymethods to the on-line nature of reinforcement learning and using them to evolve representations for value function approximators.The second approach automatically learns representations based on piecewise-constant approximations of value functions. It begins with coarse representations and gradually refines them during learning, analyzing the current policy and value function to deduce the best refinements.This book also introduces a novel method for devising input representations. This method addresses the feature selection problem by extending an algorithm that evolves the topology and weights of neural networks such that it evolves their inputs too.In addition to introducing these new methods, this book presents extensive empirical results in multiple domains demonstrating that these techniques can substantially improve performance over methodswith manual representations. 132 pp. Englisch.
Language: English
Published by Springer, Springer Okt 2010, 2010
ISBN 10: 3642139310 ISBN 13: 9783642139314
Seller: buchversandmimpf2000, Emtmannsberg, BAYE, Germany
Buch. Condition: Neu. This item is printed on demand - Print on Demand Titel. Neuware -This book presents new algorithms for reinforcement learning, a form of machine learning in which an autonomous agent seeks a control policy for a sequential decision task. Since current methods typically rely on manually designed solution representations, agents that automatically adapt their own representations have the potential to dramatically improve performance. This book introduces two novel approaches for automatically discovering high-performing representations.The first approach synthesizes temporal difference methods, the traditional approach to reinforcement learning, with evolutionary methods, which can learn representations for a broad class ofoptimization problems. This synthesis is accomplished by customizing evolutionarymethods to the on-line nature of reinforcement learning and using them to evolve representations for value function approximators.The second approach automatically learns representations based on piecewise-constant approximations of value functions. It begins with coarse representations and gradually refines them during learning, analyzing the current policy and value function to deduce the best refinements.This book also introduces a novel method for devising input representations. This method addresses the feature selection problem by extending an algorithm that evolves the topology and weights of neural networks such that it evolves their inputs too.In addition to introducing these new methods, this book presents extensive empirical results in multiple domains demonstrating that these techniques can substantially improve performance over methodswith manual representations.Springer-Verlag KG, Sachsenplatz 4-6, 1201 Wien 132 pp. Englisch.
Language: English
Published by Springer, Springer Nov 2014, 2014
ISBN 10: 3642422314 ISBN 13: 9783642422317
Seller: buchversandmimpf2000, Emtmannsberg, BAYE, Germany
Taschenbuch. Condition: Neu. This item is printed on demand - Print on Demand Titel. Neuware -This book presents new algorithms for reinforcement learning, a form of machine learning in which an autonomous agent seeks a control policy for a sequential decision task. Since current methods typically rely on manually designed solution representations, agents that automatically adapt their own representations have the potential to dramatically improve performance. This book introduces two novel approaches for automatically discovering high-performing representations.The first approach synthesizes temporal difference methods, the traditional approach to reinforcement learning, with evolutionary methods, which can learn representations for a broad class ofoptimization problems. This synthesis is accomplished by customizing evolutionarymethods to the on-line nature of reinforcement learning and using them to evolve representations for value function approximators.The second approach automatically learns representations based on piecewise-constant approximations of value functions. It begins with coarse representations and gradually refines them during learning, analyzing the current policy and value function to deduce the best refinements.This book also introduces a novel method for devising input representations. This method addresses the feature selection problem by extending an algorithm that evolves the topology and weights of neural networks such that it evolves their inputs too.In addition to introducing these new methods, this book presents extensive empirical results in multiple domains demonstrating that these techniques can substantially improve performance over methodswith manual representations.Springer-Verlag KG, Sachsenplatz 4-6, 1201 Wien 132 pp. Englisch.