Whiteson, 2006 - Google Patents
Evolutionary function approximation for reinforcement learningWhiteson, 2006
View PDF- Document ID
- 1844804372793303956
- Author
- Whiteson S
- Publication year
- Publication venue
- Journal of Machine Learning Research
External Links
Snippet
Temporal difference methods are theoretically grounded and empirically effective methods for addressing reinforcement learning problems. In most real-world reinforcement learning tasks, TD methods require a function approximator to represent the value function. However …
- 230000002787 reinforcement 0 title abstract description 42
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computer systems based on biological models
- G06N3/12—Computer systems based on biological models using genetic models
- G06N3/126—Genetic algorithms, i.e. information processing using digital simulations of the genetic system
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computer systems based on biological models
- G06N3/02—Computer systems based on biological models using neural network models
- G06N3/04—Architectures, e.g. interconnection topology
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N99/00—Subject matter not provided for in other groups of this subclass
- G06N99/005—Learning machines, i.e. computer in which a programme is changed according to experience gained by the machine itself during a complete run
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computer systems based on biological models
- G06N3/02—Computer systems based on biological models using neural network models
- G06N3/08—Learning methods
- G06N3/086—Learning methods using evolutionary programming, e.g. genetic algorithms
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N5/00—Computer systems utilising knowledge based models
- G06N5/04—Inference methods or devices
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05B—CONTROL OR REGULATING SYSTEMS IN GENERAL; FUNCTIONAL ELEMENTS OF SUCH SYSTEMS; MONITORING OR TESTING ARRANGEMENTS FOR SUCH SYSTEMS OR ELEMENTS
- G05B13/00—Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion
- G05B13/02—Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion electric
- G05B13/0265—Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion electric the criterion being a learning criterion
- G05B13/027—Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion electric the criterion being a learning criterion using neural networks only
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N5/00—Computer systems utilising knowledge based models
- G06N5/02—Knowledge representation
- G06N5/022—Knowledge engineering, knowledge acquisition
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06Q—DATA PROCESSING SYSTEMS OR METHODS, SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL, SUPERVISORY OR FORECASTING PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL, SUPERVISORY OR FORECASTING PURPOSES, NOT OTHERWISE PROVIDED FOR
- G06Q40/00—Finance; Insurance; Tax strategies; Processing of corporate or income taxes
- G06Q40/06—Investment, e.g. financial instruments, portfolio management or fund management
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05B—CONTROL OR REGULATING SYSTEMS IN GENERAL; FUNCTIONAL ELEMENTS OF SUCH SYSTEMS; MONITORING OR TESTING ARRANGEMENTS FOR SUCH SYSTEMS OR ELEMENTS
- G05B13/00—Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion
- G05B13/02—Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion electric
- G05B13/04—Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion electric involving the use of models or simulators
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Whiteson | Evolutionary function approximation for reinforcement learning | |
Karpe et al. | Multi-agent reinforcement learning in a realistic limit order book market simulation | |
Sewak | Deep reinforcement learning | |
Bernard et al. | Learning style Identifier: Improving the precision of learning style identification through computational intelligence algorithms | |
Moriarty et al. | Evolutionary algorithms for reinforcement learning | |
Hu et al. | Deep reinforcement learning assisted co-evolutionary differential evolution for constrained optimization | |
Butz et al. | Gradient descent methods in learning classifier systems: Improving XCS performance in multistep problems | |
Whiteson et al. | Critical factors in the empirical performance of temporal difference and evolutionary methods for reinforcement learning | |
Sheneman et al. | Evolving autonomous learning in cognitive networks | |
Van Moffaert | Multi-criteria reinforcement learning for sequential decision making problems | |
Malekzadeh et al. | Active inference and reinforcement learning: A unified inference on continuous state and action spaces under partial observability | |
Whiteson et al. | On-line evolutionary computation for reinforcement learning in stochastic domains | |
Whiteson | Adaptive representations for reinforcement learning | |
Carmel et al. | How to explore your opponent's strategy (almost) optimally | |
Morales | Deep Reinforcement Learning | |
Whiteson | Adaptive representations for reinforcement learning | |
De Jong | Generalized evolutionary algorithms | |
Riedel et al. | Evolving multilevel forecast combination models-an experimental study | |
Ratitch | On characteristics of Markov decision processes and reinforcement learning in large domains | |
Zheng et al. | State-dependent maximum entropy reinforcement learning for robot long-horizon task learning | |
Alkemade et al. | On social learning and robust evolutionary algorithm design in the Cournot oligopoly game | |
CN114219099B (en) | A task planning method based on symbolic options and self-learning of action models | |
McLean | Design, evaluation and comparison of evolution and reinforcement learning models | |
Marchev Jr et al. | Testing the NEAT Algorithm on a PSPACE-Complete Problem | |
Whiteson et al. | Sample-efficient evolutionary function approximation for reinforcement learning |