咨询与建议

限定检索结果

文献类型

  • 228 篇 会议
  • 4 篇 期刊文献

馆藏范围

  • 232 篇 电子文献
  • 0 种 纸本馆藏

日期分布

学科分类号

  • 98 篇 工学
    • 93 篇 计算机科学与技术...
    • 40 篇 软件工程
    • 25 篇 电气工程
    • 14 篇 控制科学与工程
    • 4 篇 机械工程
    • 1 篇 力学(可授工学、理...
    • 1 篇 信息与通信工程
    • 1 篇 建筑学
    • 1 篇 化学工程与技术
    • 1 篇 交通运输工程
  • 23 篇 理学
    • 23 篇 数学
    • 6 篇 统计学(可授理学、...
    • 4 篇 系统科学
    • 1 篇 化学
    • 1 篇 大气科学
  • 9 篇 管理学
    • 7 篇 管理科学与工程(可...
    • 3 篇 工商管理
    • 2 篇 图书情报与档案管...
  • 2 篇 经济学
    • 2 篇 应用经济学
  • 1 篇 法学
    • 1 篇 社会学

主题

  • 95 篇 dynamic programm...
  • 52 篇 learning
  • 46 篇 optimal control
  • 37 篇 reinforcement le...
  • 34 篇 learning (artifi...
  • 27 篇 equations
  • 22 篇 heuristic algori...
  • 21 篇 control systems
  • 20 篇 convergence
  • 19 篇 neural networks
  • 18 篇 function approxi...
  • 17 篇 mathematical mod...
  • 16 篇 approximation al...
  • 15 篇 vectors
  • 14 篇 markov processes
  • 14 篇 artificial neura...
  • 14 篇 cost function
  • 13 篇 stochastic proce...
  • 12 篇 algorithm design...
  • 12 篇 adaptive control

机构

  • 5 篇 school of inform...
  • 4 篇 northeastern uni...
  • 4 篇 department of el...
  • 4 篇 department of in...
  • 3 篇 department of el...
  • 3 篇 automation and r...
  • 3 篇 northeastern uni...
  • 3 篇 robotics institu...
  • 3 篇 key laboratory o...
  • 3 篇 univ illinois de...
  • 2 篇 department of ar...
  • 2 篇 school of electr...
  • 2 篇 univ groningen i...
  • 2 篇 univ texas autom...
  • 2 篇 colorado state u...
  • 2 篇 guangxi univ sch...
  • 2 篇 national science...
  • 2 篇 informatics inst...
  • 2 篇 college of infor...
  • 2 篇 school of automa...

作者

  • 7 篇 hado van hasselt
  • 7 篇 lewis frank l.
  • 7 篇 marco a. wiering
  • 7 篇 dongbin zhao
  • 6 篇 liu derong
  • 5 篇 huaguang zhang
  • 5 篇 zhang huaguang
  • 5 篇 derong liu
  • 5 篇 warren b. powell
  • 4 篇 xu xin
  • 4 篇 vrabie draguna
  • 4 篇 jagannathan s.
  • 4 篇 frank l. lewis
  • 4 篇 yanhong luo
  • 4 篇 damien ernst
  • 4 篇 jan peters
  • 4 篇 peters jan
  • 4 篇 zhao dongbin
  • 3 篇 xu hao
  • 3 篇 martin riedmille...

语言

  • 232 篇 英文
检索条件"任意字段=2009 IEEE Symposium on Adaptive Dynamic Programming and Reinforcement Learning, ADPRL 2009"
232 条 记 录,以下是71-80 订阅
排序:
Discrete-time adaptive dynamic programming using wavelet basis function neural networks
Discrete-time adaptive dynamic programming using wavelet bas...
收藏 引用
ieee International symposium on Approximate dynamic programming and reinforcement learning
作者: Jin, Ning Liu, Derong Huang, Ting Pang, Zhongyu Univ Illinois Dept Elect & Comp Engn Chicago IL 60607 USA
dynamic programming for discrete time systems is difficult due to the "curse of dimensionality": one has to find a series of control actions that must be taken in sequence, hoping that this sequence will lea... 详细信息
来源: 评论
Using ADP to understand and replicate brain intelligence: the next level design
Using ADP to understand and replicate brain intelligence: th...
收藏 引用
ieee International symposium on Approximate dynamic programming and reinforcement learning
作者: Werbos, Paul J. Natl Sci Fdn Arlington VA 22203 USA
Since the 1960's I proposed that we could understand and replicate the highest level of intelligence seen in the brain, by building ever more capable and general systems for adaptive dynamic programming (ADP) - li... 详细信息
来源: 评论
Protecting against evaluation overfitting in empirical reinforcement learning
Protecting against evaluation overfitting in empirical reinf...
收藏 引用
ieee symposium on adaptive dynamic programming and reinforcement learning
作者: Whiteson, Shimon Tanner, Brian Taylor, Matthew E. Stone, Peter Informatics Institute University of Amsterdam Netherlands Department of Computing Science University of Alberta Canada Department of Computer Science Lafayette College United States Department of Computer Science University of Texas Austin United States
Empirical evaluations play an important role in machine learning. However, the usefulness of any evaluation depends on the empirical methodology employed. Designing good empirical methodologies is difficult in part be... 详细信息
来源: 评论
Convergence of Value Iterations for Total-Cost MDPs and POMDPs with General State and Action Sets
Convergence of Value Iterations for Total-Cost MDPs and POMD...
收藏 引用
ieee symposium on adaptive dynamic programming and reinforcement learning (adprl)
作者: Feinberg, Eugene A. Kasyanov, Pavlo O. Zgurovsky, Michael Z. SUNY Stony Brook Dept Appl Math & Stat Stony Brook NY 11794 USA Natl Tech Univ Ukraine Kyiv Polytech Inst Inst Appl Syst Anal UA-03056 Kiev Ukraine Natl Tech Univ Ukraine Kyiv Polytech Inst UA-03056 Kiev Ukraine
This paper describes conditions for convergence to optimal values of the dynamic programming algorithm applied to total-cost Markov Decision Processes (MDPSs) with Borel state and action sets and with possibly unbound... 详细信息
来源: 评论
reinforcement learning algorithms for solving classification problems
Reinforcement learning algorithms for solving classification...
收藏 引用
ieee symposium on adaptive dynamic programming and reinforcement learning
作者: Wiering, Marco A. Van Hasselt, Hado Pietersma, Auke-Dirk Schomaker, Lambert Dept. of Artificial Intelligence University of Groningen Netherlands Multi-agent and Adaptive Computation Centrum Wiskunde en Informatica Netherlands
We describe a new framework for applying reinforcement learning (RL) algorithms to solve classification tasks by letting an agent act on the inputs and learn value functions. This paper describes how classification pr... 详细信息
来源: 评论
Using reward-weighted regression for reinforcement learning of task space control
Using reward-weighted regression for reinforcement learning ...
收藏 引用
ieee International symposium on Approximate dynamic programming and reinforcement learning
作者: Peters, Jan Schaal, Stefan Univ So Calif Los Angeles CA 90089 USA
Many robot control problems of practical importance, including task or operational space control, can be reformulated as immediate reward reinforcement learning problems. However, few of the known optimization or rein... 详细信息
来源: 评论
dynamic lead time promising
Dynamic lead time promising
收藏 引用
ieee symposium on adaptive dynamic programming and reinforcement learning
作者: Reindorp, Matthew J. Fu, Michael C. Department of Industrial Engineering and Innovation Sciences Eindhoven University of Technology Netherlands Robert H. Smith School of Business Institute for Systems Research University of Maryland United States
We consider a make-to-order business that serves customers in multiple priority classes. Orders from customers in higher classes bring greater revenue, but they expect shorter lead times than customers in lower classe... 详细信息
来源: 评论
On learning with imperfect representations
On learning with imperfect representations
收藏 引用
ieee symposium on adaptive dynamic programming and reinforcement learning
作者: Kalyanakrishnan, Shivaram Stone, Peter Department of Computer Science University of Texas at Austin 1616 Guadalupe St Austin TX 78701 United States
In this paper we present a perspective on the relationship between learning and representation in sequential decision making tasks. We undertake a brief survey of existing real-world applications, which demonstrates t... 详细信息
来源: 评论
reinforcement learning-based Optimal Control Considering L Computation Time Delay of Linear Discrete-time Systems
Reinforcement Learning-based Optimal Control Considering <i>...
收藏 引用
ieee symposium on adaptive dynamic programming and reinforcement learning (adprl)
作者: Fujita, Taishi Ushio, Toshimitsu
In embedded control systems, the control input is computed based on sensing data of a plant in a processor and there is a delay, called the computation time delay, due to the computation and the data transmission. Whe... 详细信息
来源: 评论
Improved neural fitted Q iteration applied to a novel computer gaming and learning benchmark
Improved neural fitted Q iteration applied to a novel comput...
收藏 引用
ieee symposium on adaptive dynamic programming and reinforcement learning
作者: Gabel, Thomas Lutz, Christian Riedmiller, Martin Machine Learning Lab Department of Computer Science University of Freiburg 79110 Freiburg Germany
Neural batch reinforcement learning (RL) algorithms have recently shown to be a powerful tool for model-free reinforcement learning problems. In this paper, we present a novel learning benchmark from the realm of comp... 详细信息
来源: 评论