咨询与建议

限定检索结果

文献类型

  • 751 篇 会议
  • 272 篇 期刊文献
  • 4 册 图书

馆藏范围

  • 1,027 篇 电子文献
  • 1 种 纸本馆藏

日期分布

学科分类号

  • 719 篇 工学
    • 523 篇 计算机科学与技术...
    • 385 篇 电气工程
    • 284 篇 控制科学与工程
    • 153 篇 软件工程
    • 83 篇 信息与通信工程
    • 41 篇 交通运输工程
    • 24 篇 仪器科学与技术
    • 21 篇 机械工程
    • 9 篇 电子科学与技术(可...
    • 9 篇 生物工程
    • 7 篇 力学(可授工学、理...
    • 7 篇 土木工程
    • 7 篇 石油与天然气工程
    • 6 篇 动力工程及工程热...
    • 4 篇 材料科学与工程(可...
    • 4 篇 生物医学工程(可授...
    • 4 篇 安全科学与工程
    • 3 篇 化学工程与技术
    • 3 篇 航空宇航科学与技...
  • 120 篇 理学
    • 98 篇 数学
    • 31 篇 系统科学
    • 22 篇 统计学(可授理学、...
    • 10 篇 生物学
    • 9 篇 物理学
    • 5 篇 化学
  • 68 篇 管理学
    • 65 篇 管理科学与工程(可...
    • 14 篇 工商管理
    • 7 篇 图书情报与档案管...
  • 5 篇 经济学
    • 4 篇 应用经济学
  • 3 篇 法学
    • 3 篇 社会学
  • 2 篇 医学
  • 1 篇 教育学

主题

  • 315 篇 reinforcement le...
  • 216 篇 dynamic programm...
  • 206 篇 optimal control
  • 110 篇 adaptive dynamic...
  • 105 篇 adaptive dynamic...
  • 97 篇 learning
  • 88 篇 neural networks
  • 79 篇 heuristic algori...
  • 67 篇 reinforcement le...
  • 58 篇 learning (artifi...
  • 54 篇 nonlinear system...
  • 52 篇 convergence
  • 52 篇 control systems
  • 51 篇 mathematical mod...
  • 48 篇 approximate dyna...
  • 44 篇 approximation al...
  • 43 篇 equations
  • 42 篇 adaptive control
  • 41 篇 cost function
  • 40 篇 artificial neura...

机构

  • 41 篇 chinese acad sci...
  • 27 篇 univ rhode isl d...
  • 17 篇 tianjin univ sch...
  • 16 篇 northeastern uni...
  • 16 篇 univ sci & techn...
  • 16 篇 univ illinois de...
  • 14 篇 beijing normal u...
  • 13 篇 northeastern uni...
  • 13 篇 guangdong univ t...
  • 12 篇 northeastern uni...
  • 9 篇 natl univ def te...
  • 8 篇 ieee
  • 8 篇 univ chinese aca...
  • 7 篇 univ chinese aca...
  • 7 篇 cent south univ ...
  • 7 篇 southern univ sc...
  • 7 篇 beijing univ tec...
  • 6 篇 chinese acad sci...
  • 6 篇 missouri univ sc...
  • 5 篇 nanjing univ pos...

作者

  • 55 篇 liu derong
  • 37 篇 wei qinglai
  • 29 篇 he haibo
  • 22 篇 wang ding
  • 21 篇 xu xin
  • 19 篇 jiang zhong-ping
  • 17 篇 lewis frank l.
  • 17 篇 yang xiong
  • 17 篇 zhang huaguang
  • 17 篇 ni zhen
  • 16 篇 zhao bo
  • 16 篇 gao weinan
  • 14 篇 zhao dongbin
  • 13 篇 zhong xiangnan
  • 12 篇 si jennie
  • 12 篇 derong liu
  • 11 篇 song ruizhuo
  • 10 篇 jagannathan s.
  • 10 篇 dongbin zhao
  • 9 篇 abouheaf mohamme...

语言

  • 970 篇 英文
  • 51 篇 其他
  • 6 篇 中文
检索条件"任意字段=IEEE Symposium on Adaptive Dynamic Programming and Reinforcement Learning"
1027 条 记 录,以下是881-890 订阅
Fitted Q iteration with CMACs
Fitted Q iteration with CMACs
收藏 引用
ieee International symposium on Approximate dynamic programming and reinforcement learning
作者: Timmer, Stephan Riedmiller, Martin Univ Osnabruck Dept Comp Sci D-4500 Osnabruck Germany
A major issue in model-free reinforcement learning is how to efficiently exploit the data collected by an exploration strategy. This is especially important in case of continuous, high dimensional state spaces, since ... 详细信息
来源: 评论
Kernelizing LSPE(λ)
Kernelizing LSPE(λ)
收藏 引用
ieee International symposium on Approximate dynamic programming and reinforcement learning
作者: Jung, Tobias Polani, Daniel Johannes Gutenberg Univ Mainz D-6500 Mainz Germany Univ Hertfordshir Hatfield Herts England
We propose the use of kernel-based methods as underlying function approximator in the least-squares based policy evaluation framework of LSPE(lambda) and LSTD(lambda). In particular we present the Ikernelization' ... 详细信息
来源: 评论
Continuous-time adaptive critics
收藏 引用
ieee TRANSACTIONS ON NEURAL NETWORKS 2007年 第3期18卷 631-647页
作者: Hanselmann, Thomas Noakes, Lyle Zaknich, Anthony Univ Melbourne Dept Elect & Elect Engn Parkville Vic 3010 Australia Univ Western Australia Sch Math & Stat Crawley WA 6009 Australia Murdoch Univ Sch Engn Sci Perth WA 6150 Australia
A continuous-time formulation of an adaptive critic design (ACD) is investigated. Connections to the discrete case are made, where backpropagation through time (BPTT) and real-time recurrent learning (RTRL) are preval... 详细信息
来源: 评论
The effect of bootstrapping in multi-automata reinforcement learning
The effect of bootstrapping in multi-automata reinforcement ...
收藏 引用
ieee International symposium on Approximate dynamic programming and reinforcement learning
作者: Peeters, Maarten Verbeeck, Katja Nowe, Ann Vrije Univ Brussel Computat Modeling Lab Pleinlaan 2 B-1050 Brussels Belgium
learning Automata are shown to be an excellent tool for creating learning multi-agent systems. Most algorithms used in current automata research expect the environment to end in an explicit end-stage. In this end-stag... 详细信息
来源: 评论
A dynamic programming approach to viability problems
A dynamic programming approach to viability problems
收藏 引用
ieee International symposium on Approximate dynamic programming and reinforcement learning
作者: Coquelin, Pieffe-Amaud Martin, Sophie Munos, Reni Ecole Polytech Ctr Math Appl Palaiseau France Approximate Dynamic Programm Paris France
Viability theory considers the problem of maintaining a system under a set of viability constraints. The main tool for solving viability problems lies in the construction of he hi viability kernel, defined as the set ... 详细信息
来源: 评论
An approximate dynamic programming approach for job releasing and Sequencing in a Reentrant manufacturing line
An approximate dynamic programming approach for job releasin...
收藏 引用
ieee International symposium on Approximate dynamic programming and reinforcement learning
作者: Ramirez-Hernandez, Jose A. Fernandez, Emmanuel Univ Cincinnati Dept Elect & Comp Engn Cincinnati OH 45221 USA Univ Cincinnati Cincinnati OH USA
This paper presents the application of an approximate dynamic programming (ADP) algorithm to the problem of job releasing and sequencing of a benchmark reentrant manufacturing line (RML). The ADP approach is based on ... 详细信息
来源: 评论
SVM viability controller active learning: Application to bike control
SVM viability controller active learning: Application to bik...
收藏 引用
ieee International symposium on Approximate dynamic programming and reinforcement learning
作者: Chapel, Laetitia Deffuant, Guillaume Cemagref LISC Aubiere France
It was shown recently that SVMs are particularly adequate to define action policies to keep a dynamical system inside a given constraint set (in the framework of viability theory). However, the training set of the SVM... 详细信息
来源: 评论
Q-learning with continuous state spaces and finite decision set
Q-learning with continuous state spaces and finite decision ...
收藏 引用
ieee International symposium on Approximate dynamic programming and reinforcement learning
作者: Barty, Kengy Girardeau, Pierre Roy, Jean-Sebastien Strugarek, Cyrille EDF R&D 1 Ave Gen Gaulle F-92141 Clamart France
This paper aims to present an original technique in order to compute the optimal policy of a Markov Decision Problem with continuous state space and discrete decision variables. We propose an extension of the Q-learni... 详细信息
来源: 评论
Opposition-based reinforcement learning in the management of water resources
Opposition-based reinforcement learning in the management of...
收藏 引用
ieee International symposium on Approximate dynamic programming and reinforcement learning
作者: Mahootchi, M. Tizhoosh, H. R. Ponnambalam, K. Univ Waterloo Sysr Design Engn 200 Univ Ave W Waterloo ON N2L 3G1 Canada
Opposition-Based learning (OBL) is a new scheme in machine intelligence. In this paper, an OBL version Q-learning which exploits opposite quantities to accelerate the learning is used for management of single reservoi... 详细信息
来源: 评论
Efficient learning in cellular simultaneous recurrent neural networks - The case of maze navigation problem
Efficient learning in cellular simultaneous recurrent neural...
收藏 引用
ieee International symposium on Approximate dynamic programming and reinforcement learning
作者: Ilin, Roman Kozma, Robert Werbos, Paul J. Univ Memphis Dept Math Sci Memphis TN 38117 USA Natl Sci Fdn Arlington VA 22230 USA
Cellular Simultaneous Recurrent Neural Networks (SRN) show great promise in solving complex function approximation problems. In particular, approximate dynamic programming is an important application area where SRNs h... 详细信息
来源: 评论