Learning to Perceive and Act by Trial and Error
ISSN: |
0885-6125
|
---|---|
Keywords: |
Reinforcement learning ; deictic representations ; sensory-motor integration ; hidden state ; non-Markov decision problems
|
Source: |
Springer Online Journal Archives 1860-2000
|
Topics: |
Computer Science
|
Notes: |
Abstract This article considers adaptive control architectures that integrate active sensory-motor systems with decision systems based on reinforcement learning. One unavoidable consequence of active perception is that the agent's internal representation often confounds external world states. We call this phoenomenon perceptual aliasingand show that it destabilizes existing reinforcement learning algorithms with respect to the optimal decision policy. We then describe a new decision system that overcomes these difficulties for a restricted class of decision problems. The system incorporates a perceptual subcycle within the overall decision cycle and uses a modified learning algorithm to suppress the effects of perceptual aliasing. The result is a control architecture that learns not only how to solve a task but also where to focus its visual attention in order to collect necessary sensory information.
|
Type of Medium: |
Electronic Resource
|
URL: |
_version_ | 1798295267962781698 |
---|---|
autor | Whitehead, Steven D. Ballard, Dana H. |
autorsonst | Whitehead, Steven D. Ballard, Dana H. |
book_url | http://dx.doi.org/10.1023/A:1022619109594 |
datenlieferant | nat_lic_papers |
hauptsatz | hsatz_simple |
identnr | NLM189898216 |
issn | 0885-6125 |
journal_name | Machine learning |
materialart | 1 |
notes | Abstract This article considers adaptive control architectures that integrate active sensory-motor systems with decision systems based on reinforcement learning. One unavoidable consequence of active perception is that the agent's internal representation often confounds external world states. We call this phoenomenon perceptual aliasingand show that it destabilizes existing reinforcement learning algorithms with respect to the optimal decision policy. We then describe a new decision system that overcomes these difficulties for a restricted class of decision problems. The system incorporates a perceptual subcycle within the overall decision cycle and uses a modified learning algorithm to suppress the effects of perceptual aliasing. The result is a control architecture that learns not only how to solve a task but also where to focus its visual attention in order to collect necessary sensory information. |
package_name | Springer |
publikationsjahr_anzeige | 1991 |
publikationsjahr_facette | 1991 |
publikationsjahr_intervall | 8009:1990-1994 |
publikationsjahr_sort | 1991 |
publisher | Springer |
reference | 7 (1991), S. 45-83 |
schlagwort | Reinforcement learning deictic representations sensory-motor integration hidden state non-Markov decision problems |
search_space | articles |
shingle_author_1 | Whitehead, Steven D. Ballard, Dana H. |
shingle_author_2 | Whitehead, Steven D. Ballard, Dana H. |
shingle_author_3 | Whitehead, Steven D. Ballard, Dana H. |
shingle_author_4 | Whitehead, Steven D. Ballard, Dana H. |
shingle_catch_all_1 | Whitehead, Steven D. Ballard, Dana H. Learning to Perceive and Act by Trial and Error Reinforcement learning deictic representations sensory-motor integration hidden state non-Markov decision problems Reinforcement learning deictic representations sensory-motor integration hidden state non-Markov decision problems Abstract This article considers adaptive control architectures that integrate active sensory-motor systems with decision systems based on reinforcement learning. One unavoidable consequence of active perception is that the agent's internal representation often confounds external world states. We call this phoenomenon perceptual aliasingand show that it destabilizes existing reinforcement learning algorithms with respect to the optimal decision policy. We then describe a new decision system that overcomes these difficulties for a restricted class of decision problems. The system incorporates a perceptual subcycle within the overall decision cycle and uses a modified learning algorithm to suppress the effects of perceptual aliasing. The result is a control architecture that learns not only how to solve a task but also where to focus its visual attention in order to collect necessary sensory information. 0885-6125 08856125 Springer |
shingle_catch_all_2 | Whitehead, Steven D. Ballard, Dana H. Learning to Perceive and Act by Trial and Error Reinforcement learning deictic representations sensory-motor integration hidden state non-Markov decision problems Reinforcement learning deictic representations sensory-motor integration hidden state non-Markov decision problems Abstract This article considers adaptive control architectures that integrate active sensory-motor systems with decision systems based on reinforcement learning. One unavoidable consequence of active perception is that the agent's internal representation often confounds external world states. We call this phoenomenon perceptual aliasingand show that it destabilizes existing reinforcement learning algorithms with respect to the optimal decision policy. We then describe a new decision system that overcomes these difficulties for a restricted class of decision problems. The system incorporates a perceptual subcycle within the overall decision cycle and uses a modified learning algorithm to suppress the effects of perceptual aliasing. The result is a control architecture that learns not only how to solve a task but also where to focus its visual attention in order to collect necessary sensory information. 0885-6125 08856125 Springer |
shingle_catch_all_3 | Whitehead, Steven D. Ballard, Dana H. Learning to Perceive and Act by Trial and Error Reinforcement learning deictic representations sensory-motor integration hidden state non-Markov decision problems Reinforcement learning deictic representations sensory-motor integration hidden state non-Markov decision problems Abstract This article considers adaptive control architectures that integrate active sensory-motor systems with decision systems based on reinforcement learning. One unavoidable consequence of active perception is that the agent's internal representation often confounds external world states. We call this phoenomenon perceptual aliasingand show that it destabilizes existing reinforcement learning algorithms with respect to the optimal decision policy. We then describe a new decision system that overcomes these difficulties for a restricted class of decision problems. The system incorporates a perceptual subcycle within the overall decision cycle and uses a modified learning algorithm to suppress the effects of perceptual aliasing. The result is a control architecture that learns not only how to solve a task but also where to focus its visual attention in order to collect necessary sensory information. 0885-6125 08856125 Springer |
shingle_catch_all_4 | Whitehead, Steven D. Ballard, Dana H. Learning to Perceive and Act by Trial and Error Reinforcement learning deictic representations sensory-motor integration hidden state non-Markov decision problems Reinforcement learning deictic representations sensory-motor integration hidden state non-Markov decision problems Abstract This article considers adaptive control architectures that integrate active sensory-motor systems with decision systems based on reinforcement learning. One unavoidable consequence of active perception is that the agent's internal representation often confounds external world states. We call this phoenomenon perceptual aliasingand show that it destabilizes existing reinforcement learning algorithms with respect to the optimal decision policy. We then describe a new decision system that overcomes these difficulties for a restricted class of decision problems. The system incorporates a perceptual subcycle within the overall decision cycle and uses a modified learning algorithm to suppress the effects of perceptual aliasing. The result is a control architecture that learns not only how to solve a task but also where to focus its visual attention in order to collect necessary sensory information. 0885-6125 08856125 Springer |
shingle_title_1 | Learning to Perceive and Act by Trial and Error |
shingle_title_2 | Learning to Perceive and Act by Trial and Error |
shingle_title_3 | Learning to Perceive and Act by Trial and Error |
shingle_title_4 | Learning to Perceive and Act by Trial and Error |
sigel_instance_filter | dkfz geomar wilbert ipn albert fhp |
source_archive | Springer Online Journal Archives 1860-2000 |
timestamp | 2024-05-06T09:33:28.966Z |
titel | Learning to Perceive and Act by Trial and Error |
titel_suche | Learning to Perceive and Act by Trial and Error |
topic | SQ-SU |
uid | nat_lic_papers_NLM189898216 |