日本語
 
Help Privacy Policy ポリシー/免責事項
  詳細検索ブラウズ

アイテム詳細

  Learning Visual Representations for Perception-Action Systems

Piater, J., Jodogne, S., Detry, R., Kraft, R., Krüger, N., Kroemer, O., & Peters, J. (2011). Learning Visual Representations for Perception-Action Systems. The International Journal of Robotics Research, 30(3), 294-307. doi:10.1177/0278364910382464.

Item is

基本情報

表示: 非表示:
資料種別: 学術論文

ファイル

表示: ファイル

関連URL

表示:

作成者

表示:
非表示:
 作成者:
Piater, J, 著者           
Jodogne, S, 著者
Detry, R, 著者
Kraft, R, 著者
Krüger, N, 著者
Kroemer, O.1, 2, 著者           
Peters, J1, 2, 著者           
所属:
1Department Empirical Inference, Max Planck Institute for Biological Cybernetics, Max Planck Society, ou_1497795              
2Max Planck Institute for Biological Cybernetics, Max Planck Society, Spemannstrasse 38, 72076 Tübingen, DE, ou_1497794              

内容説明

表示:
非表示:
キーワード: -
 要旨: We discuss vision as a sensory modality for systems that interact flexibly with uncontrolled environments. Instead of trying to build a generic vision system that produces task-independent representations, we argue in favor of task-specific, learnable representations. This concept is illustrated by two examples of our own work. First, our RLVC algorithm performs reinforcement learning directly on the visual input space. To make this very large space manageable, RLVC interleaves the reinforcement learner with a supervised classification algorithm that seeks to split perceptual states so as to reduce perceptual aliasing. This results in an adaptive discretization of the perceptual space based on the presence or absence of visual features. Its extension, RLJC, additionally handles continuous action spaces. In contrast to the minimalistic visual representations produced by RLVC and RLJC, our second method learns structural object models for robust object detection and pose estimation by probabilistic inference. To these models, the method associates grasp experiences autonomously learned by trial and error. These experiences form a non-parametric representation of grasp success likelihoods over gripper poses, which we call a grasp density. Thus, object detection in a novel scene simultaneously produces suitable grasping options.

資料詳細

表示:
非表示:
言語:
 日付: 2011-02
 出版の状態: 出版
 ページ: -
 出版情報: -
 目次: -
 査読: -
 識別子(DOI, ISBNなど): DOI: 10.1177/0278364910382464
BibTex参照ID: 6707
 学位: -

関連イベント

表示:

訴訟

表示:

Project information

表示:

出版物 1

表示:
非表示:
出版物名: The International Journal of Robotics Research
種別: 学術雑誌
 著者・編者:
所属:
出版社, 出版地: Cambridge, MA : Sage Publications, Inc.
ページ: - 巻号: 30 (3) 通巻号: - 開始・終了ページ: 294 - 307 識別子(ISBN, ISSN, DOIなど): ISSN: 0278-3649
CoNE: https://pure.mpg.de/cone/journals/resource/954925506289