• DocumentCode
    2846242
  • Title

    Acquisition of Image Feature on Collision for Robot Motion Generation

  • Author

    Okamoto, Taichi ; Kobayashi, Yuichi ; Onishi, Masaki

  • Author_Institution
    Dept. of Electr. & Electron. Eng., Tokyo Univ. of Agric. & Technol., Koganei, Japan
  • fYear
    2009
  • fDate
    Nov. 30 2009-Dec. 2 2009
  • Firstpage
    737
  • Lastpage
    742
  • Abstract
    It is important for robots that act in human-centered environments to build image processing in a bottom-up manner. This paper proposes a method to autonomously acquire image feature extraction that is suitable for motion generation while moving in unknown environment. The proposed method extracts low level features without specifying image processing for robot body and obstacles. The position of body is acquired in image by clustering of SIFT features with motion information and state transition model is generated. Based on a learning model of adaptive addition of state transition model, collision relevant features are detected. Features that emerge when the robot can not move are acquired as collision relevant features. The proposed framework is evaluated with real images of the manipulator and an obstacle in obstacle avoidance.
  • Keywords
    collision avoidance; feature extraction; manipulators; pattern clustering; robot vision; transforms; human-centered environments; image feature acquisition; image feature extraction; image processing; obstacle avoidance; robot motion generation; scale invariant feature transform features; state transition model; Computer vision; Data mining; Feature extraction; Image processing; Intelligent robots; Manipulators; Nonlinear control systems; Orbital robotics; Robot kinematics; Robot motion; Feature extraction; Obstacle avoidance; Robot motion learning;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Intelligent Systems Design and Applications, 2009. ISDA '09. Ninth International Conference on
  • Conference_Location
    Pisa
  • Print_ISBN
    978-1-4244-4735-0
  • Electronic_ISBN
    978-0-7695-3872-3
  • Type

    conf

  • DOI
    10.1109/ISDA.2009.113
  • Filename
    5365090