Pose (computer vision)

From Wikipedia, the free encyclopedia

In computer vision and in robotics, pose refers to the 3D position and orientation of a specific object relative to some coordinate frame, depicted in an image or set of images. It can also be applied to the orientation of a living being or parts of a living being, e.g., head or hands.

The pose can be described by means of a rotation transformation which brings the object from a reference pose to the observed pose. This rotation transformation can be represented in different ways, e.g., as a rotation matrix or a quaternion.

In some applications, pose may refer to the complete transformation, also including translation, of an object from a reference pose to its observed pose. In this case a transformation on homogeneous coordinates can be used.

If a vision based system should be able to manipulate the object, e.g., pick it up, pose estimation is a necessary task to be solved by the system. This can be done in different ways depending on the image sensor configuration, and choice of methodology.

[edit] External links

  • [1] Model-Based Object pose in 25 lines of code.