In this paper, we describe a method for estimating external camera parameters in real time. We investigated the effectiveness of this method for annotating real scenes with 3-D virtual objects on a wearable computer. The proposed method enables determining known natural feature points of objects through multiplied color histogram matching and template matching. This external-camera-parameter calculation method consists of three algorithms for PnP problems, and it uses each algorithm selectively. We implemented an experimental system based on our method on a wearable vision system. This experimental system can annotate real objects with 3D virtual objects by using the proposed method. The system was implemented in order to enable effective annotation in a mixed-reality environment on a wearable computing system. The system consists of an ultra small CCD camera set at the user's eye, an ultra small display, and a computer. This computer uses the proposed method to determine the camera parameters. It then renders virtual objects based on the camera parameters and synthesizes images on a display. The system works at 10 frames per second.
The copyright of the original papers published on this site belongs to IEICE. Unauthorized use of the original or translated papers is prohibited. See IEICE Provisions on Copyright for details.
Copy
Takashi OKUMA, Takeshi KURATA, Katsuhiko SAKAUE, "Real-Time Camera Parameter Estimation for 3-D Annotation on a Wearable Vision System" in IEICE TRANSACTIONS on Information,
vol. E84-D, no. 12, pp. 1668-1675, December 2001, doi: .
Abstract: In this paper, we describe a method for estimating external camera parameters in real time. We investigated the effectiveness of this method for annotating real scenes with 3-D virtual objects on a wearable computer. The proposed method enables determining known natural feature points of objects through multiplied color histogram matching and template matching. This external-camera-parameter calculation method consists of three algorithms for PnP problems, and it uses each algorithm selectively. We implemented an experimental system based on our method on a wearable vision system. This experimental system can annotate real objects with 3D virtual objects by using the proposed method. The system was implemented in order to enable effective annotation in a mixed-reality environment on a wearable computing system. The system consists of an ultra small CCD camera set at the user's eye, an ultra small display, and a computer. This computer uses the proposed method to determine the camera parameters. It then renders virtual objects based on the camera parameters and synthesizes images on a display. The system works at 10 frames per second.
URL: https://global.ieice.org/en_transactions/information/10.1587/e84-d_12_1668/_p
Copy
@ARTICLE{e84-d_12_1668,
author={Takashi OKUMA, Takeshi KURATA, Katsuhiko SAKAUE, },
journal={IEICE TRANSACTIONS on Information},
title={Real-Time Camera Parameter Estimation for 3-D Annotation on a Wearable Vision System},
year={2001},
volume={E84-D},
number={12},
pages={1668-1675},
abstract={In this paper, we describe a method for estimating external camera parameters in real time. We investigated the effectiveness of this method for annotating real scenes with 3-D virtual objects on a wearable computer. The proposed method enables determining known natural feature points of objects through multiplied color histogram matching and template matching. This external-camera-parameter calculation method consists of three algorithms for PnP problems, and it uses each algorithm selectively. We implemented an experimental system based on our method on a wearable vision system. This experimental system can annotate real objects with 3D virtual objects by using the proposed method. The system was implemented in order to enable effective annotation in a mixed-reality environment on a wearable computing system. The system consists of an ultra small CCD camera set at the user's eye, an ultra small display, and a computer. This computer uses the proposed method to determine the camera parameters. It then renders virtual objects based on the camera parameters and synthesizes images on a display. The system works at 10 frames per second.},
keywords={},
doi={},
ISSN={},
month={December},}
Copy
TY - JOUR
TI - Real-Time Camera Parameter Estimation for 3-D Annotation on a Wearable Vision System
T2 - IEICE TRANSACTIONS on Information
SP - 1668
EP - 1675
AU - Takashi OKUMA
AU - Takeshi KURATA
AU - Katsuhiko SAKAUE
PY - 2001
DO -
JO - IEICE TRANSACTIONS on Information
SN -
VL - E84-D
IS - 12
JA - IEICE TRANSACTIONS on Information
Y1 - December 2001
AB - In this paper, we describe a method for estimating external camera parameters in real time. We investigated the effectiveness of this method for annotating real scenes with 3-D virtual objects on a wearable computer. The proposed method enables determining known natural feature points of objects through multiplied color histogram matching and template matching. This external-camera-parameter calculation method consists of three algorithms for PnP problems, and it uses each algorithm selectively. We implemented an experimental system based on our method on a wearable vision system. This experimental system can annotate real objects with 3D virtual objects by using the proposed method. The system was implemented in order to enable effective annotation in a mixed-reality environment on a wearable computing system. The system consists of an ultra small CCD camera set at the user's eye, an ultra small display, and a computer. This computer uses the proposed method to determine the camera parameters. It then renders virtual objects based on the camera parameters and synthesizes images on a display. The system works at 10 frames per second.
ER -