460 results on '"Sato, Tomokazu"'
Search Results
2. Summarization of User-Generated Sports Video by Using Deep Action Recognition Features
- Author
-
Tejero-de-Pablos, Antonio, Nakashima, Yuta, Sato, Tomokazu, Yokoya, Naokazu, Linna, Marko, and Rahtu, Esa
- Subjects
Computer Science - Computer Vision and Pattern Recognition ,68T45 - Abstract
Automatically generating a summary of sports video poses the challenge of detecting interesting moments, or highlights, of a game. Traditional sports video summarization methods leverage editing conventions of broadcast sports video that facilitate the extraction of high-level semantics. However, user-generated videos are not edited, and thus traditional methods are not suitable to generate a summary. In order to solve this problem, this work proposes a novel video summarization method that uses players' actions as a cue to determine the highlights of the original video. A deep neural network-based approach is used to extract two types of action-related features and to classify video segments into interesting or uninteresting parts. The proposed method can be applied to any sports in which games consist of a succession of actions. Especially, this work considers the case of Kendo (Japanese fencing) as an example of a sport to evaluate the proposed method. The method is trained using Kendo videos with ground truth labels that indicate the video highlights. The labels are provided by annotators possessing different experience with respect to Kendo to demonstrate how the proposed method adapts to different needs. The performance of the proposed method is compared with several combinations of different features, and the results show that it outperforms previous summarization methods., Comment: 12 pages, 8 figures, 4 tables
- Published
- 2017
3. Ultra-Shallow DoF Imaging Using Faced Paraboloidal Mirrors
- Author
-
Nishi, Ryoichiro, Aoto, Takahito, Kawai, Norihiko, Sato, Tomokazu, Mukaigawa, Yasuhiro, Yokoya, Naokazu, Hutchison, David, Series editor, Kanade, Takeo, Series editor, Kittler, Josef, Series editor, Kleinberg, Jon M., Series editor, Mattern, Friedemann, Series editor, Mitchell, John C., Series editor, Naor, Moni, Series editor, Pandu Rangan, C., Series editor, Steffen, Bernhard, Series editor, Terzopoulos, Demetri, Series editor, Tygar, Doug, Series editor, Weikum, Gerhard, Series editor, Lai, Shang-Hong, editor, Lepetit, Vincent, editor, Nishino, Ko, editor, and Sato, Yoichi, editor
- Published
- 2017
- Full Text
- View/download PDF
4. ReMagicMirror: Action Learning Using Human Reenactment with the Mirror Metaphor
- Author
-
Dayrit, Fabian Lorenzo, Kimura, Ryosuke, Nakashima, Yuta, Blanco, Ambrosio, Kawasaki, Hiroshi, Ikeuchi, Katsushi, Sato, Tomokazu, Yokoya, Naokazu, Hutchison, David, Series editor, Kanade, Takeo, Series editor, Kittler, Josef, Series editor, Kleinberg, Jon M., Series editor, Mattern, Friedemann, Series editor, Mitchell, John C., Series editor, Naor, Moni, Series editor, Pandu Rangan, C., Series editor, Steffen, Bernhard, Series editor, Terzopoulos, Demetri, Series editor, Tygar, Doug, Series editor, Weikum, Gerhard, Series editor, Amsaleg, Laurent, editor, Guðmundsson, Gylfi Þór, editor, Gurrin, Cathal, editor, Jónsson, Björn Þór, editor, and Satoh, Shin’ichi, editor
- Published
- 2017
- Full Text
- View/download PDF
5. Iterative applications of image completion with CNN-based failure detection
- Author
-
Tanaka, Takahiro, Kawai, Norihiko, Nakashima, Yuta, Sato, Tomokazu, and Yokoya, Naokazu
- Published
- 2018
- Full Text
- View/download PDF
6. Effects of focused ultrasound in a "clean" mouse model of ultrasonic neuromodulation
- Author
-
Guo, Hongsun, primary, Salahshoor, Hossein, additional, Wu, Di, additional, Yoo, Sangjin, additional, Sato, Tomokazu, additional, Tsao, Doris, additional, and Shapiro, Mikhail, additional
- Published
- 2023
- Full Text
- View/download PDF
7. 4-D light field reconstruction by irradiance decomposition
- Author
-
Aoto, Takahito, Sato, Tomokazu, Mukaigawa, Yasuhiro, and Yokoya, Naokazu
- Published
- 2017
- Full Text
- View/download PDF
8. Diminished Reality Based on Image Inpainting Considering Background Geometry
- Author
-
Kawai, Norihiko, Sato, Tomokazu, Yokoya, Naokazu, Kawai, Norihiko, Sato, Tomokazu, and Yokoya, Naokazu
- Abstract
Diminished reality aims to remove real objects from video images and fill in the missing regions with plausible background textures in realtime. Most conventional methods based on image inpainting achieve diminished reality by assuming that the background around a target object is almost planar. This paper proposes a new diminished reality method that considers background geometries with less constraints than the conventional ones. In this study, we approximate the background geometry by combining local planes, and improve the quality of image inpainting by correcting the perspective distortion of texture and limiting the search area for finding similar textures as exemplars. The temporal coherence of texture is preserved using the geometries and camera pose estimated by visual-simultaneous localization and mapping (SLAM). The mask region that includes a target object is robustly set in each frame by projecting a 3D region, rather than tracking the object in 2D image space. The effectiveness of the proposed method is successfully demonstrated using several experimental environments.
- Published
- 2023
9. Real-time and accurate extrinsic camera parameter estimation using feature landmark database for augmented reality
- Author
-
Takafumi, Taketomi, Sato, Tomokazu, Yokoya, Naokazu, Takafumi, Taketomi, Sato, Tomokazu, and Yokoya, Naokazu
- Abstract
In the field of augmented reality (AR), many kinds of vision-based extrinsic camera parameter estimation methods have been proposed to achieve geometric registration between real and virtual worlds. Previously, a feature landmark-based camera parameter estimation method was proposed. This is an effective method for implementing outdoor AR applications because a feature landmark database can be automatically constructed using the structure-from-motion (SfM) technique. However, the previous method cannot work in real time because it entails a high computational cost or matching landmarks in a database with image features in an input image. In addition, the accuracy of estimated camera parameters is insufficient for applications that need to overlay CG objects at a position close to the user's viewpoint. This is because it is difficult to compensate for visual pattern change of close landmarks when only the sparse depth information obtained by the SfM is available. In this paper, we achieve fast and accurate feature landmark-based camera parameter estimation by adopting the following approaches. First, the number of matching candidates is reduced to achieve fast camera parameter estimation by tentative camera parameter estimation and by assigning priorities to landmarks. Second, image templates of landmarks are adequately compensated for by considering the local 3-D structure of a landmark using the dense depth information obtained by a laser range sensor. To demonstrate the effectiveness of the proposed method, we developed some AR applications using the proposed method.
- Published
- 2023
10. Omnidirectional Free-viewpoint Rendering Using a Deformable 3-D Mesh Model
- Author
-
Sato, Tomokazu, Hiroyuki, Koshizawa, Yokoya, Naokazu, Sato, Tomokazu, Hiroyuki, Koshizawa, and Yokoya, Naokazu
- Abstract
This paper proposes a method to render free viewpoint images from omnidirectional videos using a deformable 3-D mesh model. In the proposed method, a 3-D mesh is placed in front of a virtual viewpoint and deformed by using the pre-estimated omnidirectional depth maps that are selected on the basis of position and posture of the virtual viewpoint. Although our approach is fundamentally based on the model-based rendering approach that renders a geometrically correct virtualized world, in order to avoid the hole problem, we newly employ a viewpoint-dependent deformable 3-D model instead of the use of a unified 3-D model that is generally used in the model based rendering approach. In experiments, free-viewpoint images are generated from the omnidirectional video captured by an omnidirectional multi camera system to show the feasibility of the proposed method for walk-through applications in the virtualized environment.
- Published
- 2023
11. Efficient hundreds-baseline stereo by counting interest points for moving omni-directional multi-camera system
- Author
-
Sato, Tomokazu, Yokoya, Naokazu, Sato, Tomokazu, and Yokoya, Naokazu
- Abstract
In this article, we propose an efficient method for estimating a depth map from long-baseline image sequences captured by a calibrated moving multi-camera system. Our concept for estimating a depth map is very simple; we integrate the counting of the total number of interest points (TNIP) in images with the original framework of multiple baseline stereo. Even by using a simple algorithm, the depth can be determined without computing similarity measures such as SSD (sum of squared differences) and NCC (normalized cross correlation) that have been used for conventional stereo matching. The proposed stereo algorithm is computationally efficient and robust for distortions and occlusions and has high affinity with omni-directional and multi-camera imaging. Although expected trade-off between accuracy and efficiency is confirmed for a naive TNIP-based method, a hybrid approach that uses both TNIP and SSD improve this with realizing high accurate and efficient depth estimation. We have experimentally verified the validity and feasibility of the TNIP-based stereo algorithm for both synthetic and real outdoor scenes.
- Published
- 2023
12. AR image generation using view-dependent geometry modification and texture mapping
- Author
-
Nakashima, Yuta, Uno, Yusuke, Kawai, Norihiko, Sato, Tomokazu, Yokoya, Naokazu, Nakashima, Yuta, Uno, Yusuke, Kawai, Norihiko, Sato, Tomokazu, and Yokoya, Naokazu
- Abstract
Augmented reality (AR) applications often require virtualized real objects, i.e., virtual objects that are built based on real objects and rendered from an arbitrary viewpoint. In this paper, we propose a method for real object virtualization and AR image generation based on view-dependent geometry modification and texture mapping. The proposed method is a hybrid of model- and image-based rendering techniques that uses multiple input images of the real object as well as the object's three-dimensional (3D) model obtained by an automatic 3D reconstruction technique. Even with state-of-the-art technology, the reconstructed 3D model's accuracy can be insufficient, resulting in such visual artifacts as false object boundaries. The proposed method generates a depth map from a 3D model of a virtualized real object and expands its region in the depth map to remove the false object boundaries. Since such expansion reveals the background pixels in the input images, which is particularly undesirable for AR applications, we preliminarily extract object regions and use them for texture mapping. With our GPU implementation for real-time AR image generation, we experimentally demonstrated that using expanded geometry reduces the number of required input images and maintains visual quality.
- Published
- 2023
13. Texture Transfer Based on Energy Minimization for Painterly Rendering
- Author
-
Tsunematsu, Yuta, Kawai, Norihiko, Sato, Tomokazu, Yokoya, Naokazu, Tsunematsu, Yuta, Kawai, Norihiko, Sato, Tomokazu, and Yokoya, Naokazu
- Abstract
Non-photorealistic rendering (NPR) creates images with artistic styles of paintings. In this field, a number of methods of converting photographed images into non-photorealistic ones have been developed, and can be categorized into filter-based and exemplar-based approaches. In this paper, we focus on the exemplar-based approach and propose a novel method which transfers a style of a reference pictorial image to a photographed image. Specifically, we first input a pair of target and reference images. The target image is converted by minimizing an energy function which is defined based on the difference in intensities between an output image and a target image, and the pattern dissimilarity between an output image and a reference image. The proposed method transfers structures and colors of textures in the reference image and generates continuous textures by minimizing the energy function. In experiments, we demonstrate the effectiveness of the proposed method using a variety of images and examine the influence of parameter changes and intensity adjustment for pre-processing on resultant images.
- Published
- 2023
14. 4-D light field reconstruction by irradiance decomposition
- Author
-
Aoto, Takahito, Sato, Tomokazu, Mukaigawa, Yasuhiro, Yokoya, Naokazu, Aoto, Takahito, Sato, Tomokazu, Mukaigawa, Yasuhiro, and Yokoya, Naokazu
- Abstract
Common light sources such as an ordinary flashlight with lenses and/or reflectors make complex 4-D light field that cannot be represented by conventional isotropic distribution model nor point light source model. This paper describes a new approach to estimate 4-D light field using an illuminated diffuser. Unlike conventional works that capture a 4-D light field directly, our method decomposes observed intensities on the diffuser into intensities of 4-D light rays based on inverse rendering technique with prior knowledge. We formulate 4-D light field reconstruction problem as a non-smooth convex optimization problem for mathematically finding the global minimum.
- Published
- 2023
15. Extrinsic Camera Parameter Estimation Using Video Images and GPS Considering GPS Positioning Accuracy
- Author
-
Taketomi, Takafumi, Kume, Hideyuki, Sato, Tomokazu, Yokoya, Naokazu, Taketomi, Takafumi, Kume, Hideyuki, Sato, Tomokazu, and Yokoya, Naokazu
- Abstract
This paper proposes a method for estimating extrinsic camera parameters using video images and position data acquired by GPS. In conventional methods, the accuracy of the estimated camera position largely depends on the accuracy of GPS positioning data because they assume that GPS position error is very small or normally distributed. However, the actual error of GPS positioning easily grows to the 10m level and the distribution of these errors is changed depending on satellite positions and conditions of the environment. In order to achieve more accurate camera positioning in outdoor environments, in this study, we have employed a simple assumption that true GPS position exists within a certain range from the observed GPS position and the size of the range depends on the GPS positioning accuracy. Concretely, the proposed method estimates camera parameters by minimizing an energy function that is defined by using the reprojection error and the penalty term for GPS positioning., ICPR 2010 : 20th International Conference on Pattern Recognition , Aug 23-26, 2010 , Istanbul, Turkey
- Published
- 2023
16. AR Cultural Heritage Reconstruction Based on Feature Landmark Database Constructed by Using Omnidirectional Range Sensor
- Author
-
Taketomi, Takafumi, Sato, Tomokazu, Yokoya, Naokazu, Taketomi, Takafumi, Sato, Tomokazu, and Yokoya, Naokazu
- Abstract
This paper describes an application of augmented reality (AR) techniques to virtual cultural heritage reconstruction on the real sites of defunct constructs. To realize AR-based cultural heritage reconstruction, extrinsic camera parameter estimation is required for geometric registration of real and virtual worlds. To estimate extrinsic camera parameters, we use a pre-constructed feature landmark database of the target environment. Conventionally, a feature landmark database has been constructed in a large-scale environment using a structure -from-motion technique for omnidirectional image sequences. However, the accuracy of estimated camera parameters is insufficient for specific applications like AR-based cultural heritage reconstruction, which needs to overlay CG objects at the position close to the user’s viewpoint. This is due to the difficulty in compensation of the appearance change of close landmarks only from the sparse 3-D information obtained by structure-from-motion. In this paper, visual patterns of landmarks are compensated for by considering local shapes obtained by omnidirectional range finder to find corresponding landmarks existing close to the user. By using these landmarks with local shapes, accurate geometric registration is achieved for AR sightseeing in historic sites., ACCV 2010 : 10th Asian Conference on Computer Vision , Nov 8-12, 2010 , Queenstown, New Zealand
- Published
- 2023
17. Augmented Reality Marker Hiding with Texture Deformation
- Author
-
Kawai, Norihiko, Sato, Tomokazu, Nakashima, Yuta, Yokoya, Naokazu, Kawai, Norihiko, Sato, Tomokazu, Nakashima, Yuta, and Yokoya, Naokazu
- Abstract
Augmented reality (AR) marker hiding is a technique to visually remove AR markers in a real-time video stream. A conventional approach transforms a background image with a homography matrix calculated on the basis of a camera pose and overlays the transformed image on an AR marker region in a real-time frame, assuming that the AR marker is on a planar surface. However, this approach may cause discontinuities in textures around the boundary between the marker and its surrounding area when the planar surface assumption is not satisfied. This paper proposes a method for AR marker hiding without discontinuities around texture boundaries even under nonplanar background geometry without measuring it. For doing this, our method estimates the dense motion in the marker's background by analyzing the motion of sparse feature points around it, together with a smooth motion assumption, and deforms the background image according to it. Our experiments demonstrate the effectiveness of the proposed method in various environments with different background geometries and textures.
- Published
- 2023
18. Video summarization using textual descriptions for authoring video blogs
- Author
-
Otani, Mayu, Nakashima, Yuta, Sato, Tomokazu, Yokoya, Naokazu, Otani, Mayu, Nakashima, Yuta, Sato, Tomokazu, and Yokoya, Naokazu
- Abstract
Authoring video blogs requires a video editing process, which is cumbersome for ordinary users. Video summarization can automate this process by extracting important segments from original videos. Because bloggers typically have certain stories for their blog posts, video summaries of a blog post should take the author’s intentions into account. However, most prior works address video summarization by mining patterns from the original videos without considering the blog author’s intentions. To generate a video summary that reflects the blog author’s intention, we focus on supporting texts in video blog posts and present a text-based method, in which the supporting text serves as a prior to the video summary. Given video and text that describe scenes of interest, our method segments videos and assigns to each video segment its priority in the summary based on its relevance to the input text. Our method then selects a subset of segments with content that is similar to the input text. Accordingly, our method produces different video summaries from the same set of videos, depending on the input text. We evaluated summaries generated from both blog viewers’ and authors’ perspectives in a user study. Experimental results demonstrate the advantages to the proposed text-based method for video blog authoring.
- Published
- 2023
19. Efficient surface completion using principal curvature and its evaluation
- Author
-
Kawai, Norihiko, Sato, Tomokazu, Yokoya, Naokazu, Kawai, Norihiko, Sato, Tomokazu, and Yokoya, Naokazu
- Abstract
Surface completion is a technique for filling missing regions in 3D models measured by range scanners and videos. Conventionally, although missing regions were filled with the similar shape in a model, the completion process was fairly inefficient because the whole region in the model was searched for the similar shape. In this paper, the completion is efficiently performed using principal curvatures of local shape. In experiments, the effectiveness of the proposed method is successfully verified with subjective evaluation. In addition, the quantitative evaluation which has not been in the literature is newly performed., ICIP 2009 : 16th IEEE International Conference on Image Processing , Nov 7-10, 2009 , Cairo, Egypt
- Published
- 2023
20. Surface completion of shape and texture based on energy minimization
- Author
-
Kawai, Norihiko, Zakhor, Avideh, Sato, Tomokazu, Yokoya, Naokazu, Kawai, Norihiko, Zakhor, Avideh, Sato, Tomokazu, and Yokoya, Naokazu
- Abstract
In this paper, we propose a novel surface completion method to generate plausible shapes and textures for missing regions of 3D models. The missing regions are filled in by minimizing two energy functions for shape and texture, which are both based on similarities between the missing region and the rest of the object; in doing so, we take into account the positive correlation between shape and texture. We demonstrate the effectiveness of the proposed method experimentally by applying it to two models., ICIP2011 : 18th IEEE International Conference on Image Processing , Sep 11-14, 2011 , Brussels, Belgium
- Published
- 2023
21. Image resolution enhancement based on novel view synthesis
- Author
-
Hayashi, Yusuke, Kawai, Norihiko, Sato, Tomokazu, Yokoya, Naokazu, Hayashi, Yusuke, Kawai, Norihiko, Sato, Tomokazu, and Yokoya, Naokazu
- Abstract
This paper proposes an example-based method to increase the resolution of a low-resolution image. In the proposed method, we generate example images by a novel view synthesis technique using 3D geometry reconstruction and camera pose estimation from a video or images capturing the same scene. We then increase the resolution by minimizing an energy function by searching for the optimal example from the generated example images. The proposed method has less limitations on camera positions and geometry of the target scene than those in conventional methods. Experiments demonstrate the effectiveness of the proposed method by qualitatively comparing the results of the proposed and conventional methods., ICIP 2015 : 22th IEEE International Conference on Image Processing , Sep 27-30, 2015 , Quebec, Canada
- Published
- 2023
22. Image Inpainting Considering Brightness Change and Spatial Locality of Textures and Its Evaluation
- Author
-
Kawai, Norihiko, Sato, Tomokazu, Yokoya, Naokazu, Kawai, Norihiko, Sato, Tomokazu, and Yokoya, Naokazu
- Abstract
Image inpainting techniques have been widely investigated to remove undesired objects in an image. Conventionally, missing parts in an image are completed by optimizing the objective function using pattern similarity. However, unnatural textures are easily generated due to two factors: (1) available samples in the image are quite limited, and (2) pattern similarity is one of the required conditions but is not sufficient for reproducing natural textures. In this paper, we propose a new energy function based on the pattern similarity considering brightness changes of sample textures (for (1)) and introducing spatial locality as an additional constraint (for (2)). The effectiveness of the proposed method is successfully demonstrated by qualitative and quantitative evaluation. Furthermore, the evaluation methods used in much inpainting research are discussed., PSIVT 2009 : 3rd Pacific-Rim Symposium on Image and Video Technology , Jan 13-16, 2009 , Tokyo, Japan
- Published
- 2023
23. 3-D modeling of an outdoor scene by multi-baseline stereo using a long sequence of images
- Author
-
Sato, Tomokazu, Kanbara, Masayuki, Yokoya, Naokazu, Takemura, Haruo, Sato, Tomokazu, Kanbara, Masayuki, Yokoya, Naokazu, and Takemura, Haruo
- Abstract
Three-dimensional (3-D) models of outdoor scenes are widely used for object recognition, navigation, mixed reality, and so on. Because such models are often made manually with high costs, automatic 3-D modeling has been investigated. A 3-D model is usually generated by using a stereo method. However such approaches cannot use several hundreds images together for dense depth estimation because it is difficult to accurately calibrate a large number of cameras. In this paper we propose a 3-D modeling method that first estimates extrinsic camera. parameters of a monocular image sequence captured by a moving video camera, and then reconstructs a 3-D model of a scene. We can acquire a 3-D model of an outdoor scene accurately by using several hundreds input images., ICPR 2002 : 16th International Conference on Pattern Recognition , Aug 11-15, 2002 , Quebec, Canada
- Published
- 2023
24. Evaluation of image processing algorithms on vehicle safety system based on free-viewpoint image rendering
- Author
-
Oko, Akitaka, Sato, Tomokazu, Kume, Hideyuki, Machida, Takashi, Yokoya, Naokazu, Oko, Akitaka, Sato, Tomokazu, Kume, Hideyuki, Machida, Takashi, and Yokoya, Naokazu
- Abstract
Development of algorithms for vehicle safety systems, which support safety driving, takes a long period of time and a huge cost because it requires an evaluation stage where huge combinations of possible driving situations should be evaluated by using videos which are captured beforehand in real environments. In this paper, we address this problem by using free viewpoint images instead of the real images. More concretely, we generate free-viewpoint images from a combination of a 3D point cloud measured by laser scanners and an omni-directional image sequence acquired in a real environment. We basically rely on the 3D point cloud for geometrically correct virtual viewpoint images. In order to remove the holes caused by the unmeasured region of the 3D point cloud and to remove false silhouettes in surface reconstruction, we have developed a technique of free-viewpoint image generation that uses both a 3D point cloud and depth information extracted from images. In the experiments, we have evaluated our framework with a white line detection algorithm and experimental results have shown the applicability of free-viewpoint images for evaluation of algorithms., IV'14 : 2014 IEEE Intelligent Vehicles Symposium Proceedings , Jun 8-11, 2014 , Dearborn, MI, USA
- Published
- 2023
25. Real-time camera position and posture estimation using a feature landmark database with priorities
- Author
-
Taketomi, Takafumi, Sato, Tomokazu, Yokoya, Naokazu, Taketomi, Takafumi, Sato, Tomokazu, and Yokoya, Naokazu
- Abstract
This paper proposes a new method of diminished reality which removes AR markers from a user's view image in real time. To achieve natural marker hiding, assuming that an area around a marker is locally planar, the marker area in the first frame is inpainted using the rectified image to achieve high-quality inpainting. The unique inpainted texture is overlaid on the marker region in each frame according to camera motion for geometric consistency. Both global and local luminance changes around the marker are separately detected and reflected to the inpainted texture for photometric consistency., ICPR 2008 : 19th International Conference on Pattern Recognition , Dec 8-11, 2008 , Tampa, FL, USA
- Published
- 2023
26. Camera position and posture estimation from still image using feature landmark database
- Author
-
Sato, Tomokazu, Nishiumi, Yoshiyuki, Susuki, Mitsutaka, Nakagawa, Tomoka, Yokoya, Naokazu, Sato, Tomokazu, Nishiumi, Yoshiyuki, Susuki, Mitsutaka, Nakagawa, Tomoka, and Yokoya, Naokazu
- Abstract
Several human navigation services are currently available on the cellular phones that uses embedded GPS and 2-D map. However, 2-D map based human navigation is not always easy to understand for users because that is not intuitive. In order to realize more intuitive human navigation, AR (Augmented Reality) based navigation where guiding information is overlaid in the real image is expected to be the next generation navigation system. For AR navigation, the key problem is how to acquire the accurate position and posture of the embedded camera on the cellular phone. Many researchers have intensively tackled to the camera parameter estimation problem for AR in recent years. However, most of these methods cannot be applied to the current mobile devices because they are designed to treat video sequence where temporal information like camera parameter of the previous frame is known. In this research, we propose a novel method that estimates camera parameters of single input image using SIFT features and voting scheme., SICE 2008 : International Conference on Instrumentation, Control and Information Technology , Aug 20-22, 2008 , Chofu, Tokyo, Japan
- Published
- 2023
27. Immersive telepresence system using high-resolution omnidirectional movies and a locomotion interface
- Author
-
Ikeda, Sei, Sato, Tomokazu, Kanbara, Masayuki, Yokoya, Naokazu, Ikeda, Sei, Sato, Tomokazu, Kanbara, Masayuki, and Yokoya, Naokazu
- Abstract
Technology that enables users to experience a remote site virtually is called telepresence. A telepresence system using real environment images is expected to be used in the field of entertainment, medicine, education and so on. This paper describes a novel telepresence system which enables users to walk through a photorealistic virtualized environment by actual walking. To realize such a system, a wide-angle high-resolution movie is projected on an immersive multi-screen display to present users the virtualized environments and a treadmill is controlled according to detected user's locomotion. In this study, we use an omnidirectional multi-camera system to acquire images real outdoor scene. The proposed system provides users with rich sense of walking in a remote site., Stereoscopic Displays and Virtual Reality Systems XI , Jan 19-22, 2004 , San Jose, CA, USA
- Published
- 2023
28. Indirect augmented reality considering real-world illumination change
- Author
-
Okura, Fumio, Akaguma, Takayuki, Sato, Tomokazu, Yokoya, Naokazu, Okura, Fumio, Akaguma, Takayuki, Sato, Tomokazu, and Yokoya, Naokazu
- Abstract
Indirect augmented reality (IAR) utilizes pre-captured omnidirectional images and offline superimposition of virtual objects for achieving high-quality geometric and photometric registration. Meanwhile, IAR causes inconsistency between the real world and the pre-captured image. This paper describes the first-ever study focusing on the temporal inconsistency issue in IAR. We propose a novel IAR system which reflects real-world illumination change by selecting an appropriate image from a set of images pre-captured under various illumination. Results of a public experiment show that the proposed system can improve the realism in IAR., ISMAR 2014 : IEEE and ACM International Symposium on Mixed and Augmented Reality , Sep 10-12, 2014 , Munich, Germany
- Published
- 2023
29. Panoramic Movie Generation Using an Omnidirectional Multi-camera System for Telepresence
- Author
-
Ikeda, Sei, Sato, Tomokazu, Yokoya, Naokazu, Ikeda, Sei, Sato, Tomokazu, and Yokoya, Naokazu
- Abstract
Telepresence systems using an omnidirectional image sensor enable us to experience remote sites with rich sensation. An omnidirectional multi-camera system is more useful to acquire high-resolution omnidirectional images of outdoor scenes than a monocular camera system. However, exact calibration of the camera system is necessary to generate a panoramic movie. In this paper, we describe panoramic movie generation method based on geometric and photometric calibration of the omnidirectional multi-camera system. Some prototypes of telepresence system are also shown as applications of generated panoramic movies. These high-resolution telepresence systems have been proven to enable us to experience remote sites with high presence., SCIA 2003 : 13th Scandinavian Conference on Image Analysis , Jun 29-Jul 2, 2003 , Halmstad, Sweden
- Published
- 2023
30. Free-viewpoint AR human-motion reenactment based on a single RGB-D video stream
- Author
-
Fabian, Lorenzo Dayrit, Nakashima, Yuta, Sato, Tomokazu, Yokoya, Naokazu, Fabian, Lorenzo Dayrit, Nakashima, Yuta, Sato, Tomokazu, and Yokoya, Naokazu
- Abstract
When observing a person (an actor) performing or demonstrating some activity for the purpose of learning the action, it is best for the viewers to be present at the same time and place as the actor. Otherwise, a video must be recorded. However, conventional video only provides two-dimensional (2D) motion, which lacks the original third dimension of motion. In the presence of some ambiguity, it may be hard for the viewer to comprehend the action with only two dimensions, making it harder to learn the action. This paper proposes an augmented reality system to reenact such actions at any time the viewer wants, in order to aid comprehension of 3D motion. In the proposed system, a user first captures the actor's motion and appearance, using a single RGB-D camera. Upon a viewer's request, our system displays the motion from an arbitrary viewpoint using a rough 3D model of the subject, made up of cylinders, and selecting the most appropriate textures based on the viewpoint and the subject's pose. We evaluate the usefulness of the system and the quality of the displayed images by user study., ICME 2014 : IEEE International Conference on Multimedia and Expo , Jul 14-18, 2014 , Chengdu, China
- Published
- 2023
31. Calibration method for an omnidirectional multi-camera system
- Author
-
Ikeda, Sei, Sato, Tomokazu, Yokoya, Naokazu, Ikeda, Sei, Sato, Tomokazu, and Yokoya, Naokazu
- Abstract
Telepresence systems using an omnidirectional image sensor enable us to experience remote site. A omnidirectional multi-camera system is more useful to acquire outdoor scenes than a monocular camera system, because the multi-camera system can easily capture high-resolution omnidirectional images. However, exact calibration of the camera system is necessary to virtualize the real world accurately. In this paper, we describe a geometric and photometric camera calibration and a panorama movie generation method for the omnidirectional multi-camera system. In the geometric calibration, intrinsic and extrinsic parameters of each camera are estimated using a calibration board and a laser measurement system called total station. In the photometric calibration, the limb darkening and color balances among the cameras are corrected. The result of the calibration is used in the panorama movie generation. In experiments, we have actually calibrated the multi-camera system and have generated spherical panorama movies by using the estimated camera parameters. A telepresence system was prototyped in order to confirm that the panorama movie can be used for telepresence well. In addition, we have evaluated the discontinuity in generated panoramic images., Stereoscopic Displays and Virtual Reality Systems XII , Jan 21-23, 2003 , Santa Clara, CA, USA
- Published
- 2023
32. Teleoperation of mobile robots by generating augmented free-viewpoint images
- Author
-
Okura, Fumio, Ueda, Yuko, Sato, Tomokazu, Yokoya, Naokazu, Okura, Fumio, Ueda, Yuko, Sato, Tomokazu, and Yokoya, Naokazu
- Abstract
This paper proposes a teleoperation interface by which an operator can control a robot from freely configured viewpoints using realistic images of the physical world. The viewpoints generated by the proposed interface provide human operators with intuitive control using a head-mounted display and head tracker, and assist them to grasp the environment surrounding the robot. A state-of-the-art free-viewpoint image generation technique is employed to generate the scene presented to the operator. In addition, an augmented reality technique is used to superimpose a 3D model of the robot onto the generated scenes. Through evaluations under virtual and physical environments, we confirmed that the proposed interface improves the accuracy of teleoperation., IROS 2013 : IEEE/RSJ International Conference on Intelligent Robots and Systems , Nov 3-7, 2013 , Tokyo, Japan
- Published
- 2023
33. Position estimation of near point light sources using a clear hollow sphere
- Author
-
Aoto, Takahito, Taketomi, Takafumi, Sato, Tomokazu, Mukaigawa, Yasuhiro, Yokoya, Naokazu, Aoto, Takahito, Taketomi, Takafumi, Sato, Tomokazu, Mukaigawa, Yasuhiro, and Yokoya, Naokazu
- Abstract
We present a novel method for estimating 3-D positions of near light sources by using highlights on the outside and inside of a single clear hollow sphere. Conventionally, the positions of near light sources have been estimated by using observed highlights on multiple reference objects, e.g. mirror balls., ICPR 2012 : 21th International Conference on Pattern Recognition , Nov 11-15, 2012 , Tsukuba, Japan
- Published
- 2023
34. Augmented reality image generation with virtualized real objects using view-dependent texture and geometry
- Author
-
Nakashima, Yuta, Sato, Tomokazu, Uno, Yusuke, Yokoya, Naokazu, Kawai, Norihiko, Nakashima, Yuta, Sato, Tomokazu, Uno, Yusuke, Yokoya, Naokazu, and Kawai, Norihiko
- Abstract
Augmented reality (AR) images with virtualized real objects can be used for various applications. However, such AR image generation requires hand-crafted 3D models of that objects, which are usually not available. This paper proposes a view-dependent texture (VDT)- and view-dependent geometry (VDG)-based method for generating high quality AR images, which uses 3D models automatically reconstructed from multiple images. Since the quality of reconstructed 3D models is usually insufficient, the proposed method inflates the objects in the depth map as VDG to repair chipped object boundaries and assigns a color to each pixel based on VDT to reproduce the detail of the objects. Background pixel exposure due to inflation is suppressed by the use of the foreground region extracted from the input images. Our experimental results have demonstrated that the proposed method can successfully reduce above visual artifacts., ISMAR 2013 : IEEE and ACM International Symposium on Mixed and Augmented Reality , Oct 1-4, 2013 , Adelaide, Australia
- Published
- 2023
35. Web-Based Telepresence System Using Omni-directional Video Streams
- Author
-
Yamazawa, Kazumasa, Ishikawa, Tomoya, Sato, Tomokazu, Ikeda, Sei, Nakamura, Yutaka, Fujikawa, Kazutoshi, Sunahara, Hideki, Yokoya, Naokazu, Yamazawa, Kazumasa, Ishikawa, Tomoya, Sato, Tomokazu, Ikeda, Sei, Nakamura, Yutaka, Fujikawa, Kazutoshi, Sunahara, Hideki, and Yokoya, Naokazu
- Abstract
Recently, the telepresence which allows us to experience a remote site through a virtualized real world has been investigated. We have already proposed telepresence systems using omni-directional images which enable the user to look around a virtualized dynamic real scene. The system gives the feeling of high presence to the user. However, conventional systems are not networked but are implemented as stand-alone systems that use immersive displays. In this paper we propose a networked telepresence system which enables the user to see a virtualized real world easily in network environments. This paper describes the system implementation using real-time and stored omni-directional video streams as well as its experiments., PCM 2004 : Pacific Rim Conference on Multimedia , Nov 30-Dec 3, 2004 , Tokyo, Japan
- Published
- 2023
36. Surface completion by minimizing energy based on similarity of shape
- Author
-
Kawai, Norihiko, Sato, Tomokazu, Yokoya, Naokazu, Kawai, Norihiko, Sato, Tomokazu, and Yokoya, Naokazu
- Abstract
3D mesh models generated with range scanner or video images often have holes due to many occlusions by other objects and the object itself. This paper proposes a novel method to fill the missing parts in the incomplete models. The missing parts are filled by minimizing the energy function, which is defined based on similarity of local shape between the missing region and the rest of the object. The proposed method can generate complex and consistent shapes in the missing region. In the experiment, the effectiveness of the method is successfully demonstrated by applying it to complex shape objects with missing parts., ICIP 2008 : 15th IEEE International Conference on Image Processing , Oct 12-15, 2008 , San Diego, CA, USA
- Published
- 2023
37. Estimating Camera Position and Posture by Using Feature Landmark Database
- Author
-
Oe, Motoko, Sato, Tomokazu, Yokoya, Naokazu, Oe, Motoko, Sato, Tomokazu, and Yokoya, Naokazu
- Abstract
Estimating camera position and posture can be applied to the fields of augmented reality and robot navigation. In these fields, to obtain absolute position and posture of the camera, sensor-based methods using GPS and magnetic sensors and vision-based methods using input images from the camera have been investigated. However, sensor-based methods are difficult to synchronize the camera and sensors accurately, and usable environments are limited according to selection of sensors. On the other hand, vision-based methods need to allocate many artificial markers otherwise an estimation error will accumulate. Thus, it is difficult to use such methods in large and natural environments. This paper proposes a vision-based camera position and posture estimation method for large environments, which does not require sensors and artificial markers by detecting natural feature points from image sequences taken beforehand and using them as landmarks., SCIA 2005 : 14th Scandinavian Conference on Image Analysis , Jun 19-22, 2005 , Joensuu, Finland
- Published
- 2023
38. Interactive 3-D Modeling System Using a Hand-Held Video Camera
- Author
-
Fudono, Kenji, Sato, Tomokazu, Yokoya, Naokazu, Fudono, Kenji, Sato, Tomokazu, and Yokoya, Naokazu
- Abstract
Recently, a number of methods for 3-D modeling from images have been developed. However, the accuracy of a reconstructed model depends on camera positions and postures with which the images are obtained. In most of conventional methods, some skills for adequately controlling the camera movement are needed for users to obtain a good 3-D model. In this study, we propose an interactive 3-D modeling interface in which special skills are not required. This interface consists of “indication of camera movement” and “preview of reconstruction result.” In experiments for subjective evaluation, we verify the usefulness of the proposed 3D modeling interfaces., SCIA 2005 : 14th Scandinavian Conference on Image Analysis , Jun 19-22, 2005 , Joensuu, Finland
- Published
- 2023
39. From Image Inpainting to Diminished Reality
- Author
-
Kawai, Norihiko, Sato, Tomokazu, Yokoya, Naokazu, Kawai, Norihiko, Sato, Tomokazu, and Yokoya, Naokazu
- Abstract
Image inpainting, which removes undesired objects in a static image and fills in the missing regions with plausible textures, has been developed in the research fields of image processing. On the other hand, Diminished Reality (DR), which visually removes real objects from video images by filling in the missing regions with background textures in real time, is one of the growing topics in Virtual/Mixed Reality, and considered as the opposite of Augmented Reality. In this paper, we introduce the state-of-the-art of image inpainting methods and how to apply the image inpainting to diminished reality., VAMR 2014 : 6th International Conference Virtual, Augmented and Mixed Reality , Jun 22-27, 2014 , Heraklion, Crete, Greece
- Published
- 2023
40. Mobile AR using pre-captured omnidirectional images
- Author
-
Akaguma, Takayuki, Okura, Fumio, Sato, Tomokazu, Yokoya, Naokazu, Akaguma, Takayuki, Okura, Fumio, Sato, Tomokazu, and Yokoya, Naokazu
- Abstract
In the field of augmented reality (AR), geometric and photometric registration is routinely achieved in real time. However, real-time geometric registration often leads to misalignment (e.g., jitter and drift) due to the error from camera pose estimation. Due to limited resources on mobile devices, it is also difficult to implement state-of-the-art techniques for photometric registration on mobile AR systems. In order to solve these problems, we developed a mobile AR system in a significantly different way from conventional systems. In this system, captured omnidirectional images and virtual objects are registered geometrically and photometrically in an offline rendering process. The appropriate part of the prerendered omnidirectional AR image is shown to a user through a mobile device with online registration between the real world and the pre-captured image. In order to investigate the validity of our new framework for mobile AR, we conducted experiments using the prototype system on a real site in Todai-ji Temple, a famous world cultural heritage site in Japan., SA 2013 : SIGGRAPH Asia 2013 Symposium on Mobile Graphics and Interactive Applications , Nov 19-22, 2013 , Hong Kong
- Published
- 2023
41. 3D shape template generation from RGB-D images capturing a moving and deforming object
- Author
-
Takehara, Hikari, Nakashima, Yuta, Sato, Tomokazu, Yokoya, Naokazu, Takehara, Hikari, Nakashima, Yuta, Sato, Tomokazu, and Yokoya, Naokazu
- Abstract
Automatically reconstructing a 3D shape model of a nonrigid object using a sequence from a single commodity RGB-D sensor is a challenging problem. Some techniques use a 3D shape template of a target object; however, in order to generate the template automatically, the target object required to be stationary. Otherwise, a non-rigid ICP algorithm, which registers a pair of point clouds, can be used for reconstructing 3D geometry of a non-rigid object directly, but it often fails due to the ambiguity in point correspondences. This paper presents a method for generating a 3D shape template from a single RGB-D sequence. In order to reduce the ambiguity in point correspondences, our method leverages point trajectories obtained in the RGB images, which can be used for associating points in different point clouds. We demonstrate the capability of our method using deforming human bodies., EI 2016 : IS&T International Symposium on Electronic Imaging , Feb 14-18, 2016 , San Francisco , CA , USA
- Published
- 2023
42. Camera recovery of an omnidirectional multi-camera system using GPS positions
- Author
-
Ikeda, Sei, Sato, Tomokazu, Yokoya, Naokazu, Ikeda, Sei, Sato, Tomokazu, and Yokoya, Naokazu
- Abstract
This paper describes a novel method for estimating positions and postures of an omnidirectional multi-camera system from its multiple image sequences and sparse position data acquired by GPS. The proposed method is based on a structure-from-motion technique which is enhanced by using multiple image sequences as well as GPS position data. Moreover, the position data are also used to remove mis-tracked features. The proposed method allows us to estimate position and posture without accumulative errors and annoying effects due to moving objects in outdoor environments. The validity of the method is demonstrated through experiments using both synthetic and real outdoor scenes., First Korea-Japan Joint Workshop on Pattern Recognition , Nov 23-24, 2006 , Jeju, Korea
- Published
- 2023
43. Generation of an Omnidirectional Video without Invisible Areas Using Image Inpainting
- Author
-
Kawai, Norihiko, Machikita, Kotaro, Sato, Tomokazu, Yokoya, Naokazu, Kawai, Norihiko, Machikita, Kotaro, Sato, Tomokazu, and Yokoya, Naokazu
- Abstract
Omnidirectional cameras usually cannot capture the entire direction of view due to a blind side. Thus, such an invisible part decreases realistic sensation in a telepresence system. In this study, an omnidirectional video without invisible areas is generated by filling in the missing region using an image inpainting technique for highly realistic sensation in telepresence. This paper proposes a new method that successfully inpaints a missing region by compensating for the change in appearance of textures caused by the camera motion and determining a searching area for similar textures considering the camera motion and the shape of the scene around the missing region. In experiments, the effectiveness of the proposed method is demonstrated by inpainting missing regions in a real image sequence captured with an omnidirectional camera and generating an omnidirectional video without invisible areas., ACCV 2009 : 9th Asian Conference on Computer Vision , Sep 23-27, 2009 , Xi'an, China
- Published
- 2023
44. High-resolution panoramic movie generation from video streams acquired by an omnidirectional multi-camera system
- Author
-
Ikeda, Sei, Sato, Tomokazu, Yokoya, Naokazu, Ikeda, Sei, Sato, Tomokazu, and Yokoya, Naokazu
- Abstract
Telepresence systems using an omnidirectional image sensor enable us to experience remote sites with rich sensation. An omnidirectional multi-camera system is more useful for acquiring high-resolution omnidirectional images of outdoor scenes than a monocular camera system. However, exact calibration of the camera system is necessary to generate a panoramic movie. In this paper, we describe a panoramic movie generation method based on geometric and photometric calibration of the omnidirectional multi-camera system. A prototype of a telepresence system is also shown as applications of generated panoramic movies. This high-resolution telepresence system has been proven to enable us to experience remote sites with rich presence., MFI 2003 : International Conference on Multisensor Fusion and Integration for Intelligent Systems , Aug 1, 2003 , Tokyo, Japan
- Published
- 2023
45. High-resolution video mosaicing for documents and photos by estimating camera motion
- Author
-
Sato, Tomokazu, Ikeda, Sei, Kanbara, Masayuki, Iketani, Akihiko, Nakajima, Noboru, Yokoya, Naokazu, Yamada, Keiji, Sato, Tomokazu, Ikeda, Sei, Kanbara, Masayuki, Iketani, Akihiko, Nakajima, Noboru, Yokoya, Naokazu, and Yamada, Keiji
- Abstract
Recently, document and photograph digitization from a paper is very important for digital archiving and personal data transmission through the internet. Though many people wish to digitize documents on a paper easily, now heavy and large image scanners are required to obtain high quality digitization. To realize easy and high quality digitization of documents and photographs, we propose a novel digitization method that uses a movie captured by a hand-held camera. In our method, first, 6-DOF(Degree Of Freedom) position and posture parameters of the mobile camera are estimated in each frame by tracking image features automatically. Next, re-appearing feature points in the image sequence are detected and stitched for minimizing accumulated estimation errors. Finally, all the images are merged as a high-resolution mosaic image using the optimized parameters. Experiments have successfully demonstrated the feasibility of the proposed method. Our prototype system can acquire initial estimates of extrinsic camera parameters in real-time with capturing images., Computational Imaging II , Jan 18, 2004 , San Jose, CA, USA
- Published
- 2023
46. Immersive telepresence system with a locomotion interface using high-resolution omnidirectional videos
- Author
-
Ikeda, Sei, Sato, Tomokazu, Kanbara, Masayuki, Yokoya, Naokazu, Ikeda, Sei, Sato, Tomokazu, Kanbara, Masayuki, and Yokoya, Naokazu
- Abstract
This paper describes a novel telepresence system which enables users to walk through a photorealistic virtualized environment by actual walking. In this system, a wide-angle high-resolution video is projected on an immersive multi- screen display and a treadmill is controlled according to user’s locomotion. Calibration and motion estimation of a camera system are performed to enhance quality of pre- sented images. The proposed system provides users with rich sense of walking in a remote site., MVA 2005 : IAPR Conference on Machine Vision Applications , May 16-18, 2005 , Tsukuba, Japan
- Published
- 2023
47. An immersive telepresence system with a locomotion interface using high-resolution omnidirectional movies
- Author
-
Ikeda, Sei, Sato, Tomokazu, Kanbara, Masayuki, Yokoya, Naokazu, Ikeda, Sei, Sato, Tomokazu, Kanbara, Masayuki, and Yokoya, Naokazu
- Abstract
This paper describes a novel telepresence system which enables users to walk through a photorealistic virtualized environment by actual walking. To realize such a system, a wide-angle high-resolution movie is projected on an immersive multi-screen display to present users the virtualized environments and a treadmill is controlled according to detected user's locomotion. In this study, we use an omnidirectional multi-camera system to acquire images of a real outdoor scene. The proposed system provides users with rich sense of walking in a remote site., ICPR 2004 : 17th International Conference on Pattern Recognition , Aug 23-26, 2004 , Cambridge, UK
- Published
- 2023
48. Diminished reality considering background structures
- Author
-
Kawai, Norihiko, Sato, Tomokazu, Yokoya, Naokazu, Kawai, Norihiko, Sato, Tomokazu, and Yokoya, Naokazu
- Abstract
This paper proposes a new diminished reality method for 3D scenes considering background structures. Most conventional methods using image inpainting assumes that the background around a target object is almost planar. In this study, approximating the background structure by the combination of local planes, perspective distortion of texture is corrected and searching area is limited for improving the quality of image inpainting. The temporal coherence of texture is preserved using the estimated structures and camera pose estimated by Visual-SLAM., ISMAR 2013 : IEEE and ACM International Symposium on Mixed and Augmented Reality , Oct 1-4, 2013 , Adelaide, Australia
- Published
- 2023
49. AR marker hiding based on image inpainting and reflection of illumination changes
- Author
-
Kawai, Norihiko, Yamasaki, Masayoshi, Sato, Tomokazu, Yokoya, Naokazu, Kawai, Norihiko, Yamasaki, Masayoshi, Sato, Tomokazu, and Yokoya, Naokazu
- Abstract
This paper proposes a new method of diminished reality which removes AR markers from a user's view image in real time. To achieve natural marker hiding, assuming that an area around a marker is locally planar, the marker area in the first frame is inpainted using the rectified image to achieve high-quality inpainting. The unique inpainted texture is overlaid on the marker region in each frame according to camera motion for geometric consistency. Both global and local luminance changes around the marker are separately detected and reflected to the inpainted texture for photometric consistency., ISMAR 2012: IEEE International Symposium on Mixed and Augmented Reality , Nov 5-8, 2012 , Atlanta, GA, USA
- Published
- 2023
50. Real-time outdoor pre-visualization method for videographers - real-time geometric registration using point-based model
- Author
-
Ikeda, Sei, Taketomi, Takafumi, Okumura, Bunyo, Sato, Tomokazu, Kanbara, Masayuki, Yokoya, Naokazu, Chihara, Kunihiro, Ikeda, Sei, Taketomi, Takafumi, Okumura, Bunyo, Sato, Tomokazu, Kanbara, Masayuki, Yokoya, Naokazu, and Chihara, Kunihiro
- Abstract
This paper describes a real-time pre-visualization method using augmented reality techniques for videographers. It enables them to test camera work without real actors in a real environment. As a substitute for real actors, virtual ones are superimposed on a live video in real-time according to a real camera motion and an illumination condition. The key technique of this method is real-time motion estimation of a camera, which can be applied to unknown complex environments including natural objects. In our method, geometric and photometric registration problems for such unknown environments are solved to realize the above visualization. A prototype system demonstrates availability of the previsualization method., ICME 2008 : IEEE International Conference on Multimedia & Expo , Jun 23-26, 2008 , Hannover, Germany
- Published
- 2023
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.