Advanced
Multi-scale 3D Panor ama Content Augmented System using Depth-map
Multi-scale 3D Panor ama Content Augmented System using Depth-map
Journal of Korea Multimedia Society. 2014. Jun, 17(6): 733-740
Copyright © 2014, Korea Multimedia Society
  • Received : April 21, 2014
  • Accepted : May 07, 2014
  • Published : June 28, 2014
Download
PDF
e-PUB
PubReader
PPT
Export by style
Share
Article
Author
Metrics
Cited by
TagCloud
About the Authors
Cheeyong Kim
Dept. of Visual Information Engineering, Dong-eui University
Eung-Kon Kim
Dept. of Computer Engineering, Sunchon National University
Jong-Chan Kim
Dept. of Computer Engineering, Sunchon National University

Abstract
With the development and spread of 3D display, users can easily experience an augmented reality with 3D features. Therefore, the demand for content of an augmented reality is exponentially growing in various fields. A traditional augmented reality environment was generally created by CG(Computer Graphics) modelling production tools. However, this method takes too much time and efforts to create an augmented environment. To create an augmented environment similar to the real world, everything in the real world should be measured, gone through modeling, and located in an augmented environment. But the time and efforts spent in the creation don't produce the same environment as the real world, making it hard for users to feel the sense of reality. In this study, multi-scale 3D panorama content augmented system is suggested by using a depth-map. By finding matching features from images to add 3D features to an augmented environment, a depth-map is derived and embodied as panorama, producing high-quality augmented content system with a sense of reality. With this study, limits of 2D panorama technologies will be overcome and a sense of reality and immersion will be provided to users with a natural navigation.
Keywords
1. INTRODUCTION
3D content system is a technology which helps users accept a VR(Virtual Reality) as reality through 3D visual information and feel as if they are in the field, which are utilized in various fields such as education, advertising, home shopping, military training, and mock labs and medical industry. This kind of computer display system significantly contributes to the development of user interface, real-time object recognition, cuttingedge systems for the disabled. Various studies on 3D content including information about a sense of 3D effect and immersion are carried out home and abroad, increasing fields utilizing them [1 , 2] .
The typical technology is a image-based modeling and rendering which establish 3D augmented environment based on images. This method lacks a sense of reality since it provides realistic images of the environment or objects at arbitrary times based on information from images taken at actual environments. However, it has merits in that it doesn't need a traditional 3D modeling process, which reduces costs in th establishment and a production speed of printing images. To overcome these problems, studies on 3D image production through image synthesis and processing are carried out actively home and abroad [3 - 7] .
In this study, a multi-scale 3D panorama content augmented system using a depth-map of images is suggested. By finding matching features from images to add 3D features to an augmented environment, a depth-map is derived and embodied as panorama, producing high-quality augmented content system with a sense of reality. This system adds a high-quality panorama technology to userfriendly technology, providing 3D panorama contents which can be intuitively and easily manipulated, thereby improving a sense of immersion and reality of users.
2. RELATED WORKS
- 2.1 Panorama VR
Panorama VR is to show nearby sight at one's location. it gives an effect just like he feels in the place and it can reenact at certain days. To acquire this effect, panoramic technique is used. Images of real environment taken in 360° are called omni images or omnidirectional images. And images transformed them into cylinder forms, circular or rectangular forms are called panoramic images [8 - 10] .
To establish panorama VR, two steps are needed. The first step is to acquire panoramic images and the second step is to display them by panorama VR viewer. When users designate a certain gaze direction after receiving panoramic images, the panorama VR viewer shows the image seen from the direction by producing it from panoramic images in real time. Since panorama VR is widely used in the Internet, various kinds of panorama VR viewers have appeared. The most currently used viewers are Apple's QuickTime Player, Macromedia's Shockwave Player,Anything 3D's viewer, iSeeMedia's viewer, Zoomify's viewer, iPIX's viewer, and MG System's Viewer. Fig 1 . shows how to produce panoramic VR using a cylinder.
PPT Slide
Lager Image
Panorama VR process step.
- 2.2 Software method of production of panoramic images
Image mosaic is a method for matching several images and connecting them by general camera. First, set the starting point of photo-taking. And then rotate horizontally by overlapping 10%-50% of images and take 12-24 pictures according to the degrees of overlapping. Connect the overlapping parts to make a one piece of panorama, which is called a stitching process. For images to configure a cylinder by matching with neighboring ones, each image will be bent. Overlapped pictures should be bent in the same angle to fit perfectly. Each picture image is a flat rectangular image first, but to configure a cylinder its corner should be bent. The degrees of bending is different according to a filed of view of a used lens, and the images which become the cylinder part are more distorted. And when playing them next, the images will become flat as original. A final source image will have different dimension and ratio according to each images filed of view. The bigger the filed of view, the higher the height becomes compared to area. In other words, the smaller the filed of view, the larger the panorama. Pictures completing stitching have different tops and bottoms in length since they seem to connect pictures cut long. And then the blending process which make them a panorama picture by connecting them smoothly is car ried out. The process is shown in Fig 2 .
PPT Slide
Lager Image
Image blending step.
To produce high-quality panoramic images effectively, the exact camera direction of each input image should be found out. To do so, after finding the relative camera rotations of neighboring images through image matching, the camera direction is calculated by adding them in order. This approach has demerits in that errors occurring in image matching are accumulated and final panoramic images have gaps or overlays [10 - 13] .
3. 3D PANORAMA GENERATION ALGORITHM
Let x =( x , y ) which is a point from projection of X =( X , Y , Z ), a point in 3D(dimensional) space that is taken by lens distortion free camera, to 2D image, the relationship between X and x is modelled as follow [14] .
PPT Slide
Lager Image
PPT Slide
Lager Image
u is a vector that represents x as Homogeneous Coordinates. K , as a Camera Calibration Matrix, shows the relationship between 3D point on camera coordinates and 2D point on image within which 3D point is projected. R and t , that represent Rotation Matrix and Translation Vector each, indicate the transformation relationship between world coordinates and camera coordinates. As we call parameters included in vector K as Internal Parameter of camera, f is Focal Length, Q is Aspect Ratio of pixel, S is Skew and ( Xc , Yc ) is Principal Point.
PPT Slide
Lager Image
Rotation Matrix and Translation Vector of camera before rotating are R 0 and t 0 respectively and those of camera after rotating are R 1 and t 1 . When conforming projection center to zero point of world coordinates for convenience's sake, it becomes t 0 = t 1 =0. If homogeneous vector of a point that is from the projection of certain point X at 3 dimensional space to I 0 (original image) and I 1 (rotated image) is each u 0 and u 1 , then.
That is, as pixel correspondence relationship between two images that are from camera rotation, which is presented by a 3 × 3 matrix H, we call it Planar Homography. Considering that corresponding image point is the same even though multiplying any real number except 0 to Homogeneous Vector, H also can be defined regardless of its scale. Hence, Planar Homography is defined 8 parameters, not 9 parameters. Actually, as the number of parameter that is included in K from H = KR 01 K -1 is 5 and the number of parameter that is in R 01 is 3, we can aware that independent parameters that are composed of H is 8. This Homography is called as 8-Parameter Homography.
Given two images, finding H which is pixel correspondence relationship between the two images is Image Registration. In general, H can be calculated with more than 4 correspondent points but it takes caution to get more accurate H. Therefore, we assume that camera satisfies this normalized model. Table 1 . shows the conditions of camera based on normalized model.
Camera condition of normalized model
PPT Slide
Lager Image
Camera condition of normalized model
Therefore, the number of independent parameter included in H on normalized model is 4(focal point distance f and three rotating angles within R 01 ), which we call as 4-Parameter Homography. General rotation matrix R is divided as follow.
PPT Slide
Lager Image
In formula(4), R ( ψ ), R ( θ ), R (∅) are respective matrix that present rotation transformation from each X, Y, Z axis and ψ , θ , ∅ are called Euler Angle. Given Planar Homography H in normalized model, calculating K and R are as shown below. Matrix K can be divided as follow.
PPT Slide
Lager Image
Substituting to formula(3)
PPT Slide
Lager Image
Therefore,
PPT Slide
Lager Image
As R is Orthogonal Matrix, it satisfies following formula. After calculating f from formula(8), it gets R by substituting to formula(7).
PPT Slide
Lager Image
Camera models described above do not consider distortion of its lens. As it utilizes wide angle lens to get 360 angle panorama video, it is necessary to consider lens distortion to apply algorithm to wider area. ( xd , yd ) is coordinates of pixel in original image(that is, image that contains lens distortion) and ( xu , yu ) is corresponding coordinates of pixel in image that removes lens distortion. As ( xc , yc ) is the center coordinates, given ( xd , yd ), ( xu , yu ) is calculated as follow.
PPT Slide
Lager Image
In this moment, k 1 , k 2 are radial distortion coefficients. In common,
PPT Slide
Lager Image
To the contrary when ( xu , yu ) is given, ( xd , yd ) is as follow.
PPT Slide
Lager Image
In formula (12),
PPT Slide
Lager Image
PPT Slide
Lager Image
PPT Slide
Lager Image
4. MULTI-SCALE 3D PANORAMA CONTENT AUGMENTED SYSTEM USING DEPTHMAP
Multi-scale 3D panorama content augmented system using depth map produces one panoramic image by analyzing imput images of multi-scale images. The image stitching technique is a method for making images taken in certain angles as a stretched out form without joints. Multiple images acquired from more than one cameras are produced as one panoramic image without joints through a image synthesis process. Since cameras used to acquire a panoramic image are set in different directions and locations, overlapping parts of images should be found out and geometrical adjustments should be carried out to produce one panoramic image without joints.
Multi-scale 3D panorama augmented reality content system produces a depth-map between images internally to produce panoramic images. To derive a depth-map, median filters are applied to remove noises from images. Edges are detected through the derivation of canny edges to input images through preprocessing, and vanishing lines are detected through Hough transformation. Through the intersection points of the detected vanishing lines, vanishing points are designated. Since the depth step of the vanishing points should be different depending on the locations of images, the range of locations of corresponding vanishing points between images and vanishing points are defined. By setting the defined range as standard, a standard depth step on a base level is produced, By linear interpolation of neighboring vanishing points, a depth-map is produced.
The production of a depth-map comprises three steps. The first step, the detection of edge, reduces noise as a preprocessing for input images by applying median filters and derives geometric features by applying canny edge. The second step, the production of vanishing lines and points, derives straight lines through Hough transformation, calculates the intersection points of the straight lines, and estimates the locations of vanishing points through the rage of locations of defined vanishing points. The final step, the estimation of vanishing point locations and production of a depth-map, estimates the locations of vanishing points through the range of locations of vanishing points, sets the steps of depth, and finally produces a depth-map based on the locations of vanishing points.
To produce multi-scale panorama image, the relationships between images are set internally. Multi-scale panorama image is produced through the blending process which makes input images into panoramic pictures. According to the change of image scale, the exact corresponding locations with neighboring images are automatically calculated, securing the exactness within sub pixels and stitching the images. Fig 3 . shows a general configuration map of multi-scale 3D panorama augmented system using a depth-map suggested in this study. Fig 4 . shows the steps of stitching images derived from a depth-map, producing depthmap panorama, and embodying multi-scale 3D panorama augmented system.
PPT Slide
Lager Image
Multi-scale 3D panorama augmented system configuration map.
PPT Slide
Lager Image
Embodiment of multi-scale 3D panorama augmented system using depth-map.
5. CONCLUSION AND FUTURE WORKS
In this study, features matching images found out to add a sense of 3D effect to the augmented environment and a depth-map is used to embody 3D panorama, producing the augmented system with a sense of reality. This system will overcome the limits of 2D panorama technologies and provide users with a natural navigation, providing an improved sense of reality and immersion. Multi-scale 3D panorama augmented system using depth-map makes users feel as if they are at the location which photo-taking is taken and seeing the surroundings, in other words of a sense of 3D effect. Multi-scale 3D panorama augmented system provides highquality 3D images with a sense of reality which has a free vision of close objects and distant backgrounds. This suggests a possibility in that it can be utilized in establishing augmented reality system based on images, attracting a attention from many people.
Further studies will be carried out based on computer vision technologies including image processing and interpretation technologies and computer graphics configuring media images with a sense of reality from the interpreted images to commercialize tourist attractions tours with a sense of reality for interaction with users and embodied images and high-quality digital content of public relations office.
BIO
KIM Cheeyong
He received the B.S. degree, M.S. degree and Ph.D. degree from Inje University, Republic of Korea in 1991, 1994 and 2000 respectively. He is visiting professor at Oxford University in 2007, and visiting professor of the Digital Clothing Center at Seoul National University in 2012. Currently, he is a professor of Department of Visual Information Engineering at Dong-eui University. Prof. Kim held private exhibitions four times both in Seoul and China and joined over 210 international group exhibitions. With his deep interest in 3D Animation, Fractal & Chaos Design, Computational Simulation, and 3D virtual fashion fitting system.
Eung-Kon Kim
He received B.E. degree Chosun University in 1980, his M.E. degree from depar-tment of computer engineering, Hanyang University in 1986, his Ph.D degree form depart-ment of computer engineering, Chosun University in 1994, In 2011, he joined department of computer science, the Sunchon National University, where he is a professor. His current research interests are image processing, AR, computer graphics, multimedia and HCI.
Jong-Chan Kim
He received his B.S. degree from Sunchon National University in 2000, his M.S degree from department of computer science, Sunchon National University in 2002, his Ph.D. degree from department of computer science, Sunchon National University in 2007. His current research interests are image processing, computer graphics, AR, games and HCI.
References
Burschka D. , Cobzas D. , Dodds Z. , Hager G. , Jagersand M. , Yerex K. 2003 "Recent Methods for Image-Based Modeling and Rendering" IEEE Virtual Reality 2003 Tutorial 1 1 - 85
Shum H.K. , Kang S.B. , Chan S.C. 2003 "Survey of Image-based Representations and Compression Techniques" IEEE Transaction On Circuits and Systems for Video Technology 13 (11) 1020 - 1037    DOI : 10.1109/TCSVT.2003.817360
Shelton B.E. , Hedley N.R. 2002 "Using Augmented Reality for Teaching Earth-Sun Relationships to Undergraduate Geography Students" Proceeding of First IEEE International Augmented Reality Toolkit Workshop 1 - 8
Yan Z. , Ma X. , Tao S. 2009 "The Virtual Display Study on 3D Panorama in Tourist Areas-Take Shilin World Geopark as an Example" Proceeding of the International Symposium on Information Processing 229 - 232
Chun S.M. , Choi J.Y. , Kim S.H. , Cho Y.C. , Park K.S. 2013 "A Study on Depth Map Quantization for Multiview Image Generation" Journal of Korean Society For Computer Game 26 (4) 219 - 226
Curti S. , Sirtori D. , Vella F. 2002 "3D Effect Generation from Monocular View" IEEE Proceeding of the 3D Data Processing Visualization and Transmission 550 - 553
Hassner T. , Basri R. 2006 "Example Based 3D Reconstruction from Single 2D Images" Proceeding of the Conference on Computer Vision and Pattern Recognition Workshop 15 - 22
Juan L. , Gwun O. 2010 "SURF Applied in Panorama Image Stitching" Proceeding of 2nd International Conference on Image Processing Theory Tools and Applications 495 - 499
Brown M. , Lowe D.G. 2007 "Automatic Panoramic Image Stitching using Invariant Features" International Journal of Computer Vision 74 (1) 59 - 73    DOI : 10.1007/s11263-006-0002-3
Szeliski R. 1996 "Video Mosaics for Virtual Environments" IEEE Computer Graphics and Applications 16 (2) 22 - 30    DOI : 10.1109/38.486677
Szeliski R. , Shum H.Y. 1997 "Creating Full View Panoramic Image Mosaics and Environment Maps" Proceeding of the Conference on Computer Graphics and Interactive Techniques 251 - 258
Chiba N. , Minoh M. , Kano H. 2000 "Feature-Based Lens Distortion Correction for Image Mosaicing" Proceeding of the IAPR Workshop on Machine Vision Applications 607 - 610
Luo J. , Shin S.S. , Park H. J. , Gwun O. B. 2011 “Stitcing for Panorama based on SURF and Multi-band Blending” Journal of Korea Multimedia Society 14 (2) 201 - 209    DOI : 10.9717/kmms.2011.14.2.201
Cho S.I. , Kim K.J. , Ban K.J. , Park K.W. , Kim C.Y. , Kim E.K. 2011 "3D Panorama Generation Using Depth-Map Stitching" Journal of Information and Communication Convergence Engineering 9 (6) 780 - 784    DOI : 10.6109/ijice.2011.9.6.780