Academia.edu no longer supports Internet Explorer.

To browse Academia.edu and the wider internet faster and more securely, please take a few seconds to upgrade your browser.

Academia.eduAcademia.edu

Outline

Three-Dimensional Scene Reconstruction: A Review of Approaches

https://doi.org/10.4018/978-1-61350-326-3.CH008
visibility

...

description

23 pages

Sign up for access to the world's latest research

checkGet notified about relevant papers
checkSave papers to use in your research
checkJoin the discussion with peers
checkTrack your impact

Abstract
sparkles

AI

Three-dimensional scene reconstruction is an essential task in computer vision and robotics, involving methods to create a three-dimensional representation of a scene from two-dimensional images. Various approaches have been developed, including shape from silhouette, stereo vision, and fusion techniques. This review outlines the main methods, their advantages and limitations, and provides references for further research.

Key takeaways
sparkles

AI

  1. Three-dimensional reconstruction techniques are vital across diverse fields such as robotics and archaeology.
  2. Methods are classified into geometric, photometric, and real-aperture approaches for 3D surface generation.
  3. The pinhole camera model is foundational for geometric scene reconstruction and requires intrinsic/extrinsic parameters.
  4. Stereo vision simplifies 3D reconstruction by using epipolar constraints to enhance correspondence search efficiency.
  5. Future challenges include improving automation, precision, and integrating shape and surface reflectance data.

Related papers

3D Reconstruction from Single 2D Image

The perception of 3D scene with stereovision is the capability of human vision but it is a challenge to computer systems. The challenge is to obtain 3D geometric shape information from planar images. This is termed as 3D reconstruction. It is usually done in a piece-wise fashion by identifying various planes of the scene and from those planes constructing a representation of the whole. The image is captured from a calibrated camera. The captured image is perspectively distorted. These distortions are removed by corner point estimation. The selected points give the true dimensions in the input image using view metrology. Then a 3D geometrical model is constructed in VRML according to true dimensions from metrology. In rendering, a texture map is created to each corresponding surface of a polygon in VRML model. VRML supports for walkthrough to a rendered model, through which different views are generated.

Survey on 3D Surface Reconstruction

Journal of Information Processing Systems, 2016

The recent advent of increasingly affordable and powerful 3D scanning devices capable of capturing high resolution range data about real-world objects and environments has fueled research into effective 3D surface reconstruction techniques for rendering the raw point cloud data produced by many of these devices into a form that would make it usable in a variety of application domains. This paper, therefore, provides an overview of the existing literature on surface reconstruction from 3D point clouds. It explains some of the basic surface reconstruction concepts, describes the various factors used to evaluate surface reconstruction methods, highlights some commonly encountered issues in dealing with the raw 3D point cloud data and delineates the tradeoffs between data resolution/accuracy and processing speed. It also categorizes the various techniques for this task and briefly analyzes their empirical evaluation results demarcating their advantages and disadvantages. The paper concludes with a cross-comparison of methods which have been evaluated on the same benchmark data sets along with a discussion of the overall trends reported in the literature. The objective is to provide an overview of the state of the art on surface reconstruction from point cloud data in order to facilitate and inspire further research in this area.

State-of-the-art Algorithms for Complete 3D Model Reconstruction

The task of generating fast and accurate 3D models of a scene has applications in various computer vision fields, including robotics, virtual and augmented reality, and entertainment. So far, the computer vision scientific community has provided innovative reconstruction algorithms that exploit variant types of equipment. In this paper a survey of the recent methods that are able to generate complete and dense 3D models is given. The algorithms are classified into categories according to the equipment used to acquire the processed data.

Review of 3 D Reconstruction Methods from Single View

2015

The increase in demand for 3D content has inspired researchers to devise techniques which allow 3D models to be acquired directly from the images. Various techniques have been developed to generate 3D model from 2D images. This paper covers various approaches for 3d generation. Model based techniques use image features such as texture, shading for generating 3D model. Constraint-based techniques use geometric properties like co-linearity, coplanar, point, normal. The geometric approach deals with geometric relationships between points, lines, planes, etc under imaging. Keywords—3D Reconsruction, Shape from silhouette, 3D mesh

A review and comparison of multi-view 3D reconstruction methods

The Journal of Engineering Research, 2017

3D reconstruction from multiple views is well studied, fundamental, yet a challenging problem in the field of computer vision. There is a large variety of approaches available in the literature. The methods use different representations for input scene/object and may provide different kinds of outputs. Some methods model entire scene as voxel-set where as some use level sets or polygon mesh representation. Output may be either volume or surface representing the reconstructed object/scene. Some methods work in image space where as some methods work in object space. These methods are developed to offer a good compromise between computation speed, computation complexity and accuracy along with feasibility in implementation. Selection of a particular method depends on the requirements of application and availability of required resources. Though, earlier reviews are available in the literature, fast advances in this field demand latest review. The paper presents a review and comparison ...

3D image reconstruction from multiview images

2017 8th International Conference on Computing, Communication and Networking Technologies (ICCCNT), 2017

This paper presents an efficient method to reconstruct 3D point cloud from an arbitrary number of views of an object of interest. Nowadays, Light Detection and Ranging (LIDAR) devices are used in autonomous vehicles and robots to get the 3D measurements. LIDAR contains laser scanners which directly gives the 3D information. However, these scanners are very expensive and prone to interferences and noise due to bulk hardware present in the device. To address this problem, mobile phone cameras are used in the proposed method as they are lighter and cheaper compared to scanners. Multiviews of an object of interest are obtained using a mobile phone camera and they are given as input to the structure from motion (SFM) algorithm. In this algorithm, an appropriate set of mixed feature extraction techniques are employed to get good number of 3D inlier points. The obtained results are visualised as 3D point clouds.

2013-Ebsco DOAJ IJCSRA - Methodology for 3D reconstruction

In this paper, we propose a methodology for 3D virtual reconstruction of objects that can be applied to virtual restoration. The methodology is based on an image-based modelling technique and allows generating a textured 3D mesh from a set of images. The proposed methodology consists in the following actions: obtain images of the object, processing of the images, 3D reconstruction of the object, finishing and completing details, restoration of the 3D virtual model. First, we review several frameworks and toolkits that can be used for image-based modelling and then a detail example of 3D reconstruction is presented. The advantage of this methodology for 3D virtual reconstruction is the use of inexpensive equipment, because only common video and computing devices are needed.

Construction of 3D shapes of objects from reconstructed 3D points

2018

Estimation of 3-D objects from 2-D images is inherently performed by either motion or scene features methods as it has been described in different literatures. Structure from motion as a method employed in this study uses calibrated camera and reconstructed 3-D points from the structure of the scene for reliable and precise estimates. In this study we construct 3-D shapes using color pixels and reconstructed 3-D points to determine observable differences for the constructed 3-D images. The estimation using reconstructed 3-D points indicates that the sphere is recovered by the use of scale factor due to its known size while the one obtained by using color pixels has look similar to the former but different in the scales of the axes.

Loading...

Loading Preview

Sorry, preview is currently unavailable. You can download the paper by clicking the button above.

References (94)

  1. DeCubber, G., Nalpantidis, L., Sirakoulis, G. C., & Gasteratos, A. (2008). Intelligent robots need intelligent vision: Visual 3D perception. In RISE'08: Proceedings of the EURON/IARP International Workshop on Robotics for Risky Interventions and Surveillance of the Environ- ment. Benicassim, Spain.
  2. DeSouza, G., & Kak, A. (2002). Vision for mobile robot navigation: A survey. IEEE Transactions on Pattern Analysis and Machine Intelligence, 24(2), 237-267. doi:10.1109/34.982903
  3. Dhond, U., & Aggarwal, J. (1989). Structure from stereo-A review. IEEE Transactions on Sys- tems, Man, and Cybernetics, 19(6), 1489-1510. doi:10.1109/21.44067
  4. Durou, J., Falcone, M., & Sagona, M. (2008). Numerical methods for shape-from-shading: A new survey with benchmarks. Computer Vi- sion and Image Understanding, 109(1), 22-43. doi:10.1016/j.cviu.200709003
  5. El-Hakim, S., Beraldin, J., Picard, M., & Courn- oyer, L. (2008). Surface reconstruction of large complex structures from mixed range data-the erechtheion experience. In ISPRS'08: Proceedings of the XXI Congress of the International Society for Photogrammetry and Remote Sensing (vol. 37, pp. 1077-1082). Lemmer, The Netherlands: Reed Business.
  6. Esteban, C. H., & Schmitt, F. (2004). Silhouette and stereo fusion for 3D object modeling. Com- puter Vision and Image Understanding, 96(3), 367-392. doi:10.1016/j.cviu.2004年03月01日6
  7. Favaro, P. (2007). Shape from focus and defocus: Convexity, quasiconvexity and defocus-invariant textures. In ICCV'07: Proceedings of the 11th IEEE International Conference on Computer Vision (pp. 1-7). Los Alamitos, CA: IEEE Com- puter Society.
  8. Favaro, P., & Soatto, S. (2005). A geometric ap- proach to shape from defocus. IEEE Transactions on Pattern Analysis and Machine Intelligence, 27(3), 406-417. doi:10.1109/TPAMI.2005.43
  9. Favaro, P., Soatto, S., Burger, M., & Osher, S. (2008, March). Shape from defocus via diffusion. IEEE Transactions on Pattern Analysis and Ma- chine Intelligence, 30(3), 518-531. doi:10.1109/ TPAMI.2007.1175
  10. Fischler, M. A., & Bolles, R. C. (1981). Random sample consensus: A paradigm for model fitting with applications to image analysis and automated cartography. Communications of the ACM, 24(6), 381-395. doi:10.1145/358669.358692
  11. Forsyth, D. A., & Ponce, J. (2002). Computer vi- sion: A modern approach (US ed.). Prentice Hall Professional Technical Reference.
  12. Galasso, F., & Lasenby, J. (2007). Shape from tex- ture of developable surfaces via Fourier analysis. In G. Bebis, et al. (Eds.), In ISVC'07: Proceedings of the 3rd International Symposium on Advances in Visual Computing (vol. 4841, pp. 702-713). Heidelberg, Germany: Springer.
  13. Gennery, D. (2006). Generalized camera calibra- tion including fish-eye lenses. International Jour- nal of Computer Vision, 68, 239-266. doi:10.1007/ s11263-006-5168-1
  14. Grauman, K., Shakhnarovich, G., & Darrell, T. (2003). Inferring 3D structure with a statistical image-based shape model. In ICCV'03: Proceed- ings of the 9th IEEE International Conference on Computer Vision (vol. 1, pp. 641-647). Los Alamitos, CA: IEEE Computer Society.
  15. Grossberg, S., Kuhlmann, L., & Mingolla, E. (2007). A neural model of 3D shape-from-texture: Multiple-scale filtering, boundary grouping, and surface filling-in. Vision Research, 47(5), 634-672. doi:10.1016/j.visres.2006年10月02日4
  16. Harker, M., & O'Leary, P. (2008). Least squares surface reconstruction from measured gradient fields. In CVPR'03: Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, (pp.1-7). Los Alamitos, CA: IEEE Computer Society.
  17. Hays, J., Leordeanu, M., Efros, A. A., & Liu, Y. (2006). Discovering texture regularity as a higher- order correspondence problem. In A. Leonardis, H. Bischof, & A. Pinz (Eds.), Eccv'06: Proceedings of the 9th European Conference on Computer Vision (vol. 3952, pp. 522-535). Graz, Austria: Springer.
  18. Hernandez Esteban, C., Vogiatzis, G., & Cipolla, R. (2008). Multiview photometric stereo. IEEE Transactions on Pattern Analysis and Machine Intelligence, 30(3), 548-554. doi:10.1109/ TPAMI.2007.70820
  19. Horn, B. K. P. (1989). Obtaining shape from shading information. Shape from shading (pp. 123-171). Mit Press Series Of Artificial Intel- ligence Series.
  20. Horn, B. K. P. (1990). Height and gradient from shading. International Journal of Computer Vi- sion, 5(1), 37-75. doi:10.1007/BF00056771
  21. Jacques, L., De Vito, E., Bagnato, L., & Vander- gheynst, P. (2008). Shape from texture for omnidi- rectional images. In EUSIPCO'08: Proceedings of the 16th European Signal Processing Conference.
  22. Janoos, F., Mosaliganti, K., Xu, X., Machiraju, R., Huang, K., & Wong, S. (2009). Robust 3D reconstruction and identification of dendritic spines from optical microscopy imaging. Medical Image Analysis, 13(1), 167-179. doi:10.1016/j. media.2008年06月01日9
  23. Jin, H., Soatto, S., & Yezzi, A. J. (2000). Stereo- scopic shading: Integrating multi-frame shape cues in a variational framework. In CVPR'00 [). Los Alamitos, CA: IEEE Computer Society.].
  24. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 1, 1169.
  25. Kannala, J., & Brandt, S. S. (2006). A generic cam- era model and calibration method for conventional, wide-angle, and fisheye lenses. IEEE Transactions on Pattern Analysis and Machine Intelligence, 28, 1335-1340. doi:10.1109/TPAMI.2006.153
  26. Kolev, K., & Cremers, D. (2008). Integration of multiview stereo and silhouettes via convex functionals on convex domains. In D. A. Forsyth, P. H. S. Torr, & A. Zisserman (Eds.), ECCV'08: Proceedings of the 10th European Conference on Computer Vision (vol. 5302, pp. 752-765). Heidelberg, Germany: Springer.
  27. Kru ̈ger, L. E., Wohler, C., Wurz-Wessel, A., & Stein, F. (2004). In-factory calibration of multiocu- lar camera systems . In Osten, W., & Takeda, M. (Eds.), Optical metrology in production engineer- ing (Vol. 5457, pp. 126-137). SPIE.
  28. Kruppa, E. (1913). Zur ermittlung eines objektes aus zwei perspektiven mit innerer orientierung. Sitzungsberichte der Mathematisch Naturwis- senschaftlichen Kaiserlichen Akademie der Wis- senschaften, 122, 1939-1948.
  29. Ladikos, A., Benhimane, S., & Navab, N. (2008). Efficient visual hull computation for real-time 3d reconstruction using Cuda. In Proceedings of the 2008 Conference on Computer Vision and Pattern Recognition Workshops (pp. 1-8). Los Alamitos, CA: IEEE Computer Society.
  30. Laussedat, A. (1898). Recherches sur les instru- ments: Les méthodes et le dessin topographiques. Gauthier-Villars.
  31. Lemaire, T., Berger, C., Jung, I.-K., & Lacroix, S. (2007). Vision-based slam: Stereo and monocular approaches. International Journal of Computer Vision, 74(3), 343-364. doi:10.1007/s11263- 007-0042-3
  32. Li, J., & Allinson, N. M. (2008). A comprehensive review of current local features for computer vi- sion. Neurocomputing, 71(10-12), 1771-1787. doi:10.1016/j.neucom.2007年11月03日2
  33. Liu, X., Yao, H., Chen, X., & Gao, W. (2008). Shape from silhouettes based on a centripetal pentahedron model. Graphical Models, 70(6), 133-148. doi:10.1016/j.gmod.200606003
  34. Liu, Y., Collins, R., & Tsin, Y. (2004). A com- putational model for periodic pattern perception based on frieze and wallpaper groups. IEEE Transactions on Pattern Analysis and Machine Intelligence, 26(3), 354-371. doi:10.1109/ TPAMI.2004.1262332
  35. Liu, Y., Lin, W.-C., & Hays, J. (2004, August). Near-regular texture analysis and manipulation. ACM Transactions on Graphics, 23(3), 368-376. doi:10.1145/1015706.1015731
  36. Liu, Z., & Klette, R. (2008, November 25-28). Dynamic programming stereo on real-world se- quences. In ICONIP'08: Proceedings of the 15th International Conference on Advances in Neuro- Information Processing (vol. 5507, pp. 527-534). Auckland, New Zealand: Springer.
  37. Lobay, A., & Forsyth, D. (2006). Shape from texture without boundaries. International Journal of Computer Vision, 67(1), 71-91. doi:10.1007/ s11263-006-4068-8
  38. Loh, A., & Hartley, R. (2005). Shape from non- homogeneous, non-stationary, anisotropic, per- spective texture. In BMCV'05: Proceedings of the British Machine Vision Conference (pp. 69-78).
  39. Lou, Y., Favaro, P., Bertozzi, A. L., & Soatto, S. (2007, 18-23 June). Autocalibration and uncali- brated reconstruction of shape from defocus. In CVPR'07: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (pp.
  40. Lourakis, M. I. A., & Argyros, A. A. (2004). The design and implementation of a generic sparse bundle adjustment software package based on the Levenberg-Marquardt algorithm (Technical Report 340). FORTH, Heraklion Crete, Greece, Institute of Computer Science.
  41. Lourakis, M. I. A., & Argyros, A. A. (2009). SBA: A software package for generic sparse bundle ad- justment. ACM Transactions on Mathematical Soft- ware, 36(1), 1-30. doi:10.1145/1486525.1486527
  42. Luhmann, T. (2003). Nahbereichsphotogrammetrie Grundlagen, Methoden, Anwendungen. 2. Heidel- berg: Auflage, Wichmann Verlag.
  43. MacLean, W., Sabihuddin, S., & Islam, J. (2010). Leveraging cost matrix structure for hardware implementation of stereo disparity computation using dynamic programming. Computer Vision and Image Understanding, 114(11). doi:10.1016/j. cviu.2010年03月01日1
  44. Masrani, D. K., & MacLean, W. J. (2006). A real- time large disparity range stereo system using fpgas. In ICVS'06: Proceedings of the International Conference on Computer Vision Systems (p. 13). Los Alamitos, CA, USA: IEEE Computer Society.
  45. Matusik, W., Buehler, C., & McMillan, L. (2001). Polyhedral visual hulls for real-time rendering. In S. J. Gortler & K. Myszkowski (Eds.), Proceedings of the 12th Eurographics Workshop on Rendering Techniques (vol. 1, pp. 115-126). Springer.
  46. Meydenbauer, A. (1867). Ueber die Anwendung der Photographie zur Architektur-und Terrain- Aufnahme. Zeitschrift für Bauwesen, 17, 61-70.
  47. Mikolajczyk, K., Tuytelaars, T., Schmid, C., Zisser- man, A., Matas, J., & Schaffalitzky, F. (2005). A comparison of affine region detectors. Interna- tional Journal of Computer Vision, 65(1-2), 43-72. doi:10.1007/s11263-005-3848-x
  48. Moreels, P., & Perona, P. (2007). Evaluation of features detectors and descriptors based on 3S objects. International Journal of Computer Vision, 73(3), 263-284. doi:10.1007/s11263-006-9967-1
  49. Nalpantidis, L., Chrysostomou, D., & Gasteratos, A. (2009). Obtaining reliable depth maps for robotic applications from a quad-camera system. In M. Xie, Y. Xiong, C. Xiong, H. Liu, & Z. Hu (Eds.), ICIRA '09: Proceedings of the 2nd International Confer- ence on Intelligent Robotics and Applications (vol. 5928, pp. 906-916). Berlin, Germany: Springer.
  50. Nalpantidis, L., Kostavelis, I., & Gasteratos, A. (2009). Stereovision-based algorithm for obstacle avoidance. In M. Xie, Y. Xiong, C. Xiong, H. Liu, & Z. Hu (Eds.), ICIRA '09: Proceedings of the 2nd International Conference on Intelligent Robotics and Applications (vol. 5928, pp. 195-204). Berlin, Germany: Springer.
  51. Nalpantidis, L., Sirakoulis, G., & Gasteratos, A. (2008). Review of stereo vision algorithms: From software to hardware. International Journal of Optomechatronics, 2(4), 435-462. doi:10.1080/15599610802438680
  52. Nayar, S., Ikeuchi, K., & Kanade, T. (1991). Shape from interreflections. International Journal of Computer Vision, 6(3), 173-195. doi:10.1007/ BF00115695
  53. Nayar, S., Watanabe, M., & Noguchi, M. (1996). Real-time focus range sensor. IEEE Transactions on Pattern Analysis and Machine Intelligence, 18(12), 1186-1198. doi:10.1109/34.546256
  54. Nehab, D., Rusinkiewicz, S., Davis, J., & Ra- mamoorthi, R. (2005). Efficiently combining positions and normals for precise 3D geometry.
  55. ACM Transactions on Graphics, 24(3), 543. doi:10.1145/1073204.1073226
  56. Nevado, M. M., Garcia-Bermejo, J. G., & Casa- nova, E. Z. (2004). Obtaining 3D models of indoor environments with a mobile robot by estimating local surface directions. Robotics and Autonomous Systems, 48(2-3), 131-143.
  57. Nister, D. (2004, June). An efficient solution to the five-point relative pose problem. IEEE Transac- tions on Pattern Analysis and Machine Intelligence, 26(6), 756-777. doi:10.1109/TPAMI.2004.17
  58. Park, M., Brocklehurst, K., Collins, R., & Liu, Y. (2009). Deformed lattice detection in real-world images using mean-shift belief propagation. IEEE Transactions on Pattern Analysis and Machine Intelligence, 31(10), 1804-1816. doi:10.1109/ TPAMI.2009.73
  59. Pradeep, K., & Rajagopalan, A. (2007). Improving shape from focus using defocus cue. IEEE Trans- actions on Image Processing, 16(7), 1920-1925. doi:10.1109/TIP.2007.899188
  60. Ramnath, K., & Rajagopalan, A. N. (2009). Discontinuity-adaptive shape from focus using a non-convex prior. In J. Denzler, G. Notni, & H. Su ̈ße (Eds.), DAGM'09: Proceedings of the 31st DAGM Symposium (vol. 5748, pp. 181-190). Heidelberg, Germany: Springer.
  61. Remondino, F., El-Hakim, S., Baltsavias, E., Picard, M., & Grammatikopoulos, L. (2008). Image-based 3D modeling of the Erechteion, Acropolis of Athens. In ISPRS'08: Proceedings of the XXI Congress of the International Society for Photogrammetry and Remote Sensing (vol. 37, pp. 1083-1091). Lemmer, The Netherlands: Reed Business.
  62. Sahay, R. R., & Rajagopalan, A. N. (2009). Real aperture axial stereo: Solving for correspondences in blur. In J. Denzler, G. Notni, & H. Su ̈ße (Eds.), DAGM'09: Proceedings of the 31st DAGM Sym- posium (vol. 5748, pp. 362-371). Springer.
  63. Scharstein, D., & Szeliski, R. (2002). A tax- onomy and evaluation of dense two-frame ste- reo correspondence algorithms. International Journal of Computer Vision, 47(1-3), 7-42. doi:10.1023/A:1014573219977
  64. Sˇeatovi'c, D. (2008). A segmentation approach in novel real time 3d plant recognition system. In A. Gasteratos, M. Vincze, & J. K. Tsotsos (Eds.), Proceedings6th International Computer Vision Systems Conference, Santorini, Greece, May 12- 15, 2008, (vol. 5008, pp. 363-372). Heidelberg, Germany: Springer.
  65. Shim, S.-O., & Choi, T.-S. (2010). A novel iterative shape from focus algorithm based on combinato- rial optimization. Pattern Recognition, 43(10), 3338-3347. doi:10.1016/j.patcog.2010年05月02日9
  66. Sturm, P., & Barreto, J. A. (2008). General imaging geometry for central catadioptric cameras. In D. A. Forsyth, P. H. S. Torr, & A. Zisserman (Eds.), ECCV'08: Proceedings of the 10th European Conference on Computer Vision (vol. 5305, pp. 609-622). Berlin, Germany: Springer.
  67. Sturm, P., & Ramalingam, S. (2004). A generic concept for camera calibration. In T. Pajdla & J. Matas (Eds.), ECCV'04: Proceedings of the 8th European Conference on Computer Vision (vol. 3022, pp. 1-13). Berlin, Germany: Springer.
  68. Szeliski, R. (1991). Fast shape from shading. CVGIP: Image Understanding, 53(2), 129-153. doi:10.1016/1049-9660(91)90023-I Tepper, O., Karp, N., Small, K., Unger, J., Rudolph, L., & Pritchard, A. (2008). Three-dimensional imaging provides valuable clinical data to aid in unilateral tissue expander-implant breast recon- struction. The Breast Journal, 14(6), 543-550. doi:10.1111/j.1524-4741.2008.00645.x
  69. Todd, J., Thaler, L., Dijkstra, T., Koenderink, J., & Kappers, A. (2007). The effects of viewing angle, camera angle, and sign of surface curvature on the perception of three-dimensional shape from texture. Journal of Vision (Charlottesville, Va.), 7(12). doi:10.1167/7.12.9
  70. Triggs, B., McLauchlan, P. F., Hartley, R. I., & Fitzgibbon, A. W. (2000). Bundle adjustment -A modern synthesis. In B. Triggs, A. Zisserman, & R. Szeliski (Eds.), ICCV '99: Proceedings of the International Workshop on Vision Algorithms: Theory and Practice (vol. 1883, pp. 298-372). London, UK: Springer-Verlag.
  71. Tsai, P.-S., & Shah, M. (1994). Shape from shad- ing using linear approximation. Image and Vision Computing, 12(8), 487-498. doi:10.1016/0262- 8856(94)90002-7
  72. Tuytelaars, T., & Mikolajczyk, K. (2008). Local invariant feature detectors: A survey. Foundations and Trends in Computer Graphics and Vision, 3(3), 177-280. doi:10.1561/0600000017
  73. Vlasic, D., Baran, I., Matusik, W., & Popovic, J. (2008). Articulated mesh animation from multi- view silhouettes. ACM Transactions on Graphics, 27(3), 1-9. doi:10.1145/1360612.1360696
  74. Vogiatzis, G., Esteban, C. H., Torr, P. H. S., & Cipolla, R. (2007). Multiview stereo via volu- metric graph-cuts and occlusion robust photo- consistency. IEEE Transactions on Pattern Analysis and Machine Intelligence, 29(12), 2241-2246. doi:10.1109/TPAMI.2007.70712
  75. White, R., Crane, K., & Forsyth, D. A. (2007). Capturing and animating occluded cloth. [TOG].
  76. ACM Transactions on Graphics, 26(3), 34. doi:10.1145/1276377.1276420
  77. White, R., & Forsyth, D. A. (2006). Combining cues: Shape from shading and texture. CVPR'06 . Proceedings of the IEEE Computer Vision and Pattern Recognition, 2, 1809-1816.
  78. Wilhelmy, J., & Kru ̈ger, J. (2009). Shape from shading using probability functions and belief propagation. International Journal of Computer Vision, 84(3), 269-287. doi:10.1007/s11263-009- 0236-y Woodham, R. (1981). Analysing images of curved surfaces. Artificial Intelligence, 17(1-3), 117-140. doi:10.1016/0004-3702(81)90022-9
  79. Xu, J., Xi, N., Zhang, C., & Shi, Q. (2009). Wind- shield shape inspection using structured light patterns from two diffuse planar light sources. In IEEE/RSJ International Conference on Intelligent Robots and Systems (pp. 721-726). Piscataway, NJ: IEEE Press.
  80. Xu, J., Xi, N., Zhang, C., Shi, Q., & Gregory, J. (2010). Real-time 3D shape inspection system of automotive parts based on structured light pattern. Optics & Laser Technology, 43(1), 1-8. doi:10.1016/j.optlastec.201004008
  81. Xu, Y., & Aliaga, D. G. (2009). An adaptive cor- respondence algorithm for modeling scenes with strong interreflections. IEEE Transactions on Vi- sualization and Computer Graphics, 15, 465-480. doi:10.1109/TVCG.2008.97
  82. Yemez, Y., & Wetherilt, C. J. (2007). A volumetric fusion technique for surface reconstruction from silhouettes and range data. Computer Vision and Im- age Understanding, 105(1), 30-41. doi:10.1016/j. cviu.200607008
  83. Zhang, J., & Smith, S. (2009). Shape similarity matching with octree representations. Journal of Computing and Information Science in Engineer- ing, 9(3), 034503. doi:10.1115/1.3197846
  84. Zhang, R., Tsai, P.-S., Cryer, J. E., & Shah, M. (1999). Shape from shading: A survey. IEEE Transactions on Pattern Analysis and Machine In- telligence, 21(8), 690-706. doi:10.1109/34.784284
  85. Zhou, K., Gong, M., Huang, X., & Guo, B. (2010).
  86. Data-parallel octrees for surface reconstruction.
  87. IEEE Transactions on Visualization and Computer Graphics, 99.
  88. Zhu, B., Lu, J., Luo, Y., Tao, Y., & Cheng, X. (2009). 3D surface reconstruction and analysis in automated apple stem-end/calyx identification. International Journal of Pattern Recognition and Artificial Intelligence, 52(5), 1775-1784. ADDITIONAL READING
  89. Cyganek, B. (2007). An introduction to 3d computer vision techniques and algorithms. John Wiley & Sons.
  90. Davies, E. R. (2004). Machine vision: Theory, algorithms, and practicalities. San Francisco, CA, USA: Morgan Kaufmann Publishers Inc.
  91. Faugeras, Olivier, Luong, Quang-Tuan, and Pa- padopoulou, T. (2001). The Geometry of Multiple Images: The Laws That Govern The Formation of Images of A Scene and Some of Their Applications. MIT Press, Cambridge, MA, USA.
  92. Forsyth, D. A., & Ponce, J. (2002). Computer vi- sion: A modern approach (US ed.). Prentice Hall Professional Technical Reference Hartley, R. I., & Zisserman, A. (2004). Mul- tiple view geometry in computer vision (2nd ed.). Cambridge University Press. doi:10.1017/ CBO9780511811685
  93. Paragios, N., Chen, Y., & Faugeras, O. (2005). Handbook of mathematical models in computer vision. Secaucus, NJ, USA: Springer-Verlag New York, Inc.
  94. Steger, C., Ulrich, M., & Wiedemann, C. (2008). Machine vision algorithms and Applications. Wiley VCH.

FAQs

sparkles

AI

What explains the advantages of photometric approaches in dense depth map generation?add

Photometric methods yield dense depth maps by evaluating pixel intensity distributions when illumination conditions and surface reflectance properties are well defined, as shown in numerous applications by 2023.

How has the evolution of camera calibration methods impacted reconstruction accuracy?add

Recent camera calibration advancements, such as those by Furukawa in 2015, achieved remarkable pixel accuracy using multi-view stereo and bundle adjustment techniques, significantly enhancing reconstruction outcomes.

When did the exploration of structure from motion techniques begin in 3D reconstruction?add

Structure from motion methods gained prominence in the 1980s, revolutionizing 3D reconstruction by estimating object shapes from multiple images captured from different viewpoints.

Why do geometric reconstruction methods face challenges with non-textured objects?add

Geometric methods struggle with non-textured objects due to their reliance on feature matching, which becomes problematic in low-contrast scenarios, as highlighted in ongoing 2023 research.

What current limitations prevent widespread adoption of 3D reconstruction technologies?add

Challenges such as the complexity of algorithms, real-time processing requirements, and the need for user-friendly systems hinder the mainstream adoption of 3D reconstruction technologies, according to recent findings.

Related papers

Computer vision based 3D reconstruction : A review

International Journal of Electrical and Computer Engineering (IJECE), 2019

3D reconstruction are used in many fields starts from the object reconstruction such as site, cultural artifacts in both ground and under the sea levels, medical imaging data, nuclear substantional. The scientist are beneficial for these task in order to learn, keep and better visual enhancement into 3D data. In this paper we differentiate the algorithm used depends on the input image: single still image, RGB-Depth image, multiperspective of 2D images, and video sequences. The prior works also explained how the 3D reconstruction perform in many fields and using various algorithms. 1. INTRODUCTION 3D Reconstruction task is one of the interesting task that meet its maturity already. These can be seen from the commercial products such as product from Agisoft and Pix4D that are capable of produced high quality of large scale 3D models. Furthermore, the hardware such as the computer vision has been developed and improve since then. There are some setup camera introduced in the research such as stereo camera and Kinect. In addition to the vision setup, kinect camera shows a great positive feedback from the researchers, proved by common vision setup that can be found in the literature review. Not only that, stereo camera setup can be found among the literature review. In addition to the stereo camera, custom stereo camera are quite popular among the researchers by combining two equals web camera that positioned by period of distance. The algorithm to perform 3D reconstruction between these camera are different due to the produced images are different as well. Kinect abilities allows RGB image and depth map produced, on the other hand Stereo camera has to perform another depth map acquisition algorithm by combining 2 RGB images. Numerous numbers of 3D reconstruction task can be found in capturing the site, cultural artifacts both in ground and under the sea levels [1]. The extinction factor is the most prominent issue in these area. Moreover, 3D imaging data also could help improve the accuracy of the anatomical features in order to observe some areas before coming to the surgery action .Furthermore, in order to perform 3D reconstruction, there are multiple approaches found in the literature review such as from the broad ranges of vision setup, various types of inputted image to construct 3D reconstruction. Thus, In this paper will describe more on those approaches. The great numbers of the researchers along with the hardware supports allows such algorithm to do high processing calculation in order to perform reconstruction task. There are some sections mentioned in part 2.. The benefits of reconstruction are to perform 3D recording, visualization, representation and reconstruction [2]. Moreover Tsiafaki and Michailidou explained that, there are 6 benefits in performing reconstruction and visualization: limiting the destructive nature of excavating, placing excavation data into the bigger picture, limiting fragmentation of archaeological remains, classifying archaeological finds, limiting subjectivity and publication delays, enriching and extending archaeological research.

3D Reconstruction Algorithms Survey

Advances in multimedia and interactive technologies book series, 2018

Over the past years, 3D reconstruction has proved to be a challenge. With augmented reality and robotics attracting more attention, the demand for efficient 3D reconstruction algorithms has increased. 3D reconstruction presents a problem in computer vision and as a result, much work has been dedicated to solving it. Different design choices were made to consider different components of the process. Examples of these differences are how the scanning process is tackled, how the 3D reconstructed world is represented, among other aspects. Therefore, an evaluation of these algorithms is necessary. This chapter focuses on the properties that facilitate the evaluation of 3D reconstruction algorithms and provides an evaluation of the various algorithms.

A New Fast 3D Reconstruction Approach using Multiple View Images

– The extract key points and matching the pictures are the most paramount reconstruction 3D factors. They almost two-thirds the time of reconstruction. This paper presents a method to extract the most paramount key points, through the use of GrabCut algorithm that elimintes considerable parts of images that does not have its prominence in the reconstructio. Moreover, the proposed algorithm uses siftGPU algorithm that runs parallel to any process more than one image at a time to extract key points and carry out matching process. The experiments show that the proposed system increase the speed of reconstruction and thoroughly good. Keywords – 3D Reconstruction, S tructure From Motion (S FM), Mash Reconstructionand Multi-View S tereo (MVS). I. INTRODUCTİON 3D reconstruction is one of the classical and difficu lt problems in co mputer vision, and finds its applications in a variety of different fields. In recent years, large scale 3D reconstruction from co mmunity photo collections has become an emerging research topic, wh ich is attracting more and more researchers fro m academy and industry. However, 3D reconstruction is extremely co mputationally expensive. For examp le, it may cost more than a day in a single mach ine to reconstruct an object with only one thousand pictures. In the Structure from Motion (SfM) model [1, 2], 3D reconstruction pipeline can be divided into various steps: feature extraction, image matching, track generation and geometric estimation, etc. Among them, image matching occupies the fundamental computational cost, even more than half of all in some case. Moreover, inexact matching results might lead to washout of reconstruction. Therefore, fast and accurate image matching is crit ical for 3D reconstruction. There are various ways to build reconstruction For example Reconstruction manually is most Statute method to reconstruct a 3D model for an object real world. but is a method ponderous and very intensive. Level of realism can be achieved [3]. The other way tried to eliminate the voltage on the user. 3D Scanner Variant Gu ide to reconstruction is to let co mputers to take some work, and is a well-established method of 3D scanning. The 3D the scanner is the device that apprehends the detailed informat ion for shape and appearance[4]. Modern developments in techniques scanners and Laser able to apprehend point clouds of scenes the real world, And also Automatically can reveal scene planes and create 3D models without the help of the user which can generate Points dense cloud from total images by photogrammetry tools [5].To create a point clouds typically sharing the same problems fro m noisy and lost data. Makes it very hard to apply the methods of surface reconstruction the direct [6,7], Points cloud doesn't contain the specific edges and borders. Last method offered by the this paperPhotogrammetry reconstruction regains 3D informat ion fro m a single or more of images. Main ly focused on rebuilding Photos mu lti view called stereo vision. Epipolar geo metry describes the features and the linkages between the scene and the 3D geometric projections on two or more images of 2D. Figure 1 shows the idealistic workflow for photogrammetric reconstruction. The first step of photogrammetric reconstruction includes the registration of all input images. This procedure is called structure-fro m-mot ion and includes the computation of intrinsic and extrinsic camera parameters. For reg istered images, it is possible to compute 3D positions from t wo or mo re corresponding image points. Multi-view stereo algorithms use these conditions and compute dense point clouds or triangulated meshes fro m the input scene. A B Fig.1. Reconstruction Photogrammet ry are recording mu ltip le images (A): is created by the structure fro m motion (B) and 3D geo metry by dense mult i view stereo [8]. The terms Multi-view Stereo (MVS) simu lates the sense of human sight distance and 3D objects. It uses two or more images fro m various points of view to get the 3D structure of the scene and distance information. Many algorith ms stereo mult iview [9, 10] used all the images at the same time to rebuild the 3D model. It requires a high expense and also lacks scalability. Furukawa [ 10] suggested PMVS (mult iple stereoscopic vision correction) and took mu ltiple p ictures of various views of the body to extract feature points. It is then expanded abroad to find more of the interview points. Furukawa CM VS also suggested (Views comp ilation of mult iple stereo) [ 12] in 2010, has been used to ameliorate the image co mb ines numerous susceptibility in order to see the stereo, and sustainable forest management PM VS b roker contacts. RGB-D systems have been developed due to the advent of RGB-D sensors, such as the Microsoft Kinect.

3D reconstruction

3D RECONSTRUCTION, 2020

Multi software used to get the results of multiple images. The (SFM) and (CMVS) used to obtain the sparse and dense reconstruction of sequential images collections. By taken images of engine through robot from different view to quantify the quality of (SFM) data in relation. We have implemented a number of comparison b/w different parts of engine in order to assess the mesh deviation and the reconstruction accuracy. The (SFM) and (CMVS) techniques and methods were mostly questioned by creating a complete 3D digital replication of engine.

A Review of the Existing Literature on 3D Reconstruction Methods

International Journal for Research in Applied Science & Engineering Technology (IJRASET), 2023

Using deep neural networks, supervised 3D reconstruction has made significant progress. However, large-scale annotations of 2D/3D data are necessary for this performance boost [4]. How to effectively represent 3D data to feed deep networks remains a challenge in 3D deep learning. Volumetric or point cloud representations have been used in recent works, but these methods have a number of drawbacks, including computational complexity, unorganized data, and a lack of finer geometry [7]. The ultrasound (US) examination is one method used to diagnose carotid artery disease. The flow conditions in the artery also play a role in the onset and progression of vascular diseases [5]. Stenting the carotid artery is a treatment option for carotid atherosclerosis. It is impossible to know for sure where an injection with a needle will begin. The place of the conduits is in the body, thusly, deciding the beginning stage of needle infusion is finished by assessment just and can't not entirely set in stone. The first thing that must be done is to locate the carotid artery in order to identify it. To determine it, we propose a modified template matching based on the ellipse feature for a 3D reconstruction of the carotid artery [1]. Data acquisition, pre-processing, segmentation, outlier selection for ellipse parameter fitting, and visualization are all used to process it. In comparison to the template matching method and the Hough Circle method, the proposed procedure with pre-processing produces the highest accuracy [1]. The objective of this research was to create three-dimensional (3D) ultrasound imaging of the carotid arteries to lessen the variability of volume measurements between and within examiners during follow-up scans of atherosclerotic plaques.

Review Paper on Techniques of 2D to 3D Image Reconstruction

International Journal for Research in Applied Science and Engineering Technology, 2023

The review paper emphasize on reducing technologies for 2D and 3D imaging, as well as model conversion. Although the popularity of 3D hardware is growing rapidly in the present era, 3D content is still dominated by its 2D counterpart. There are two main categories of image processing now available in the market, namely analogue and digital image processing. To produce hard copies such as scanned pictures and printouts, with images being the most common output, the analogue IP technique is used. On the other hand, Digital IP is used to manipulate digital images using computers, the outputs are often information related to images, mainly being data on features, edging characteristics, or masks. Image processing techniques, including Machine Learning and Deep Learning, can get more powerful.

3-D Visual Reconstruction : A System Perspective

2011

This paper presents a comprehensive study of all the challenges associated with the design of a platform for the reconstruction of 3-D objects from stereo images. The overview covers: Design and Calibration of a physical setup, Rectification and disparity computation for stereo pairs, and construction and rendering of meshes in 3-D for visualization. We outline some challenges and research in each of these areas, and present some results using some basic algorithm and a system build with off-the-shelf components.

Implementation of 3D shape reconstruction from range images for object digital modeling

ABCM Symposium Series in Mechatronics, 2006

The necessity of obtaining geometric models in three-dimension that represent with precision a real world object is becoming common each day. For this, one has to recur to methods of 3D Modeling. Three-dimension models have application in several areas, amongst which one can cite photogrammetry, archaeology, reverse engineering, robotic guidance, virtual reality, medicine, cinema, game programming, and others. A current challenge is the construction of 3D models digitized with precision enough to be used in manufacturing systems or numerical simulation of the performance of machines and components in operation, such as turbines and flows in non-circular ducts when the geometric model is not available. The reconstruction of 3D shapes of objects or scenes from range images, also known as depth maps, is preferable than using intensity images or stereoscopy. These maps represent information of distances measured from an observer (optical sensor or camera) to the scene in a rectangular grid. Therefore, the 3D information is explicit and will not need to be recovered as in the case of intensity images. The reconstruction process presents three stages. The first one is sampling of the real world in depth maps. The second stage is the alignment of several views within the same coordinate system, known as image registration. The third stage is the integration of the views for the generation of surface meshes, named merging. The current challenges converge to searching methods that meet with the highest number of desirable properties, such as robustness to outliers, efficiency of time and space complexity and precision of results. This work consists in the discussion of different methods dealing with 3D shape reconstruction from range images found in the literature and in the implementation of the second phase of 3D reconstruction: range image registration.

Methodology for 3D Scene Reconstruction from Digital Camera Images

2012

Digital images provide today an important source of data that deserves a careful statistical analysis. This paper concerns methods for retrieval of 3D information, including shape and texture, from cheap digital camera imaging outputs. It includes a three step reconstruction of a 3D scene with texture, from arbitrary partial views, in absence of occlusions. In Patrangenaru and Patrangenaru [23], Mardia et. al. and Patrangenaru and Mardia [24] a planar scene was reconstructed using image fusion, around representatives of sample mean projective shapes or sample mean affine shapes of landmark configurations shared by a number of partial views of the scene. In this paper we first analyze the advantages and limitations of such a reconstruction of a close to planar remote scene from its partial aerial views, by specializing this algorithm to affine transformations. Furthermore, we combine a projective shape reconstruction of a finite 3D configuration from its uncalibrated camera views, as developed in Patrangenaru, Liu and Sughatadasa [22], with a VRML technique, to reconstruct projectively a 3D scene with texture from a pair of digital camera images, thus allowing a more detailed statistical analysis of the scene pictured. We give three such examples of 3D reconstructions.

Academia
580 California St., Suite 400
San Francisco, CA, 94104

AltStyle によって変換されたページ (->オリジナル) /