LLMpediaThe first transparent, open encyclopedia generated by LLMs

COLMAP (software)

Generated by GPT-5-mini
Note: This article was automatically generated by a large language model (LLM) from purely parametric knowledge (no retrieval). It may contain inaccuracies or hallucinations. This encyclopedia is part of a research project currently under review.
Article Genealogy
Parent: SURF Hop 5
Expansion Funnel Raw 1 → Dedup 0 → NER 0 → Enqueued 0
1. Extracted1
2. After dedup0 (None)
3. After NER0 ()
4. Enqueued0 ()
COLMAP (software)
NameCOLMAP
DeveloperJohannes L. Schönberger
Released2016
Programming languageC++, CUDA
Operating systemLinux, macOS, Microsoft Windows
LicenseBSD

COLMAP (software) is a general-purpose photogrammetry and 3D reconstruction system for structure-from-motion and multi-view stereo. It provides a graphical user interface and command-line tools to process unordered image collections into sparse and dense 3D models, and integrates with widely used tools in computer vision and graphics. COLMAP is used across research groups, academic institutions, and industry labs for mapping, cultural heritage, robotics, and visual effects.

Overview

COLMAP was created by Johannes L. Schönberger and is associated with the Visual Geometry Group and multiple research laboratories at institutions such as the Massachusetts Institute of Technology, University of Oxford, and ETH Zurich. The project sits alongside prominent software and projects like OpenMVG, OpenMVS, MeshLab, Agisoft Metashape, and VisualSFM in the photogrammetry ecosystem. It targets researchers and practitioners in computer vision, robotics, and photogrammetry who require reproducible Structure from Motion and dense reconstruction pipelines compatible with datasets used in benchmarks like Middlebury, DTU, and KITTI.

Features and Components

COLMAP bundles components for feature detection and matching, geometric verification, incremental and global Structure from Motion, sparse reconstruction visualization, multi-view stereo, and surface reconstruction. The feature extraction modules compare with detectors and descriptors such as SIFT, SURF, ORB, and relate to tools like VLFeat, OpenCV, and the CUDA-accelerated implementations used in GPU research. Matching and retrieval leverage vocabulary trees and techniques similar to those used in the Bag-of-Words models developed by Oxford researchers, and the geometric verification pipeline uses concepts from RANSAC variants originally proposed in the context of robust estimation. Dense stereo and patch-based multi-view stereo modules produce point clouds suitable for meshing with Poisson Surface Reconstruction and tools like CGAL and PCL.

Algorithms and Methodology

The core algorithms combine classical and modern approaches in multi-view geometry and optimization. Feature detection and description are inspired by SIFT and its derivatives, while pairwise matching uses nearest-neighbor search strategies comparable to FLANN and KD-tree approaches. Geometric verification employs RANSAC and robust pose estimation methods related to five-point and eight-point algorithms developed in multiple computer vision labs. Incremental Structure from Motion performs camera pose estimation and triangulation with bundle adjustment implemented via sparse nonlinear least squares solvers similar to Ceres Solver and g2o. Global Structure from Motion leverages rotation and translation averaging techniques that trace their lineage to optimization work from ETH Zurich and Stanford researchers. Dense reconstruction uses PatchMatch-like stereo propagation and photometric consistency formulations akin to algorithms from Microsoft Research and INRIA for multi-view stereo.

Workflow and Usage

A typical workflow starts with image import and EXIF-based initialization, follows with feature extraction and exhaustive or vocabulary-tree-based matching, then proceeds to sparse reconstruction via incremental or global SfM, and concludes with dense reconstruction and meshing. Users interact through a Qt-based GUI or a command-line interface that integrates with scripting environments used in academia and industry, including Python-based pipelines at institutions like Carnegie Mellon University, Stanford University, and University of California labs. Outputs—camera poses, sparse point clouds, dense point clouds, and meshes—are compatible with visualization and postprocessing tools such as Blender, CloudCompare, and Meshlab. The software supports GPU acceleration for intensive stages, aligning with practices seen in high-performance computing centers and graphics laboratories.

Performance and Evaluation

COLMAP has been benchmarked against datasets produced and curated by research groups at Middlebury, ETH Zurich, and KITTI, and comparisons often involve metrics used in challenges hosted by CVPR, ECCV, and ICCV. Performance evaluations consider reconstruction accuracy, completeness, and runtime, often contrasting incremental versus global SfM strategies similar to studies published by groups at Microsoft Research and Google. GPU-accelerated stages demonstrate significant speedups on NVIDIA hardware used in many vision labs, while memory and scalability characteristics reflect design trade-offs examined in computational geometry and large-scale mapping projects, including those at NASA and ESA for photogrammetric tasks.

History and Development

The project originated in the mid-2010s from research efforts by Johannes L. Schönberger during collaborations and affiliations with academic groups at institutions such as the University of North Carolina and Columbia University, and it evolved through contributions from students and researchers connected to research labs at ETH Zurich, TU Graz, and the University of Washington. COLMAP's development has been influenced by landmark works and software in photogrammetry and computer vision from researchers at Oxford, Stanford, and Microsoft Research, and it has been cited in publications presented at conferences like CVPR, ECCV, and ICCV. Ongoing development continues through community contributions and integration with academic datasets and benchmarks maintained by universities and research consortia.

Category:Photogrammetry software