Computer Vision Unit 4 Notes
Computer Vision Unit 4 Notes
Photography
1. Introduction
In computational photography and computer vision, self-calibration is a technique used to
estimate intrinsic parameters of a camera—such as focal length, skew, principal point, and
distortion—without requiring a physical calibration object like a checkerboard. Instead, it
relies on analyzing multiple uncalibrated images of a static 3D scene, typically during
Structure-from-Motion (SfM) or multi-view reconstruction.
These parameters are represented in a matrix called the Intrinsic Matrix (K):
3. What is Self-Calibration?
Self-calibration is the process of estimating the camera's intrinsic parameters using only image
correspondences across multiple views, without any known geometry or calibration target.
It assumes:
This process extracts camera intrinsics from epipolar geometry, particularly the Fundamental
Matrix (F) or the Essential Matrix (E) derived from matched image features.
4. Mathematical Background
4.1 Projection Model
A 3D point XXX in homogeneous coordinates projects to a 2D point xxx using the projection matrix P:
6. Real-Time Example: Structure from Motion (SfM)
In SfM, multiple images from different viewpoints are used to reconstruct 3D scenes. Here:
This integration enables full 3D reconstruction with only images as input—ideal for robotics, AR/VR,
and drones.
8. Benefits of Self-Calibration
✅ No special equipment required
✅ Enables 3D reconstructions in uncontrolled environments
✅ Works on existing datasets (e.g., historical photos)
9. Limitations
Requires sufficient camera motion (degenerate motion leads to failure)
Noise-sensitive
Doesn’t estimate lens distortion well
Numerically unstable if poorly conditioned
11. Summary
Self-calibration is a powerful tool in computational photography that enables intrinsic parameter
estimation directly from images. While not as precise as traditional methods, it is indispensable in
scenarios where no calibration pattern is available.
Projective Reconstruction in
Computational Photography
1. Introduction
Projective Reconstruction is the process of recovering 3D structure and camera motion
from multiple 2D images without requiring knowledge of the camera's intrinsic or
extrinsic parameters. It provides a reconstruction that is accurate up to a projective
transformation, meaning the 3D geometry is recovered but not with real-world metric
information (like exact angles or distances).
Add more cameras and apply Bundle Adjustment to refine all 3D points and camera matrices.
6. Diagram: Projective Reconstruction Overview
Image 1 Image 2
↓ ↓
+----------------+------------------+
| Feature Matching and F Matrix |
+----------------+------------------+
↓
Estimate P1 and P2 from F
↓
Triangulate 3D Points
↓
Projective 3D Point Cloud (X)
7. Applications
3D Modeling from uncalibrated photo collections (e.g., tourist photos).
Archaeology and cultural heritage digitization.
Augmented Reality where metric accuracy is not critical.
Movie production (3D scenes from set photos).
Robotics / Drones for basic scene understanding.
8. Limitations
Cannot recover real-world scale or shape.
Sensitive to feature correspondence errors.
Requires at least 2 images (3+ for robustness).
Needs good camera motion (not planar or degenerate).
9. Comparison with Other Reconstructions
Type of Reconstruction What it Recovers Requirement
Shape up to projective
Projective Uncalibrated images
transformation
Metric / Euclidean Real-world distances and angles Full calibration (intrinsics + extrinsics)
10. Summary
Projective reconstruction is a foundational technique in computational photography that enables 3D
scene recovery from images without camera calibration. While not metrically accurate, it allows for
visually consistent modeling and forms the basis for more refined reconstructions.