
(Auto)stitching Mosaics
Varun Neal Srivastava
Correspondences were manually selected. In order to help this process, I built a fine-tuning tool that would automatically "snap" a correspondence to a nearby pixel with a higher NCC. This resulted in fast and accurate correspondence-selection.
Finding the keypoint correspondences between two planar surfaces allows us to compute
the homography matrix H that transforms one planar surface to the other.
One application of this is to "rectify"—front align—a given image. Below is such
an example on my pencil box.
This works quite well. It is worth noting an interpolation is needed
in order to complete the warp.
Below is another example of rectification on Holbein's 1533 Painting
The Ambassadors. Particularly, we are rectifying the skull
present in the bottom left of the image.
We are having a lot of fun rectifying! I am going to use this technique to make a meme.
We have the tools to stitch together mosaics!
We will use
a distance transform as our mask — this will allow us to determine
the boundaries for our images, and the behavior of the mosaic on the regions
where two images overlap.
Oh no! The lighting is all fucked up. Instead of using a binary mask, let's apply a Gaussian Blur to the mask to smooth over transitions.
Much better! Let's use this technique to make more mosaics:
Anthropology building:
It is clear in the above example that it matters which perspective
we choose to align each image. In the above example, the first and second
images are greatly distorted to match the plane of the third image.
In an attempt to do some fun stuff, I tried to make myself floating.
Unfortunately, the lighting wasn't great here, and there were not many
good corespondences, but the strange warping gave me the added effect of looking taller!
It has been incredibly fun to start thinking of images in terms of the projective planes that they lay on. In order to implement the homography matrix, I spent a lot of time playing with matrices by-hand to see how they would map planes (parallelograms) to each other.
We use Harris Corners and the ANMS refinement technique to identify interest points automatically.
Next we extract feature vectors around each interest point, which consist of a rescaled, blurred 40 x 40 vector around the point, resulting in a 64-dimensional descriptor. Below are some example description vectors.
Now we have to match features to derive correspondences. We do this by:
Now we can use these correspondences to compute our mosaics. Here are comparisons of some of our "by-hand" panoramics and corresponding automosaics:
Given a set of unordered images, we attempt to recognize which ones form a panorama,
and then proceed to stitch those images together into a panorama. The technique
I used for this was naive: if two images have many correspondence points,
they can be stitched together. Putting together all the images that can be stitched
together (with some consideration to order) will form a panorama.
I took 20 random-ish images (pictured below with similar images placed next to each other),
a few of which depicted the same object from
multiple perspectives, and algorithmically extracted the mosaics that could be turned
into panorammics.
This was successful and yielded the following panoramas:
The last panoramic (of my parents' wedding) was not actually of the same image, so it makes sense that the resulting panoramic looks incorrect. That said, it makes sense that they were detected as potential candidates for a panoramic and the algorithm attempted to match them.
One of the coolest techniques I learned from this project was Lowe Thresholding. I am surprised at how effective it is as a metric for quality selection, as it seems to be empirically very reliable. I am also surprised that a naive nearest neighbors worked, without implementing rotational invariance, on my feature vectors. I think that this speaks to the power of simple metrics being sufficiently effective much of the time.