Thursday, December 27, 2007
Computer Vision (37): Sensing through Seismics, The Golden Mole
FORSAKEN FANFARE
Here are some of the glimpses of kerala (Cochin, Attirapalli and Alleppey backwaters) through my camera: http://www.flickr.com/photos/57078108@N00/.
Tuesday, October 30, 2007
Computer Vision (36): Mechanical or Knowledge based CORRESPONDENCE
In spite of expending sleepless nights giving deep thoughts on what could be the technique behind our brain solving the problem of depth perception, my brain only gave me a drowsier day ahead. So I started to filter out the possibilities to narrow down to the solution. The question I asked to myself was; is our brain using knowledge to correspond the left and the right images, or is it something that happens more mechanically? I had tried out a lot of knowledge based approaches, but only in vain and even the discussion that we had in the earlier post concluded to nothing. I wanted to take a different route by thinking of a more mechanical and less of a knowledge based approach. My brain then pointed me to the age old theory proposed by Thomas Young to explain the wave nature (interference) of light, “The double Slit Experiment”. How could this be of use to solve a seemingly unrelated problem of depth perception? On comparing you will find a few things in common between the two setups. Both are trying to deal with light and both of them are trying to pass the surrounding light through two openings and combine them later. I excitedly thought, have I unlocked the puzzle?
Let’s analyze and understand it better to know if I really did! I am neither an expert in physics nor biology, so I can only build a wrapper around this concept and not verify its complete path.
λ is the wavelength of the light
s is the separation of the (slits/eyes)
x is the distance between the bands of light (also called fringe distance)
D is the distance from the (slits to the screen/eye and retina)
If a light source is placed at 3 different locations equidistant from the center of the slits, the one at red would produce a symmetric pattern and the other two I guess would not. I have not experimented this and hence the letters NX (Not eXperimented). If my guess is right, a light source placed anywhere in the 3D space would produce a unique pattern on the screen!!! This means an analysis of this pattern would tell us the exact location of the source in the 3D space.
Tuesday, October 16, 2007
Photography and Travel: Kudremukha
Wednesday, October 3, 2007
Computer Vision (35): Segmentation Verses Stereo Correspondence
Some people might argue that it’s almost impossible to achieve this from a 2D image, since there is no way to distinguish the plain of the dog from that of the ground. Remember, your brain has already done it! In a real scenario even if we come across such a view our stereo vision would ensure that the dog forms an image separate to the plain of the ground and hence would get segmented due to the variation in depth. Our brain can still do it from the 2D image here due to the tremendous amount of knowledge it has gathered over the years. In short, stereo vision helped us build this knowledge over the years and this knowledge is now helping us to segment objects even from a 2D image. The BIG question is, how do we do it in a computer?
Whenever I start to think about a solution to the problem of stereo correspondence the problem of segmentation would barricade it. This is why. The first step to understand or solve for stereo correspondence is to experiment with two cameras taking images at an offset. Below is a sample image.
Thursday, September 20, 2007
Computer Vision and Photography (34): The Focus Story Continues…
Saturday, August 25, 2007
Computer Vision and Photography (33): Capturing stereo images using a single camera
Chak De
Given a camera and a requirement to take stereo images, this arrangement was anyone’s mind game. I thought simple things like these need not be documented. A few months after this finding I saw a paper on exactly the same concept from a university in Switzerland. Do we really need PhDs to build this school level optics? Now I know the reason behind the very poor numbers behind India’s contribution towards the world’s papers and patents. Are we too fainéant to put our findings on paper or do we think we are not up to the mark in creating new things when compared to others? That too when others are confident of such simple findings.
CHAK DE
Saturday, August 11, 2007
Computer Vision (32): Monocular Cues
All these cues supplemented with our knowledge will always give us if not accurate a misty information about depth even in a 2D scenario.
Thursday, July 26, 2007
Ideas and Technology: More Intelligent Alarms
Sunday, July 22, 2007
Computer Vision (31): "Seeing" through ears
Even Nature can only produce best designs and not perfect ones and the Owls will definitely have to starve if their prey manages to remain silent. To make its design more reliable and worthy, Nature has never allowed a prey to have this very thought in its mind.
Saturday, July 21, 2007
Computer Vision (30): Why wasn't our face designed like this?
2. http://puneethbc.blogspot.com/2007/04/computer-vision-13.html
Wednesday, July 18, 2007
Photography: Effects of Aperture variation on Clarity
Also I have a few sample images from my multishot image merging software here: http://www.flickr.com/photos/57078108@N00/I will be launching it as soon as the GUI gets ready!
Sunday, July 8, 2007
Photography: Effects of aperture variation on sharpness
Taken with f36
To make the difference more apparent I selected the max and min apertures available in my camera in telephoto and kept the focus plane fixed, still I see a wider aperture giving better sharpness in all regions. Still, at this point of time I am more towards my concepts being true. My experiment turns out to be junk because in order to verify my concepts, the experimental setup should have had no other variables other than the aperture settings between the two shots. Here are the two things that changed between the two shots unknowingly.
- Here I tried to shoot the clouds with min and max apertures and the shutter speed that I got was 1/80 and 1/4000 respectively. Even though the shots were taken only a few seconds apart we can easily see that the clouds have significantly changed their pattern in this time. This places a high probability of more cloud movement in the lower aperture shot than the wider one, which might have caused more blurness in the former.
- Both the shots were taken with the camera hand held from on top of a pretty high building under windy conditions, which might have caused more handshake and in turn blurring in the lower aperture shot than the higher one.
Sunday, June 24, 2007
DSLR's and Photography
Wednesday, June 20, 2007
Photography and Travel: Bhimeshwari and Shivanasamudra
Thursday, June 14, 2007
Photography, Programming and Algorithms: Merging Multishot Motion Images
Tuesday, June 5, 2007
Computer Vision (29): Motion Segmentation
Wednesday, May 30, 2007
Photography and Travel: Ooty
Thursday, May 24, 2007
Computer Vision (28): Motion Detection
Wednesday, May 16, 2007
Photography and Travel: Bellandur Lake
Computer Vision (27), Optics and Photography
Tuesday, May 15, 2007
Computer Vision (26) and Optics
Monday, May 14, 2007
Photography and Travel: Nagarhole
Friday, May 11, 2007
Computer Vision (25) and Optics
In the first image of the sequence, the focus point was moved just behind the LED and we see a similar image as when the focus point was placed between the matchstick and the LED. But now the rays have actually crisscrossed which is not observed here since the cone is symmetric. To demonstrate the crisscross nature, I placed an opaque object and covered the left half of the lens, which made the right semicircle of the circular projection of the cone, disappear! To come back to our proper cone I moved the focus point back to the matchstick and did the same experiment. Now covering the left portion of the lens masks the left semicircle of the LED! This means there no crisscross!
Wednesday, May 9, 2007
Computer Vision (24) and Optics
Tuesday, May 8, 2007
Computer Vision (23) and Optics
Sunday, May 6, 2007
Computer Vision (22) and Optics
The intensity and frequency of the reflected light from these various points can be different and hence get summed up at a point on the retina. This scenario can happen for every pixel on the sensor and hence the image that you will get will just be the summation of the intensities and frequencies of the rays coming out from various points around you. As a result of this you will always end up with a uniform patch of light on the sensor if you try to take an image without a lens.
If you didn’t have a lens in your eyes, you would only be able to know the amount of light present in the surrounding and not the objects present in front of you. The various objects wouldn’t be distinguishable at all.
Friday, April 27, 2007
Computer Vision (21) and Optics
Monday, April 23, 2007
Photography and Travel
Saturday, April 21, 2007
Computer Vision (20)
A camera takes the horizontal projection of objects (horizontal line pointing towards you), and so the distance between the horizontal bars along this direction cannot be shown in this 2D image. The vertical distance between them is ‘v’. In other words, if we try to capture this 3D setup in a stereo image pair to get the horizontal depth between the bars you will end up with exactly the same image in the left and right. There is no use seeing it stereoscopically, because, which point in the two images will the brain correspond? Since the camera is moved horizontally, the vertical distance between the two bars remains the same in the stereo image.
Photography
Sunday, April 15, 2007
Computer Vision (19)
The red lines are traced when the eyes combine the rectangle and the green lines when they combine the circle. The point of intersection of the red lines gives the 3D location of the rectangle and the green lines that of the circle. As mentioned earlier the circle is in front of the rectangle when cross viewed. One of the points for the formation of the triangle comes from the point of intersection of either the red or the green lines and the other two points are the two eyes. The distance of the point of intersection of lines from the two eyes (d), depends on the separation between the images, so the absolute distance of the objects remains unknown in the stereo image pair. The relative depth of different objects from one another is obtained by corresponding objects form the two images, which moves the point of intersection of the lines according to the 3D placement of the objects (similar to red and green lines).