Archive for the ‘ Just Interesting ’ Category

Face Recognition by Yima and Features of Andrew Ng’s recent work.


Was thinking in terms of Andrew Ng’s “Building High-level Features Using Large Scale Unsupervised Learning (NIPS 2012)  and the Yima’s Robust Face Recognition (code present my other blog)

What could be benefits of using the features coming from the Andrew Ng and explicitly modeling them using the Sparse dictionary learning. Definitely one cannot use the the Dictionary as done by the Yima, since that is not feasible for huge amount of data and people. So will the features coming from the Andrew Ng’s work provide the robustness when used for the dictionary learning and then the coding?

Or Group Sparse coding and Block Dictionary learning could be used to better model the network itself, thus reducing the complexity and time required to train the network?

Just a thought.

CVPR 2012, attending Sebastian Thrun’s talk


{being updated as the talk goes on}

Sebastian is wearing a version of google glasses and showing the Google driver-less car video. The project is interesting but the story behind is much more interesting, they shows video of one of researcher driving it, who is legally blind.

It is showing how the Change Detection and segmentation will become an important problem.

If you guys can find the video link do share, talk is good enough to re listen

Talking about California the motorbikes can pass through very near to the car or inbetween the two cars to overtake them, so it becomes difficult to track them. Especially when they become so near to the car that they appear to be one. Mentions that same kind of problem could be seen in the Kinect and if someone working on the Tracking can solve more efficiently this problem it could be life saving. 

Saying the driverless car is lot safer than the human in case of collision but mentions where there are situations where computer cannot fully understand the situation and reacts improperly.

Showing one more application, used by the motor patrolling personals. So that they don’t have to give much attention to car and can do their job.

Excellent talk, although we were discussing a case of such cars in cities like Lahore, Delhi or even New York. In some crowded cities such situations arise quite commonly where the such “safe” cars might end up in deadlock.

Meet Edward H Adelson


If you have seen following image you have met Edward H Adelson,

He is faculty member at M.I.T Dept. of Brain and Cognitive Sciences, was recently going over some part of his paper “On Seeing Stuff: The Perception of Materials by Humans  and Machines“, quite interesting paper. Talks about why recognizing materials is important and points out that machines are not able to do that. In his paper one example is of icecream, that due to it’s texture can be recognized even by the child where as a machine cannot do that. This paper was published in 2001, I feel still the machines cannot recognize ice-cream.

There are many illusions on his webpage, have a look, http://persci.mit.edu/people/adelson

Can Computer Vision do that?


I was listening to one Bach as recorded for BBC Proms Bach Day ‘Passacaglia and fugue’, in the comments of youtube video there was reference to the Flute player.  I wanted to know where the video captured the Flute player so I had to walk through whole video (somewhat Binary search manner) until I found the video clip.

While searching I began wondering can recent technology solve problem of finding that part of video where it shows certain instrument is being “played”. Note the PLAYED part.

The naive way is to find the frames where e.g. Flute is being shown, then use the sound analysis to find whether in  those frames we can hear the Flute sound or not. It appears to be good solution, however the problem is finding where the Flute is being played is not easy when many other instruments are also playing. Secondly it’s not necessarily that the flute person that is being shown is actually playing flute also.

Question is Can we judge (both using the sound and not using sound) whether in given clip some instrument is being “PLAYED” or is being just shown,

Try this by looking at this video while you enjoy the amazing Bach (From 4:37) http://www.youtube.com/watch?v=bXwHORPeOsg&t=4m37s

Bean Machine; demonstrating Central Limit Theorem


During searching for some material related to Gauss I came through this interesting machine called Bean Machine.  Following Image is from the Wikipedia.

Bean Machine; From Wikipedia

Bean Machine

The idea is that balls dropped from top and gets randomly moved as they bounce off from the pins. Finally after the end of the  pins the are collected by one of the bins. After sufficient number of balls have been dropped the Bell Shape starts to appear, made by the height of how many balls are in each bin.

http://en.wikipedia.org/wiki/Bean_machine

However this more interesting figure has been taken from the Wikipedia’s Normal Distribution article entry.