In this project you will create a face recognition system. The program reduces each face image to a vector, then uses Principal Component Analysis (PCA) to find the space of faces. This space is spanned by just a few vectors, which means each face can be defined by just a set of coefficients weighting these vectors.
You are given a skeleton program that provides a sophisticated command-line interface as well as most of the image and vector classes and operations you will need. Your job will be to fill in the functions that do PCA, projection into face space, determining if a vector represents a face, verifying a user based on a face, finding a face match given a set of user face information, and finding the size and position of a face in an image.
PCA is a technique by which we reduce the dimensionality of data points. For example, consider the space of all 20 by 30 pixel grayscale images. This is 600 dimensional space because 600 data points are required to represent the values of the 600 pixels. But suppose we only consider images that are valid faces. Such images may lie in a small subspace of the set of all images. If this subspace is a hyperplane spanned by k vectors, then we can represent any point that lies on this plane with only a set of k coefficients.
To use PCA to do pattern recognition you first take a representative sample of data points and find the subspace with the desired dimensionality that best fits the sample data. This step needs to be done only once. Once it has been done, you can determine if a new point is a valid data point by projecting it onto the subspace and measuring the distance between the point and its projection.To use PCA for face recognition we must represent each face image as a vector of pixel values. To generate this vector the face image must be cropped and scaled, and its intensity must be normalized. The cropping can either be done by hand or by searching images for faces. This second option is only possible once the face recognition algorithm has already been trained with some input faces.
The input data you will be using is divided into several sets. They are included as folders in the skeleton code zip file:
smiling_cropped: Already cropped images of smiling vision students, taken in class.
nonsmiling_cropped: Already cropped in-class images of vision students wearing their normal sullen expressions.
eigenfaces_skeleton.zip: Skeleton
code
files.zip: All the images. You should
unzip them and put them with the skeleton code.
The skeleton code is large, but please take the time to get some familiarity
with the classes and methods it provides. There is a lot of useful functionality
included, like vector and image operations, which would take a lot of
time to write yourself. The program uses a simple command line interface,
and it is simple to add your own functionality in case you want to tackle
some extra credit, or add commands to facilitate your experimentation. At
the minimum you will only need to modify two files, faces.cpp, and eigfaces.cpp,
but you will still need to understand the contents of most of the other
classes.
Here is a description of the classes in the skeleton code. You will need to read this page to understand the organization of the code. It also gives information on numerous methods that you will need to call in your code, so please read it very carefully.
All the required work can be done in eigfaces.cpp and faces.cpp. The functions you must fill in are members of classes Faces and EigFaces. A Faces object stores a set of user faces loaded from images. EigFaces is derived from Faces. An EigFaces object stores a set of eigenfaces as well as the average face.
Most of the methods you will have to implement are called from main.cpp, so you should study main.cpp closely to see how these methods are used.
Each of the items below have an upper bound on the number of lines of code needed, not including comments. If you have many more lines of code for a given step than the upper bound, then you may be doing something wrong, or at least making more work for yourself. Be sure to look in vector.h, image.h, and face.h to see what useful functions are already provided for you.
Note that points will not be taken off for making your code longer than necessary, as long as it is correct and well coded. However, you should try your best not to go too far over the upper bound, as it means unnecessary work for you and will make your code more difficult to read and harder to grade.
void jacobi(double **a, int n, double d[], double **v, int *nrot); Computes all eigenvalues and eigenvectors of a real symmetric matrix a[1..n][1..n]. On output, elements of a above the diagonal are destroyed. d[1..n] returns the eigenvalues of a. v[1..n][1..n] is a matrix whose columns contain, on output, the normalized eigenvectors of a. nrot returns the number of Jacobi rotations that were required.
This is a complicated problem, so don't expect to get perfect results. However, you should be able to find the faces in the given group1.tga image; after all, the faces were used to form the eigenspace! Your code will be tested on this.
Lines required: 78
Here are general tips, requirements, and things to watch out for that may come in handy as you are coding your project or running your experiments.
See the usage function in Main for a list of commands and their parameters. Or you can just run the program with no arguments and this usage info will be output. There are examples of every command listed in ReadMe.txt. If you add any of your own commands please modify the usage info and ReadMe.txt.
Use eigenface width and height of 25 by 25 or greater for your experiments. You can use smaller eigenfaces such as 15 by 15 for debugging so you won't have to wait a long time. Note that if the eigenfaces you use are too large, Jacobi will run too slowly to get results in any reasonable number of days (or even years!).
Always resize a Face to the proper size before loading a targa file into it. You don't need to worry about this when loading into an Image, as images just resize automatically to the size of the loaded image.
If you do anything non-obvious in your code, comment it. Otherwise your TA won't be able to figure out what you were trying to do and won't be able to give you partial credit if it has a bug.
To aid you in debugging it is helpful to output debug images that you can look at and verify that the image contains what you expect. Use the saveTarga function to save Image or Face objects to file. For images such as eigenfaces which may not have pixel values arranged neatly from zero to 255, you will have to normalize to between 0.0 and 255.0 first by calling the normalize method.
The working directory when running from inside Visual Studio is eigenfaces\files. Thus any filename arguments you give should be relative to this directory. Of course you can change the working directory if you want to.
It is recommended that you don't try computing eigenfaces or doing face detection in debug mode, your code will run too slowly. If you need to debug something you might want to modify the code to skip to the point you want to debug so that you don't have to wait too long.
Your turned-in executable MUST be compiled in Release mode, NOT debug mode!
You must complete all of the following experiments to test your code. The results of each experiment, including the answers to all the questions, should be written in your Write-up, along a discussion of why you think the experiment turned out the way it did, as well as any unexpected results.
Note that some of the experiments may require you to run a command over and over again. One way to accomplish this is to use a batch file. Or, if you wish, you can easily add your own command that takes the place of a string of commands. See the description of class Main above to find out how to do this.
Your write-up should be an HTML document (here are some tips); the main document should be entitled index.html, and at the top level in the artifact turnin directory. Make sure you have all images you used for face finding and cropping, and your plots! Also remember the average face and eigenfaces in the first experiment. Make sure any images that need it are normalized so they can be seen clearly. Finally, make sure to fully document any extra credit on this web page.
Feel free to put your write-up in multiple files if you want, but make sure that they are all accessible via links from the main page.
Implement morphing; given two images, compute an animation that shows one being transformed continuously into the other, using the eigenfaces representation. Also, try transforming an image "away" from another one, to exaggerate differences between two faces (make a face more male, more female...). Make a video from your morphed faces, e.g. using Adobe Premier.
Try using
eigenfaces to recognize images other than faces or to recognize animal
faces.
Implement verifyFace as described in the list of methods to implement above. Perform the following experiment.
Extend eigenfaces to handle changes
in illumination and/or viewpoint. You can find links to additional face databases that may contain such effects here (look towards the bottom) and here.
Try using kernel PCA, or some non-linear variant of PCA. A good starting point is the following
survey paper. Compare the results using standard PCA.