How to do face recognition in OpenCV and how to improve it by aligning faces

After a long time here I come with some great news. In my semestral project for computer vision I worked on application which can improve face recognition by rotating faces so they align with training set. Without further adue here it comes:

Face recognition consists of these steps:

  1. Create training set for face recognition
  2. Load training set for face recognition
  3. Train faces and create model
  4. Capture/load image where you want to recognize people
  5. Find face/s
  6. Adjust the image for face recognition (greyscale, crop, resize, rotate …)
  7. Use trained model for face recognition
  8. Display result

This post requires you to have already set up environment with OpenCV in c++. Part with reading directories is designed for windows operating system. You can easily replace that code with your implementation suited for your operating system. If you did not setup your environment with opencv go check out some tutorials: OpenCV on Windows installation

Creating training set

To recognize faces you first need to train faces and create model for each person you want to be recognized. You can do this by manually cropping faces and adjusting them, or you can just save adjusted face from step 6 with name of the person. It is simple as that. I store this information in the name of file, which may not be the best option so there is room for improvement here.

string result;
unsigned long int sec = time(NULL);
result << "facedata/" << user << "_" << sec << ".jpg"; imwrite(result, croppedImage); capture = false;

As you can see I add timestamp to the name of the image so they have different names. And string user is read from console just like this:

user.clear(); cin >> user;

Loading training set

Working with directories in windows is a bit tricky because String is not suitable since directories and files can contain different diacritics and locales. Working with windows directories in c++ requires the use of WString for file and directories names.

vector get_all_files_names_within_folder(wstring folder)
{
  vector names;
  TCHAR search_path[200];
  StringCchCopy(search_path, MAX_PATH, folder.c_str());
  StringCchCat(search_path, MAX_PATH, TEXT("\\*"));
  WIN32_FIND_DATA fd;
  HANDLE hFind = ::FindFirstFile(search_path, &fd);
  if (hFind != INVALID_HANDLE_VALUE)
  {
    do
    {
      if (!(fd.dwFileAttributes & FILE_ATTRIBUTE_DIRECTORY))
      {
        wstring test = fd.cFileName;
        string str(test.begin(), test.end());
        names.push_back(str);
      }
    } while (::FindNextFile(hFind, &fd));
    ::FindClose(hFind);
  }
  return names;
}
void getTrainData(){
  wstring folder(L"facedata/");
  vector files = get_all_files_names_within_folder(folder);
  string fold = "facedata/";
  int i = 0;
  for (std::vector::iterator it = files.begin(); it != files.end(); ++it) {
    images.push_back(imread(fold + *it, 0));
    labelints.push_back(i);
    string str = *it;
    unsigned pos = str.find("_");
    string str2 = str.substr(0, pos);
    labels.push_back(str2);
    i++;
  }
}

I create 3 sets for face recognition and mapping. variable labelints is used in face recognition model, then its value serves as index for finding proper string representation of person and his training image.

Train faces and create model

Face recognition in opencv has three available implementations: Eiegenfaces, Fisherfaces and Local Binary Patterns Histograms. In this stage you choose which one you want to use. I found out that LBPH has the best results but is really slow. You can find out more about which one to choose in opencv’s face recognition tutorial.

void learnFacesEigen(){
  model = createEigenFaceRecognizer();
  model->train(images, labelints);
}
void learnFacesFisher(){
  model = createFisherFaceRecognizer();
  model->train(images, labelints);
}
void learnFacesLBPH(){
  model = createLBPHFaceRecognizer();
  model->train(images, labelints);
}

Capture/load image where you want to recognize people

You can load image from file as was showed before on training set or you can capture frames from your webcam. Some webcams are a bit slow and you might end up adding some sleep between initialising the camera and capturing frames. If you don’t get any frames try increasing the sleep or change the stream number.

VideoCapture stream1(1);
//-- 2. Read the video stream
if (!stream1.isOpened()){
  cout << "cannot open camera";
}
Sleep(2000);
while (true)
{
  bool test = stream1.read(frame);
  if (test)
  {
    detectAndDisplay(frame, capture, user, recognize);
  }
  else
  {
    printf(" --(!) No captured frame -- Break!"); break;
  }

}

stream1.release();

You can play with number in stream1(number) to choose the webcam you need or pick -1 to open a window with webcam selection and 0 is default.

Find face/s

Facedetection in opencv is usually done by using haar cascades. You can learn more about it in opencv’s Cascade Classifier post
Code is explained there so I will skip this part.

Adjust the image for face recognition

The most interesting part and the part where there is still much to do is this one. Face recognition in OpenCV works only on images with same size and greyscale. The more aligned faces are the better face recognition results are. So we need to convert the image to greyscale.

cvtColor(frame, frame_gray, CV_BGR2GRAY);

Then rotate face to vertical position so it is aligned perfectly. I do this by computing height difference between the eyes and when eyes are shut I use histogram of oriented gradients for nose to get its orientation. First thing first we need to find eyes on picture. I cropped the image to just the face part so the classifier has easier job finding eyes and doesn’t throw false positives.

int tlY = faces[i].y;
if (tlY < 0){
  tlY = 0;
}
int drY = faces[i].y + faces[i].height;
if (drY>frame.rows)
{
  drY = frame.rows;
}
Point tl(faces[i].x, tlY);
Point dr(faces[i].x + faces[i].width, drY);

Rect myROI(tl, dr);
Mat croppedImage_original = frame(myROI);

I tried different crops. But the best one seems to be the one with dropped out chin and a little bit of forehead which is defaultly recognized by OpenCV’s face haar classifier. Then I use different classifier to find the eyes and I decide from x position which one is left, which one is right.

eye_cascade.detectMultiScale(croppedImageGray, eyes, 1.1, 3, CV_HAAR_DO_CANNY_PRUNING, Size(croppedImageGray.size().width*0.2, croppedImageGray.size().height*0.2));

int eyeLeftX = 0;
int eyeLeftY = 0;
int eyeRightX = 0;
int eyeRightY = 0;
for (size_t f = 0; f &lt; eyes.size(); f++)
{
  int tlY2 = eyes[f].y + faces[i].y;
  if (tlY2 &lt; 0){
    tlY2 = 0;
  }
  int drY2 = eyes[f].y + eyes[f].height + faces[i].y;
  if (drY2&gt;frame.rows)
  {
    drY2 = frame.rows;
  }
  Point tl2(eyes[f].x + faces[i].x, tlY2);
  Point dr2(eyes[f].x + eyes[f].width + faces[i].x, drY2);

  if (eyeLeftX == 0)
  {

//rectangle(frame, tl2, dr2, Scalar(255, 0, 0));
    eyeLeftX = eyes[f].x;
    eyeLeftY = eyes[f].y;
  }
  else if (eyeRightX == 0)
  {

////rectangle(frame, tl2, dr2, Scalar(255, 0, 0));
    eyeRightX = eyes[f].x;
    eyeRightY = eyes[f].y;

   }

}
// if lefteye is lower than right eye swap them
if (eyeLeftX &gt; eyeRightX){
  croppedImage = cropFace(frame_gray, eyeRightX, eyeRightY, eyeLeftX, eyeLeftY, 200, 200, faces[i].x, faces[i].y, faces[i].width, faces[i].height);
}
else{
  croppedImage = cropFace(frame_gray, eyeLeftX, eyeLeftY, eyeRightX, eyeRightY, 200, 200, faces[i].x, faces[i].y, faces[i].width, faces[i].height);
}

After that I rotate the face by height difference of eyes, drop it and resize it to the same size all training data is.

Mat dstImg;
Mat crop;
if (!(eyeLeftX == 0 &amp;&amp; eyeLeftY == 0))
{

  int eye_directionX = eyeRightX - eyeLeftX;
  int eye_directionY = eyeRightY - eyeLeftY;
  float rotation = atan2((float)eye_directionY, (float)eye_directionX) * 180 / PI;
  if (rotation_def){
    rotate(srcImg, rotation, dstImg);
  }
  else {
    dstImg = srcImg;
  }
}
else
{

  if (noseDetection)
  {
    Point tl(faceX, faceY);
    Point dr((faceX + faceWidth), (faceY + faceHeight));

    Rect myROI(tl, dr);
    Mat croppedImage_original = srcImg(myROI);

    Mat noseposition_image;
    resize(croppedImage_original, noseposition_image, Size(200, 200), 0, 0, INTER_CUBIC);
    float rotation = gradienty(noseposition_image);
    if (rotation_def){
      rotate(srcImg, rotation, dstImg);
    }
    else {
      dstImg = srcImg;
    }
  }
  else{
    dstImg = srcImg;
  }

}
std::vector faces;
face_cascade.detectMultiScale(dstImg, faces, 1.1, 3, CV_HAAR_DO_CANNY_PRUNING, Size(dstImg.size().width*0.2, dstImg.size().height*0.2));

for (size_t i = 0; i &amp;lt; faces.size(); i++)
{

  int tlY = faces[i].y;
  if (tlY &lt; 0){
    tlY = 0;
  }
  int drY = faces[i].y + faces[i].height;
  if (drY&gt;dstImg.rows)
  {
    drY = dstImg.rows;
  }
  Point tl(faces[i].x, tlY);
  Point dr(faces[i].x + faces[i].width, drY);

  Rect myROI(tl, dr);
  Mat croppedImage_original = dstImg(myROI);
  Mat croppedImageGray;
  resize(croppedImage_original, crop, Size(width, height), 0, 0, INTER_CUBIC);
  imshow(&quot;test&quot;, crop);
}

As you can see I use another face detection to find cropping area. It is probably not the best option and it is not configured for more than one face, but after few enhancements it is suffitient. The next part is rotation by nose. This is purely experimental and doesn’t give very good results. I had to use average of 4 frames to determine the rotation and it is quite slow.

int plotHistogram(Mat image)
{
  Mat dst;

  /// Establish the number of bins
  int histSize = 256;

  /// Set the ranges
  float range[] = { 0, 256 };
  const float* histRange = { range };

  bool uniform = true; bool accumulate = false;

  Mat b_hist, g_hist, r_hist;
  /// Compute the histograms:
  calcHist(&amp;image, 1, 0, Mat(), b_hist, 1, &amp;histSize, &amp;histRange, uniform, accumulate);

  int hist_w = 750; int hist_h = 500;
  int bin_w = cvRound((double)hist_w / histSize);

  Mat histImage(hist_h, hist_w, CV_8UC3, Scalar(0, 0, 0));

  /// Normalize the result to [ 0, histImage.rows ]
  normalize(b_hist, b_hist, 0, histImage.rows, NORM_MINMAX, -1, Mat());
  int sum = 0;
  int max = 0;
  int now;
  int current = 0;
  for (int i = 1; i &lt; histSize; i++)
  {

    now = cvRound(b_hist.at(i));
    // ak su uhly v rozsahu 350-360 alebo 0-10 dame ich do suctu
    if ((i &lt; 5))
    {
      max += now;
      current = i;
    }

  }

  return max;

}
float gradienty(Mat frame)
{

  Mat src, src_gray;
  int scale = 1;
  int delta = 0;
  src_gray = frame;
  Mat grad_x, grad_y;
  Mat abs_grad_x, abs_grad_y;
  Mat magnitudes, angles;
  Mat bin;
  Mat rotated;
  int max = 0;
  int uhol = 0;
  for (int i = -50; i &lt; 50; i++)
  {
    rotate(src_gray, ((double)i / PI), rotated);
    Sobel(rotated, grad_x, CV_32F, 1, 0, 9, scale, delta, BORDER_DEFAULT);
    Sobel(rotated, grad_y, CV_32F, 0, 1, 9, scale, delta, BORDER_DEFAULT);
    cartToPolar(grad_x, grad_y, magnitudes, angles);
    angles.convertTo(bin, CV_8U, 90 / PI);
    Point tl((bin.cols / 2) - 10, (bin.rows / 2) - 20);
    Point dr((bin.cols / 2) + 10, (bin.rows / 2));
    Rect myROI(tl, dr);
    Mat working_pasik = bin(myROI);
    int current = 0;
    current = plotHistogram(working_pasik);
    if (current &gt; max)
    {
      max = current;
      uhol = i;
    }
  }
  noseQueue.push_back(uhol);
  int suma = 0;
  for (std::list::iterator it = noseQueue.begin(); it != noseQueue.end(); it++)
  {

    suma = suma + *it;
  }
  int priemer;
  priemer = (int)((double)suma / (double)noseQueue.size());
  if (noseQueue.size() &gt; 3)
  {
    noseQueue.pop_front();
  }
  cout &lt;&lt; priemer;
  cout &lt;&lt; &quot;\n&quot;;

  return priemer;

}

You can simply delete those “couts” those are just for debugging purposes. Main idea behind this is to compute vertical and horizontal sobel for nose part and find the angle between them. Then I determine which angle is dominant with help of histogram and I use its peak value in finding the best rotation of face. This part can be improved by normalizing the histogram on start and then using just values from one face rotation angle to determine the angle between vertical position and current angle.

Rotation is done simply by this function

void rotate(cv::Mat& src, double angle, cv::Mat& dst)
{
  int len = max(src.cols, src.rows);
  cv::Point2f pt(len / 2., len / 2.);
  cv::Mat r = cv::getRotationMatrix2D(pt, angle, 1.0);

  cv::warpAffine(src, dst, r, cv::Size(len, len));
}

And in the code earlier I showed you how to crop the face and then resize it to default size.

Use trained model for face recognition

We now have model of faces trained and enhanced image of face we want to recognize. Tricky part here is determining when the face is new (not in training set) and when it is recognized properly. So I created some sort of treshold for distance and found out that it lies between 11000 and 12000.

int predictedLabel = -1;
double predicted_confidence = 0.0;
model->set("threshold", treshold);
model->predict(croppedImage, predictedLabel, predicted_confidence);

Display result

After we found out whether the person is new or not we show results:

if (predictedLabel > -1)
{

  text = labels[predictedLabel];
  putText(frame, text, tl, fontFace, fontScale, Scalar::all(255), thickness, 8);
}

sources

Full source code of this project (I am really sorry it is so messy I don’t like c++ :D):

[cpp collapse=”true”]
#include &lt;opencv2/opencv.hpp&gt;
#include &lt;iostream&gt;
#include &lt;fstream&gt;
#include &lt;sstream&gt;
#include &lt;windows.h&gt;
#include &lt;tchar.h&gt;
#include &lt;stdio.h&gt;
#include &lt;strsafe.h&gt;
#include &lt;time.h&gt;
#include &lt;list&gt;

#define PI 3.14159265

using namespace cv;
using namespace std;

Ptr&lt;FaceRecognizer&gt; model;
vector&lt;Mat&gt; images;
vector&lt;int&gt; labelints;
vector&lt;string&gt; labels;
/**
** CONFIGURATION SECTION
**/
int method = 1; // 1 eigenfaces, 2 fisherfaces, 3 HSV
bool info = true; // Print info in console
bool rotation_def = true; // enable/disable face rotation
const int treshold_slider_max = 100000; // maximum treshold
int treshold_slider = 15000; // default treshold for face recognition
int logCountMax = 50; //each x frame will be logged
bool fixingData = false; // fixing training images
bool noseDetection = true; // rotation according to nose position
/**
*************************
**/
std::list&lt;int&gt; noseQueue;
double treshold = treshold_slider;
int logCount = 0;
void detectAndDisplay(Mat frame, bool &amp;capture, string user, bool recognize);
void noseposition(Mat frame);
void learnFacesEigen();
void learnFacesFisher();
float gradienty(Mat frame);
void learnFacesLBPH();
void getTrainData();
void fixData();
Mat cropFace(Mat srcImg, int eyeLeftX, int eyeLeftY, int eyeRightX, int eyeRightY, int width, int height, int faceX, int faceY, int faceWidth, int faceHeight);

String face_cascade_name = &quot;haarcascade_frontalface_alt.xml&quot;;
String eye_cascade_name = &quot;haarcascade_eye.xml&quot;;

CascadeClassifier face_cascade;
CascadeClassifier nose_cascade;
CascadeClassifier eye_cascade;
string window_name = &quot;Capture – Face detection&quot;;

ofstream logFile;

void on_trackbar(int, void*)
{
treshold = treshold_slider;
}
void callbackButton(int, void*)
{

}

static Mat norm_0_255(InputArray _src) {
Mat src = _src.getMat();
// Create and return normalized image:
Mat dst;
switch (src.channels()) {
case 1:
cv::normalize(_src, dst, 0, 255, NORM_MINMAX, CV_8UC1);
break;
case 3:
cv::normalize(_src, dst, 0, 255, NORM_MINMAX, CV_8UC3);
break;
default:
src.copyTo(dst);
break;
}
return dst;
}

int main(int argc, const char** argv)
{

Mat image;
Mat frame;

logFile = ofstream(&quot;logs.csv&quot;, std::ios_base::app);
if (!logFile.is_open())
{
cout &lt;&lt; &quot;Cannot open log file&quot;;
return 0;
}

//– 1. Load the cascades
if (!face_cascade.load(face_cascade_name)){ printf(&quot;–(!)Error loading\n&quot;); return -1; };
if (!eye_cascade.load(eye_cascade_name)){ printf(&quot;–(!)Error loading\n&quot;); return -1; };
// Get train data
getTrainData();

if (labels.size() &gt; 0)
{
//learn faces
if (method == 1)
{
learnFacesEigen();

}
else if
(method == 2){
learnFacesFisher();
}
else{
learnFacesLBPH();
}
}

VideoCapture stream1(1);
//– 2. Read the video stream
if (!stream1.isOpened()){
cout &lt;&lt; &quot;cannot open camera&quot;;
}

Sleep(2000);
bool capture = false;
bool recognize = false;
/*
** GUI
*/
namedWindow(window_name, 1);

char TrackbarName[50];
sprintf(TrackbarName, &quot;Treshold %d&quot;, treshold_slider_max);
// trackbar for treshold
createTrackbar(TrackbarName, window_name, &amp;treshold_slider, treshold_slider_max, on_trackbar);
string user;
/*fixingData = true;
fixData();
fixingData = false;*/
while (true)
{
bool test = stream1.read(frame);

//– 3. Apply the classifier to the frame
if (test)
{
detectAndDisplay(frame, capture, user, recognize);
}
else
{
printf(&quot; –(!) No captured frame — Break!&quot;); break;
}

int c = waitKey(10);
if ((char)c == ‘n’){
user.clear();
cin &gt;&gt; user;
}
// press r for face recognition
if ((char)c == ‘r’){ recognize = !recognize; }
// press a for face capture
if ((char)c == ‘a’){ capture = true; }
// press e to print out eigenfaces
if ((char)c == ‘e’) { info = false; }
//press c to exit
if ((char)c == ‘c’) { break; }
}

stream1.release();
logFile.close();
waitKey(0);
return 0;
}

void detectAndDisplay(Mat frame, bool &amp;capture, string user, bool recognize)
{
std::vector&lt;Rect&gt; faces;
std::vector&lt;Rect&gt; hss;
Mat frame_gray;
Mat croppedImage;
Mat workingImage;
int fontFace = FONT_HERSHEY_PLAIN;
double fontScale = 2;
int thickness = 1;
if (!fixingData)
{

cvtColor(frame, frame_gray, CV_BGR2GRAY);
//equalizeHist(frame_gray, frame_gray);
}
else
{

frame_gray = frame;
}

//– Detect faces
face_cascade.detectMultiScale(frame_gray, faces, 1.1, 3, CV_HAAR_DO_CANNY_PRUNING, Size(frame.size().width*0.2, frame.size().height*0.2));

for (size_t i = 0; i &lt; faces.size(); i++)
{

int tlY = faces[i].y;
//tlY -= (faces[i].height / 3);
if (tlY &lt; 0){
tlY = 0;
}
int drY = faces[i].y + faces[i].height;
//drY += +(faces[i].height / 6);
if (drY&gt;frame.rows)
{
drY = frame.rows;
}
Point tl(faces[i].x, tlY);
Point dr(faces[i].x + faces[i].width, drY);

Rect myROI(tl, dr);
Mat croppedImage_original = frame(myROI);
Mat croppedImageGray;

//resize(croppedImage_original, croppedImage_original, Size(200, 292), 0, 0, INTER_CUBIC);
if (!fixingData)
{
cvtColor(croppedImage_original, croppedImageGray, CV_RGB2GRAY);
}
else
{
croppedImageGray = croppedImage_original;
}

std::vector&lt;Rect&gt; eyes;
// detect eyes
eye_cascade.detectMultiScale(croppedImageGray, eyes, 1.1, 3, CV_HAAR_DO_CANNY_PRUNING, Size(croppedImageGray.size().width*0.2, croppedImageGray.size().height*0.2));

int eyeLeftX = 0;
int eyeLeftY = 0;
int eyeRightX = 0;
int eyeRightY = 0;
for (size_t f = 0; f &lt; eyes.size(); f++)
{
int tlY2 = eyes[f].y + faces[i].y;
if (tlY2 &lt; 0){
tlY2 = 0;
}
int drY2 = eyes[f].y + eyes[f].height + faces[i].y;
if (drY2&gt;frame.rows)
{
drY2 = frame.rows;
}
Point tl2(eyes[f].x + faces[i].x, tlY2);
Point dr2(eyes[f].x + eyes[f].width + faces[i].x, drY2);

if (eyeLeftX == 0)
{

//rectangle(frame, tl2, dr2, Scalar(255, 0, 0));
eyeLeftX = eyes[f].x;
eyeLeftY = eyes[f].y;
}
else if (eyeRightX == 0)
{

////rectangle(frame, tl2, dr2, Scalar(255, 0, 0));
eyeRightX = eyes[f].x;
eyeRightY = eyes[f].y;

}

}
// if lefteye is lower than right eye swap them
if (eyeLeftX &gt; eyeRightX){
croppedImage = cropFace(frame_gray, eyeRightX, eyeRightY, eyeLeftX, eyeLeftY, 200, 200, faces[i].x, faces[i].y, faces[i].width, faces[i].height);
}
else{
croppedImage = cropFace(frame_gray, eyeLeftX, eyeLeftY, eyeRightX, eyeRightY, 200, 200, faces[i].x, faces[i].y, faces[i].width, faces[i].height);
}

if (capture &amp;&amp; !user.empty())
{
// save captured image to file
string result;
unsigned long int sec = time(NULL);
result &lt;&lt; &quot;facedata/&quot; &lt;&lt; user &lt;&lt; &quot;_&quot; &lt;&lt; sec &lt;&lt; &quot;.jpg&quot;;
imwrite(result, croppedImage);
capture = false;
}
if (recognize&amp;&amp;!croppedImage.empty())
{
//recognize face
logCount++;
int predictedLabel = -1;
double predicted_confidence = 0.0;
model-&gt;set(&quot;threshold&quot;, treshold);
model-&gt;predict(croppedImage, predictedLabel, predicted_confidence);
string text;
if (logCount &gt; logCountMax)
{
if (predictedLabel &gt; -1)
{
// log to file
text = labels[predictedLabel];
logCount = 0;
logFile &lt;&lt; text;
logFile &lt;&lt; &quot;;&quot;;
logFile &lt;&lt; predicted_confidence;
logFile &lt;&lt; &quot;\n&quot;;
}
}
cout &lt;&lt; predicted_confidence;
cout &lt;&lt; &quot;\n&quot;;
if (!info&amp;&amp;method == 1)
{
// save Eigenvectors to file
info = true;
// Here is how to get the eigenvalues of this Eigenfaces model:
Mat eigenvalues = model-&gt;getMat(&quot;eigenvalues&quot;);
// And we can do the same to display the Eigenvectors (read Eigenfaces):
Mat W = model-&gt;getMat(&quot;eigenvectors&quot;);
// Get the sample mean from the training data
Mat mean = model-&gt;getMat(&quot;mean&quot;);
// Display or save:
imwrite(format(&quot;%s/mean.png&quot;, &quot;eigenfaces&quot;), norm_0_255(mean.reshape(1, 292)));
// Display or save the Eigenfaces:
for (int i = 0; i &lt; min(10, W.cols); i++) {
string msg = format(&quot;Eigenvalue #%d = %.5f&quot;, i, eigenvalues.at&lt;double&gt;(i));
cout &lt;&lt; msg &lt;&lt; endl;
// get eigenvector #i
Mat ev = W.col(i).clone();
// Reshape to original size &amp; normalize to [0…255] for imshow.
Mat grayscale = norm_0_255(ev.reshape(1, 292));
// Show the image &amp; apply a Jet colormap for better sensing.
Mat cgrayscale;
applyColorMap(grayscale, cgrayscale, COLORMAP_JET);
// Display or save:

imwrite(format(&quot;%s/eigenface_%d.png&quot;, &quot;eigenfaces&quot;, i), norm_0_255(cgrayscale));

}
}

// display name of recognized person
if (predictedLabel &gt; -1)
{

text = labels[predictedLabel];
putText(frame, text, tl, fontFace, fontScale, Scalar::all(255), thickness, 8);
}

}
rectangle(frame, tl, dr, Scalar(255, 0, 255));

}

imshow(window_name, frame);
}
std::wstring s2ws(const std::string&amp; s)
{
int len;
int slength = (int)s.length() + 1;
len = MultiByteToWideChar(CP_ACP, 0, s.c_str(), slength, 0, 0);
wchar_t* buf = new wchar_t[len];
MultiByteToWideChar(CP_ACP, 0, s.c_str(), slength, buf, len);
std::wstring r(buf);
delete[] buf;
return r;
}

vector&lt;string&gt; get_all_files_names_within_folder(wstring folder)
{
vector&lt;string&gt; names;
TCHAR search_path[200];
StringCchCopy(search_path, MAX_PATH, folder.c_str());
StringCchCat(search_path, MAX_PATH, TEXT(&quot;\\*&quot;));
WIN32_FIND_DATA fd;
HANDLE hFind = ::FindFirstFile(search_path, &amp;fd);
if (hFind != INVALID_HANDLE_VALUE)
{
do
{
// read all (real) files in current folder
// , delete ‘!’ read other 2 default folder . and ..
if (!(fd.dwFileAttributes &amp; FILE_ATTRIBUTE_DIRECTORY))
{
//_tprintf(TEXT(&quot; %s &lt;DIR&gt;\n&quot;), fd.cFileName);
wstring test = fd.cFileName;
string str(test.begin(), test.end());
names.push_back(str);
}
} while (::FindNextFile(hFind, &amp;fd));
::FindClose(hFind);
}
return names;
}
void getTrainData(){
wstring folder(L&quot;facedata/&quot;);
vector&lt;string&gt; files = get_all_files_names_within_folder(folder);
string fold = &quot;facedata/&quot;;
int i = 0;
for (std::vector&lt;string&gt;::iterator it = files.begin(); it != files.end(); ++it) {
images.push_back(imread(fold + *it, 0));
labelints.push_back(i);
string str = *it;
unsigned pos = str.find(&quot;_&quot;);
string str2 = str.substr(0, pos);
labels.push_back(str2);
i++;
}
}
void fixData(){
Mat fixImg;
wstring folder(L&quot;facedata2/&quot;);
vector&lt;string&gt; files = get_all_files_names_within_folder(folder);
string fold = &quot;facedata2/&quot;;
for (std::vector&lt;string&gt;::iterator it = files.begin(); it != files.end(); ++it) {
fixImg = imread(fold + *it, 0);
string str = *it;
unsigned pos = str.find(&quot;_&quot;);
string str2 = str.substr(0, pos);
bool capture = true;
detectAndDisplay(fixImg, capture, str2, true);
waitKey(1000);
}
}
void learnFacesEigen(){

model = createEigenFaceRecognizer();
model-&gt;train(images, labelints);
}
void learnFacesFisher(){
model = createFisherFaceRecognizer();
model-&gt;train(images, labelints);
}
void learnFacesLBPH(){
model = createLBPHFaceRecognizer();
model-&gt;train(images, labelints);
}
void rotate(cv::Mat&amp; src, double angle, cv::Mat&amp; dst)
{
int len = max(src.cols, src.rows);
cv::Point2f pt(len / 2., len / 2.);
cv::Mat r = cv::getRotationMatrix2D(pt, angle, 1.0);

cv::warpAffine(src, dst, r, cv::Size(len, len));
}

int plotHistogram(Mat image)
{
Mat dst;

/// Establish the number of bins
int histSize = 256;

/// Set the ranges
float range[] = { 0, 256 };
const float* histRange = { range };

bool uniform = true; bool accumulate = false;

Mat b_hist, g_hist, r_hist;
/// Compute the histograms:
calcHist(&amp;image, 1, 0, Mat(), b_hist, 1, &amp;histSize, &amp;histRange, uniform, accumulate);

int hist_w = 750; int hist_h = 500;
int bin_w = cvRound((double)hist_w / histSize);

Mat histImage(hist_h, hist_w, CV_8UC3, Scalar(0, 0, 0));

/// Normalize the result to [ 0, histImage.rows ]
normalize(b_hist, b_hist, 0, histImage.rows, NORM_MINMAX, -1, Mat());
int sum = 0;
int max = 0;
int now;
int current = 0;
for (int i = 1; i &lt; histSize; i++)
{
/*line(histImage, Point(bin_w*(i – 1), hist_h – cvRound(b_hist.at&lt;float&gt;(i – 1))),
Point(bin_w*(i), hist_h – cvRound(b_hist.at&lt;float&gt;(i))),
Scalar(255, 255, 255), 2, 8, 0);*/
now = cvRound(b_hist.at&lt;float&gt;(i));
// ak su uhly v rozsahu 350-360 alebo 0-10 dame ich do suctu
if ((i &lt; 5))
{
max += now;
current = i;
}
/*if ((i &gt; 175))
{
max += now;
current = i;
}*/

}

return max;

}
float gradienty(Mat frame)
{

Mat src, src_gray;
int scale = 1;
int delta = 0;
src_gray = frame;
/// Generate grad_x and grad_y
Mat grad_x, grad_y;
Mat abs_grad_x, abs_grad_y;
Mat magnitudes, angles;
Mat bin;
Mat rotated;
int max = 0;
int uhol = 0;
// otocime obrazok do uhlov v ktorych je este mozne tvar detekovat
for (int i = -50; i &lt; 50; i++)
{
rotate(src_gray, ((double)i / PI), rotated);
// aplikujeme vertikalny a horizontalny sobel
Sobel(rotated, grad_x, CV_32F, 1, 0, 9, scale, delta, BORDER_DEFAULT);
Sobel(rotated, grad_y, CV_32F, 0, 1, 9, scale, delta, BORDER_DEFAULT);
// vypocitame uhly
cartToPolar(grad_x, grad_y, magnitudes, angles);
// skonvertujeme do stupnov sedej pre lepsiu reprezentaciu 1 stupenm sedej = 2 stupne (este mozno zmenit!!!)
angles.convertTo(bin, CV_8U, 90 / PI);
// vyrezeme nos ktory sa nachadza v strede hlavy +-
Point tl((bin.cols / 2) – 10, (bin.rows / 2) – 20);
Point dr((bin.cols / 2) + 10, (bin.rows / 2));
Rect myROI(tl, dr);
Mat working_pasik = bin(myROI);
int current = 0;
// vypocitame histogram a pocet uhlov v norme
current = plotHistogram(working_pasik);
// vyberieme maximum
if (current &gt; max)
{
max = current;
uhol = i;
}
}
//pridame uhol do queue
noseQueue.push_back(uhol);
int suma = 0;
// spocitame vsetky uhly v queue
for (std::list&lt;int&gt;::iterator it = noseQueue.begin(); it != noseQueue.end(); it++)
{

suma = suma + *it;
}
//cout &lt;&lt; &quot;suma &quot; &lt;&lt; suma &lt;&lt; &quot;\n&quot;;
int priemer;
// vypocitame priemerny uhol za posledne 4 snimky
priemer = (int)((double)suma / (double)noseQueue.size());
// ak je moc v queue posledne odstranime
if (noseQueue.size() &gt; 3)
{
noseQueue.pop_front();
}
// vypiseme priemer
cout &lt;&lt; priemer;
cout &lt;&lt; &quot;\n&quot;;

return priemer;

}

Mat cropFace(Mat srcImg, int eyeLeftX, int eyeLeftY, int eyeRightX, int eyeRightY, int width, int height, int faceX, int faceY, int faceWidth, int faceHeight){
Mat dstImg;
Mat crop;
// Ak sa detekuju oci, otacaj podla vyskoveho rozdielu oci
if (!(eyeLeftX == 0 &amp;&amp; eyeLeftY == 0))
{

int eye_directionX = eyeRightX – eyeLeftX;
int eye_directionY = eyeRightY – eyeLeftY;
float rotation = atan2((float)eye_directionY, (float)eye_directionX) * 180 / PI;
if (rotation_def){
rotate(srcImg, rotation, dstImg);
}
else {
dstImg = srcImg;
}
} // Ak sa oci nedetekuju a je zapnuta detekcia nosa otacaj podla histogramu orientovanych gradientov nosa
else
{

if (noseDetection)
{
Point tl(faceX, faceY);
Point dr((faceX + faceWidth), (faceY + faceHeight));

Rect myROI(tl, dr);
Mat croppedImage_original = srcImg(myROI);

Mat noseposition_image;
//Zmensime oblast orezanim a dame ju na standardnu velkost
resize(croppedImage_original, noseposition_image, Size(200, 200), 0, 0, INTER_CUBIC);
// vypocitame uhol rotacie pomocou histogramu orientovanych gradientov
float rotation = gradienty(noseposition_image);
if (rotation_def){
rotate(srcImg, rotation, dstImg);
}
else {
dstImg = srcImg;
}
}
else{
dstImg = srcImg;
}

}
std::vector&lt;Rect&gt; faces;
// Este raz detekujeme tvar na otocenom obrazku kvoli lepsej presnosti rozpoznavania tvare
face_cascade.detectMultiScale(dstImg, faces, 1.1, 3, CV_HAAR_DO_CANNY_PRUNING, Size(dstImg.size().width*0.2, dstImg.size().height*0.2));

for (size_t i = 0; i &lt; faces.size(); i++)
{

int tlY = faces[i].y;
//tlY -= (faces[i].height / 3);
if (tlY &lt; 0){
tlY = 0;
}
int drY = faces[i].y + faces[i].height;
//drY += +(faces[i].height / 6);
if (drY&gt;dstImg.rows)
{
drY = dstImg.rows;
}
Point tl(faces[i].x, tlY);
Point dr(faces[i].x + faces[i].width, drY);

Rect myROI(tl, dr);
// Orezeme na ksicht
Mat croppedImage_original = dstImg(myROI);
Mat croppedImageGray;
// zmensime na standardnu velkost
resize(croppedImage_original, crop, Size(width, height), 0, 0, INTER_CUBIC);
//face_cascade.detectMultiScale(dstImg, faces, 1.1, 3, CV_HAAR_DO_CANNY_PRUNING, Size(dstImg.size().width*0.2, dstImg.size().height*0.2));
imshow(&quot;test&quot;, crop);
}

return crop;
}
[/cpp]

11 Comments

  1. Frans

    Hey, thanks for sharing!! it helps me to learn more about face recognition using C++ !
    I want to ask you, why didn’t you use CSV to predict the Users? and this code couldn’t recognize more than 2 people?
    Thanks a lot 🙂

    Reply
    1. drndos (Post author)

      Hello, csv would be a great idea. This was school project so I focused on aligning faces instead of making it perfectly usable :D.
      As far as I remember it should work with more than two faces.

      Reply
  2. tj

    hi dude,

    can you give csv file to run face recognition

    Reply
  3. crepus

    Hey I was using Opencv for face recognition on Android. In the training part, I thought the idea was to assign same label to different images of same person, but every time I do that prediction just goes wrong. Do you have any idea where the problem might be?

    Reply
  4. hosein

    hi
    thanks for your useful post
    but i give lots of error in this two include and below functions

    #include
    #include

    std::wstring s2ws(const std::string& s)
    {
    int len;
    int slength = (int)s.length() + 1;
    len = MultiByteToWideChar(CP_ACP, 0, s.c_str(), slength, 0, 0);
    wchar_t* buf = new wchar_t[len];
    MultiByteToWideChar(CP_ACP, 0, s.c_str(), slength, buf, len);
    std::wstring r(buf);
    delete[] buf;
    return r;
    }

    vector get_all_files_names_within_folder(wstring folder)
    {
    vector names;
    TCHAR search_path[200];
    StringCchCopy(search_path, MAX_PATH, folder.c_str());
    StringCchCat(search_path, MAX_PATH, TEXT(“\\*”));
    WIN32_FIND_DATA fd;
    HANDLE hFind = ::FindFirstFile(search_path, &fd);
    if (hFind != INVALID_HANDLE_VALUE)
    {
    do
    {
    // read all (real) files in current folder
    // , delete ‘!’ read other 2 default folder . and ..
    if (!(fd.dwFileAttributes & FILE_ATTRIBUTE_DIRECTORY))
    {
    //_tprintf(TEXT(” %s \n”), fd.cFileName);
    wstring test = fd.cFileName;
    string str(test.begin(), test.end());
    names.push_back(str);
    }
    } while (::FindNextFile(hFind, &fd));
    ::FindClose(hFind);
    }
    return names;
    }
    i am in fedora linux,what that seems this code work only on windows

    Reply
    1. drndos (Post author)

      Hello, yes this is part of code for windows to read files and to do that I had to convert some special windows strings, you don’t need this function on linux. Just google how to read all files in directory in linux in c and replace that function.

      Reply
  5. blaisexen

    I used Delphi-OpenCV, source code can help.

    thank you

    Reply
  6. blaisexen

    Source code has many errors:
    eq.,
    if (eyeLeftX == 0)
    {
    //rectangle(frame, tl2, dr2, Scalar(255, 0, 0));
    eyeLeftX = eyes[f].x;
    eyeLeftY = eyes[f].y;
    }
    else if (eyeRightX == 0)
    {
    ////rectangle(frame, tl2, dr2, Scalar(255, 0, 0));
    eyeRightX = eyes[f].x;
    eyeRightY = eyes[f].y;
    }

    should be:
    if (eyeLeftX == 0)
    {
    eyeLeftX = 0;
    eyeLeftY = eyes[f].y;
    eyeRightX = eyes[f].x;
    eyeRightY = eyes[f].y;
    }
    else if (eyeRightX == 0)
    {
    eyeRightX = 0;
    eyeRightY = eyes[f].y;
    eyeLeftX = eyes[f].x;
    eyeLeftY = eyes[f].y;
    }

    and lot’s of errors, I guess source code was hi-jack!

    Reply
  7. blaisexen

    now, source code is very wrong,
    int eyeLeftX = 0;
    int eyeLeftY = 0;
    int eyeRightX = 0;
    int eyeRightY = 0;
    if (eyeLeftX == 0)

    else if (eyeRightX == 0)

    //will alwasy return 0;

    Reply
  8. Just here Reading

    The codes are compromised by the thread Master! to keep your minds twist. lol

    Reply
  9. ben

    Is it possible to do a training set with a single image?

    Reply

Leave a Reply to blaisexen Cancel reply

Your email address will not be published. Required fields are marked *