SlideShare uma empresa Scribd logo
1 de 5
Baixar para ler offline
The International Journal Of Engineering And Science (IJES)
|| Volume || 3 || Issue || 5 || Pages || 18-22 || 2014 ||
ISSN (e): 2319 – 1813 ISSN (p): 2319 – 1805
www.theijes.com The IJES Page 18
 Process Automation of Human Emotion Detection
1,
Namrata Mahajan, 2,
Harshad Mahajan
1,
Department of Electronics and Communication, VIT Nagpur
2,
Department of Electronics and Communication, GCOE Amravati
--------------------------------------------------------------ABSTRACT------------------------------------------------------
Human can see emotion as associated with mood, temperament, personality and disposition. To detect
emotions is easy for humans but it’s quite difficult for computer as world is three dimensional but Computer
has only two dimensions. Computer seeks to emulate the human emotions by digital image analysis. Humans
can use vision to identify objects quickly and accurately. Human can detect emotion using voice based on
different parameters like tone, pitch, pace and volume; but in case of digital images detecting emotion just by
analysing images is a novel way.
This algorithm has two major parts. First, Template database generation and another is emotion
detection. In this, we first extracts face from an image using some basic image processing operations and
color models in it. Here we define thresholds to separate our region of interest. Then we perform lip detection
on cropped face and this extracted lips are stored in our database with emotion name in database generation
phase while in emotion detection phase this extracted lips are compared with series of stored template in
database and on the basis of best correlated template emotion is recognized. This method of detecting
emotions is simple and fast as compared to previous methods i.e. using brain activity or by speech. Size of
database will affect the effectiveness of the proposed algorithm.
Keywords - Emotion Detection, Skin based Segmentation, Automation and Lip Segmentation
--------------------------------------------------------------------------------------------------------------------------------------
Date of Submission: 02 May 2014 Date of Publication: 15 May 2014
--------------------------------------------------------------------------------------------------------------------------------------
I. INTRODUCTION
Human has ability to see emotions by temperament, disposition, personality and mood. Computer
seeks to emulate the human emotions by digital image analysis. The problem with the computer vision is due
to the fact that world is three dimensional but Computer has only two dimensions.
In our day to day life we interact with each other directly (for e.g. face to face) or indirectly (for e.g.
phone calls). In some profession interaction like call centers interaction with people is important. With great
advancement in technology in terms of different techniques of people interacting with each other it is quite
necessary that one should be aware of current emotions of the person he/she is interacting.
It is widely accepted from psychological theory that human emotions can be classified into six archetypal
emotions: surprise, fear, disgust, anger, happiness, and sadness. Facial motion and the tone of the speech play
a major role in expressing these emotions.
The muscles of the face can be changed and the tone and the energy in the production of the speech
can be intentionally modified to communicate different feelings. Human beings can recognize these signals
even if they are subtly displayed, by simultaneously processing information acquired by ears and eyes. Based
on psychological studies, which show that visual information modifies the perception of speech [1], it is
possible to assume that human emotion perception follows a similar trend.
This paper analyses the use of digital image to recognize four different human emotions: sadness,
happiness, anger and neutral state, using a database recorded .the more challenging task of capturing salient
visual information directly from conventional videos is a topic for future work but is hoped to be informed by
studies such as in this report. The primary purpose of this paper is to define a simple and fast automated
emotion recognition system.
Process Automation of Human Emotion Detection
www.theijes.com The IJES Page 19
II. PROCESS AUTOMATION
In order to implement an automated emotion detection system, a series of logical steps have to be
developed. The flow of steps in this automated system is defined below. Our project is mainly divided in two
parts Database Generation (Fig.1) and Emotion Recognition (Fig.2). In Emotion recognition we are going to
match resulting output of input image with our templates which already exist in our database.
A. Face Detection
We faced many problems related to the face detection. To locate the face image should consists of
single face. Hence in this paper our assumption is that an input image is having single face. If this assumption
is not satisfied it will take either of any one image from input not any specific one. Although we provide a
single face image for face detection but face detection means identifying a region of image which contains face
no matter what is the background of image, its lighting conditions or orientation. This task is bit critical as
different faces have different colours, shapes and texture.
At this time, many approaches involve automatic face detection followed by lip detection .But most of
approaches define the scope of upright view of faces such approaches failed to detect faces rotated in plane or
out of image plane. Also sometimes it failed while detecting faces from image when only part of face is there in
image.
Hence in this project we are using thresholding to separate skin region from image in order to detect face.
1. Skin Color Classification
Skin color and textures are important cues that people use consciously or unconsciously to infer
variety of culture-related aspects about each other. Skin color and texture can be an indication of race, health,
age, wealth, beauty, etc. [3]. However, such interpretations vary across cultures and across the history. In
images, skin color is an indication of the existence of humans in such media.
The color of human skin is created by a combination of blood (red) and melanin (yellow, brown). Skin
colors lie between these two extreme hues and are somewhat saturate Facial skin takes on is clearly a subspace
of the total color space. The human skin is a fraction of the actual color cube, about 0.25 % of the total colors.
Except for extremely hairy subjects, which are rare, skin has only low-amplitude texture. Three color spaces
i.e. RGB, HIS and YCrCb are defined in literature and they are useful to achieve our goal.
2. Skin Based Segmentation
To detect Human faces, their skin color is unique and efficient feature. Although skin color may vary
person to person, major difference lies largely between their intensity rather than their chrominance. Most of
the face detection algorithms are on the basis of skin based segmentation. Many of these algorithms uses the
algorithms use a range of color values to define skin color.
Process Automation of Human Emotion Detection
www.theijes.com The IJES Page 20
3. Segmentation Rules
Segmentation procedures partition an image into its constituent parts or objects. A rugged
segmentation procedure brings the process a long way toward successful solution of imaging problems that
require objects to be identified individually. In general, the more accurate the segmentation, the more likely
recognition is to succeed. Segmentation rules determine the formation of regions. These rules can be defined
on the basis of color and edge properties of region. Using skin based segmentation we can reduce the efforts of
first determining faces within image in which faces are out of frame or tilted.
4. Segmentation Rules for different Colour Models
After some experiments we found following rules for different color models which give one step
closure to our goal.
RGB Model: 0.836G – 14 < B < 0.836G + 44 => Skin OR 0.79G -67 < B < 0.78G + 42 => Skin
HIS/HSV Model: 19 < H < 240 => Not skin
YCrCb Model: 140 < Cr < 165 => Skin OR 140 < Cr < 195 => Skin
Fig 3.Orignal Image Fig.4 Image after Skin Thresholding
5. Erosion and Dilation
After skin based segmentation, for fast processing we convert our output image into binary image.
The noises in this black and white image are removed using suitable window sizes. The pixels in this image
which were falsely detected as skin color pixels are removed by erosion and dilation by moving window over
image matrix. If it is found that the number of white pixels is more than fixed threshold then whole window is
made black. The face detection is performed on skin detected areas. Hence to locate face we always consider
the occurrence of white region in output image after erosion and dilation. This process has assumption that
image on which we are going to perform process should have single face with minimum skin exposure.
Skin colors in this project are using RGB, HIS/HSV and YCrCb models. Values are unique for this
project obtained using fresh toned digital photos. For larger skin tone range HIS model gives better results but
these models made it difficult to constraint color for flesh-like tones. RGB models are preferred in such
situations.
Fig.5 Image after Dilation Fig. 6 Image after Erosion
Process Automation of Human Emotion Detection
www.theijes.com The IJES Page 21
B. Cropping Facial Region
After detection of Face we are going to extract lip region .We are aware about the thing that lip region
contents maximum of red color content. If red color content is present in the image outside the facial region, it
will interfere in our lip extraction process it may also cause increase in time and complexity. Thus we decide to
crop the region of our interest only. So focusing on this limited area, results will be more accurate. Also
Computational time is directly proportional to the size of image. If processing image size is reduced due to
cropping region of interest, computational time will also reduce.
Before cropping we used to apply erosion and dilation operations on detected face image. These
functions help us to remove noise which is quite helpful for region cropping. Major work in cropping is to find
four boundary points. We found first top, bottom, right and left four points. On the basis of these four points
height and width of detected region can be calculated. After calculating height and width we create rectangle
and then crop required area of facial region.
C. Lip Detection
After face detection, we have our region of interest; on this image we applied our most important
algorithm of lip detection. Human emotions reflect changes in lip region and eyes. But maximum changes
around lips are visible rather than eyes. Hence a detected lip gives us accurate results.
After face detection and getting our region of interest, we now move towards the most important part
of the project i.e. lip detection. As already discussed in the 1st chapter, the variations seen during different
emotions are around the lip region as well as the eyes region. But the maximum changes that are visible will be
around the lip region, thus giving us more accurate results.
Same as we used RGB thresholds in skin based se3gmentations to detect face. Here we used
thresholds on red color as lip region contains maximum of red colors.
In field of computer vision, automation of the process of locating and tracking person‟s lip is
somewhat difficult task. There several methods proposed for lip detection in last 10-12 years. But we found
important and effective way of detecting lip based on the segmentation directly from the color space. In this
algorithm we can enlarge difference between skin and lip using color filters or color transformations.
The processing time in this algorithm is main advantage .Lip segmentation is very sensitive to
changes in colors. Hence it may create problem if for some faces lip and skin contrast is lower.
Fig. 7 Image after Lip Detection
D. Lip Segmentation
After completion of lip detection, now our next step is lip segmentation. After this segmentation we
will use result of lip segmentation to generate template if we are creating database otherwise it will be used to
detect emotion if we are in emotion detection phase of project. Thus this cropped lip region makes us closer to
our aim.
Again the process of lip segmentation is same as face detection .In which we need to define first four boundary
points i.e. top, bottom, left and right points of detected lip. From this point we will find width and height of lip
region which is further used to create rectangle and extract lip region.
It may possible in case of noise occurs on the image, it will create confusion between lips and any
other area on face. To avoid this confusion we used a fact that width of height is 3 times than lips height. But
for some worst case we modified this condition as lip width should be greater than the twice of its width, if this
condition is true then we create rectangle window and crop lip region.
E. Template Database Generation
Our project is mainly divided in two parts. 1. Database Generation and 2. Emotion Recognition. In
Emotion recognition we are going to match resulting output of input image with our templates which already
exist in our database. But for this we need to create database. In order to generate database, templates have to
be generated. Roipoly function is used to create template. It is out of box function in MATLAB. Roiply
function specifies the polygonal region of interest (ROI) within an image. Its output is in the form of binary
image which is used as mask for filtering.
Process Automation of Human Emotion Detection
www.theijes.com The IJES Page 22
BW = roipoly creates an interactive polygon tool, associated with the image displayed in the current
figure, called the target image. With the polygon tool active, the pointer changes to cross hairs,„+ „ when you
move the pointer over the image in the figure. Using the mouse, you specify the region by selecting vertices of
the polygon or any area of interest. You can move or resize the polygon or figure using the mouse. When you
are finished positioning and sizing the polygon/figure, create the mask by double-clicking, or by right-clicking
inside the region and selecting Create mask from the context menu. Roipoly returns the mask as a binary
image, BW, the same size as I. In the mask image, roipoly sets pixels inside the region to 1 and pixels outside
the region to 0. This binary image BW so returned is in logical format and as such cannot be used for
implementation. It needs to be converted to a format that can be processed on i.e. double or uint8. After this
the masks are cropped in the similar manner used for face cropping and lip cropping i.e. finding four points
and creating rectangle and finally cropping.[5]
F. Emotion Recognition
In Emotion recognition we are going to match resulting output of input image with our templates
which already exist in our database. This comparison is made by using cross-correlation.
This option calculates the cross correlation function for two images. Each of the images is divided into
rectangular blocks. Each block in the first image is correlated with its corresponding block in the second image
to produce the cross correlation as a function of position.
The cross correlation may be used to determine the degree of similarity between two similar images,
or, with the addition of a linear offset to one of the images, the spatial shift or spatial correlation between the
images.
G. Emotion Detection
After finding cross-correlation of the detected and cropped lip region with templates, we have to find
the index number on which we get maximum value offered by cross-correlation of two images. Emotion is
detected by checking the corresponding emotion across the index number with maximum value of cross-
correlation
III. FINAL REMARKS
We found that this algorithm is faster and easy to implement than existing algorithm based on speech
recognition and by ECG in which physical contact with human brain is needed which is quite expensive than
this. We worked on five emotions i.e. .happiness, grief, anger, surprise and neutral. These emotions give us
successful results on majority of images. Generally no false emotion were found or wrongly interpreted.
REFERENCES
[1] “Human Emotion Recognition using Frequency & Statistical Measures of EEG Signal” , Monira Islam, Tazrin Ahmed, Sheikh Shanawaz
Mostafa*, Md. Salah Uddin Yusuf, and Mohiuddin Ahmad
[2] “Human Emotion Recognition using Frequency & Statistical Measures of EEG Signal”, Monira Islam, Tazrin Ahmed, Sheikh Shanawaz
Mostafa*, Md. Salah Uddin Yusuf, and Mohiuddin Ahmad
[3] Bernhard Fink, K.G., Matts, P.J.: Visible skin color distribution plays a role in the perception of age, attractiveness, and health in
female faces. Evolution and Human Behavior 27(6) (2006) 433–442
[4] Jones, M.J., Rehg, and J.M.: Statistical color models with application to skin detection. International Journal of Computer Vision
(IJCV) 46(1) (2002) 81–96
[5] http://www.mathworks.in/help/images/ref/roipoly.html
[6] A. Yuille, “Feature Extraction from Faces Using Deformable Templates,” Int. Journal of Computer Vision,(1992), 8(2):99-111.
[7] Bracewell R.N.: “The Hartley transform”, Oxford University Press, 1986.
[8] Gocke R., Millar J.B., Zelinsky A., Robert-Ribes J, “Automatic extraction of lip feature points”. Proc. Austral. Conf. Robotics and
Automation, Melbourne, Australia, August 2000, pp. 31–36.

Mais conteúdo relacionado

Mais procurados

Research Inventy : International Journal of Engineering and Science
Research Inventy : International Journal of Engineering and ScienceResearch Inventy : International Journal of Engineering and Science
Research Inventy : International Journal of Engineering and Scienceresearchinventy
 
An interactive image segmentation using multiple user input’s
An interactive image segmentation using multiple user input’sAn interactive image segmentation using multiple user input’s
An interactive image segmentation using multiple user input’seSAT Publishing House
 
A novel approach for performance parameter estimation of face recognition bas...
A novel approach for performance parameter estimation of face recognition bas...A novel approach for performance parameter estimation of face recognition bas...
A novel approach for performance parameter estimation of face recognition bas...IJMER
 
Face expression recognition using Scaled-conjugate gradient Back-Propagation ...
Face expression recognition using Scaled-conjugate gradient Back-Propagation ...Face expression recognition using Scaled-conjugate gradient Back-Propagation ...
Face expression recognition using Scaled-conjugate gradient Back-Propagation ...IJMER
 
Deep Neural NEtwork
Deep Neural NEtworkDeep Neural NEtwork
Deep Neural NEtworkVai Jayanthi
 
Final Year Project - Enhancing Virtual Learning through Emotional Agents (Doc...
Final Year Project - Enhancing Virtual Learning through Emotional Agents (Doc...Final Year Project - Enhancing Virtual Learning through Emotional Agents (Doc...
Final Year Project - Enhancing Virtual Learning through Emotional Agents (Doc...Sana Nasar
 
Facial expression recognition using pca and gabor with jaffe database 11748
Facial expression recognition using pca and gabor with jaffe database 11748Facial expression recognition using pca and gabor with jaffe database 11748
Facial expression recognition using pca and gabor with jaffe database 11748EditorIJAERD
 
REAL TIME INTELLIGENT EMOTIONAL MUSIC PLAYER USING ANDROID
REAL TIME INTELLIGENT EMOTIONAL MUSIC PLAYER USING ANDROIDREAL TIME INTELLIGENT EMOTIONAL MUSIC PLAYER USING ANDROID
REAL TIME INTELLIGENT EMOTIONAL MUSIC PLAYER USING ANDROIDJournal For Research
 
IRJET- A Review on Face Recognition using Local Binary Pattern Algorithm
IRJET- A Review on Face Recognition using Local Binary Pattern AlgorithmIRJET- A Review on Face Recognition using Local Binary Pattern Algorithm
IRJET- A Review on Face Recognition using Local Binary Pattern AlgorithmIRJET Journal
 
Thermal Imaging Emotion Recognition final report 01
Thermal Imaging Emotion Recognition final report 01Thermal Imaging Emotion Recognition final report 01
Thermal Imaging Emotion Recognition final report 01Ai Zhang
 
Face recognition across non uniform motion blur, illumination, and pose
Face recognition across non uniform motion blur, illumination, and poseFace recognition across non uniform motion blur, illumination, and pose
Face recognition across non uniform motion blur, illumination, and posePvrtechnologies Nellore
 
Predicting Emotions through Facial Expressions
Predicting Emotions through Facial Expressions  Predicting Emotions through Facial Expressions
Predicting Emotions through Facial Expressions twinkle singh
 
Vision based human computer interface using colour detection
Vision based human computer interface using colour detectionVision based human computer interface using colour detection
Vision based human computer interface using colour detectioneSAT Journals
 
Happiness Expression Recognition at Different Age Conditions
Happiness Expression Recognition at Different Age ConditionsHappiness Expression Recognition at Different Age Conditions
Happiness Expression Recognition at Different Age ConditionsEditor IJMTER
 
Emotion based music player
Emotion based music playerEmotion based music player
Emotion based music playerNizam Muhammed
 
A02414491453
A02414491453A02414491453
A02414491453IJMER
 
Facial Expression Recognition
Facial Expression Recognition Facial Expression Recognition
Facial Expression Recognition Rupinder Saini
 
Facial Expression Recognition System using Deep Convolutional Neural Networks.
Facial Expression Recognition  System using Deep Convolutional Neural Networks.Facial Expression Recognition  System using Deep Convolutional Neural Networks.
Facial Expression Recognition System using Deep Convolutional Neural Networks.Sandeep Wakchaure
 

Mais procurados (20)

Research Inventy : International Journal of Engineering and Science
Research Inventy : International Journal of Engineering and ScienceResearch Inventy : International Journal of Engineering and Science
Research Inventy : International Journal of Engineering and Science
 
An interactive image segmentation using multiple user input’s
An interactive image segmentation using multiple user input’sAn interactive image segmentation using multiple user input’s
An interactive image segmentation using multiple user input’s
 
An optimal face recoginition tool
An optimal face recoginition toolAn optimal face recoginition tool
An optimal face recoginition tool
 
A novel approach for performance parameter estimation of face recognition bas...
A novel approach for performance parameter estimation of face recognition bas...A novel approach for performance parameter estimation of face recognition bas...
A novel approach for performance parameter estimation of face recognition bas...
 
Face expression recognition using Scaled-conjugate gradient Back-Propagation ...
Face expression recognition using Scaled-conjugate gradient Back-Propagation ...Face expression recognition using Scaled-conjugate gradient Back-Propagation ...
Face expression recognition using Scaled-conjugate gradient Back-Propagation ...
 
Deep Neural NEtwork
Deep Neural NEtworkDeep Neural NEtwork
Deep Neural NEtwork
 
Final Year Project - Enhancing Virtual Learning through Emotional Agents (Doc...
Final Year Project - Enhancing Virtual Learning through Emotional Agents (Doc...Final Year Project - Enhancing Virtual Learning through Emotional Agents (Doc...
Final Year Project - Enhancing Virtual Learning through Emotional Agents (Doc...
 
Facial expression recognition using pca and gabor with jaffe database 11748
Facial expression recognition using pca and gabor with jaffe database 11748Facial expression recognition using pca and gabor with jaffe database 11748
Facial expression recognition using pca and gabor with jaffe database 11748
 
REAL TIME INTELLIGENT EMOTIONAL MUSIC PLAYER USING ANDROID
REAL TIME INTELLIGENT EMOTIONAL MUSIC PLAYER USING ANDROIDREAL TIME INTELLIGENT EMOTIONAL MUSIC PLAYER USING ANDROID
REAL TIME INTELLIGENT EMOTIONAL MUSIC PLAYER USING ANDROID
 
1705.07543
1705.075431705.07543
1705.07543
 
IRJET- A Review on Face Recognition using Local Binary Pattern Algorithm
IRJET- A Review on Face Recognition using Local Binary Pattern AlgorithmIRJET- A Review on Face Recognition using Local Binary Pattern Algorithm
IRJET- A Review on Face Recognition using Local Binary Pattern Algorithm
 
Thermal Imaging Emotion Recognition final report 01
Thermal Imaging Emotion Recognition final report 01Thermal Imaging Emotion Recognition final report 01
Thermal Imaging Emotion Recognition final report 01
 
Face recognition across non uniform motion blur, illumination, and pose
Face recognition across non uniform motion blur, illumination, and poseFace recognition across non uniform motion blur, illumination, and pose
Face recognition across non uniform motion blur, illumination, and pose
 
Predicting Emotions through Facial Expressions
Predicting Emotions through Facial Expressions  Predicting Emotions through Facial Expressions
Predicting Emotions through Facial Expressions
 
Vision based human computer interface using colour detection
Vision based human computer interface using colour detectionVision based human computer interface using colour detection
Vision based human computer interface using colour detection
 
Happiness Expression Recognition at Different Age Conditions
Happiness Expression Recognition at Different Age ConditionsHappiness Expression Recognition at Different Age Conditions
Happiness Expression Recognition at Different Age Conditions
 
Emotion based music player
Emotion based music playerEmotion based music player
Emotion based music player
 
A02414491453
A02414491453A02414491453
A02414491453
 
Facial Expression Recognition
Facial Expression Recognition Facial Expression Recognition
Facial Expression Recognition
 
Facial Expression Recognition System using Deep Convolutional Neural Networks.
Facial Expression Recognition  System using Deep Convolutional Neural Networks.Facial Expression Recognition  System using Deep Convolutional Neural Networks.
Facial Expression Recognition System using Deep Convolutional Neural Networks.
 

Destaque

D0372027037
D0372027037D0372027037
D0372027037theijes
 
The International Journal of Engineering and Science (The IJES)
The International Journal of Engineering and Science (The IJES)The International Journal of Engineering and Science (The IJES)
The International Journal of Engineering and Science (The IJES)theijes
 
The International Journal of Engineering and Science (The IJES)
The International Journal of Engineering and Science (The IJES)The International Journal of Engineering and Science (The IJES)
The International Journal of Engineering and Science (The IJES)theijes
 
H0373058071
H0373058071H0373058071
H0373058071theijes
 
B02610509
B02610509B02610509
B02610509theijes
 
The International Journal of Engineering and Science (The IJES)
 The International Journal of Engineering and Science (The IJES) The International Journal of Engineering and Science (The IJES)
The International Journal of Engineering and Science (The IJES)theijes
 
G03702054060
G03702054060G03702054060
G03702054060theijes
 
The International Journal of Engineering and Science (The IJES)
The International Journal of Engineering and Science (The IJES)The International Journal of Engineering and Science (The IJES)
The International Journal of Engineering and Science (The IJES)theijes
 
The International Journal of Engineering and Science (The IJES)
The International Journal of Engineering and Science (The IJES)The International Journal of Engineering and Science (The IJES)
The International Journal of Engineering and Science (The IJES)theijes
 
The International Journal of Engineering and Science (The IJES)
The International Journal of Engineering and Science (The IJES)The International Journal of Engineering and Science (The IJES)
The International Journal of Engineering and Science (The IJES)theijes
 
E0362028035
E0362028035E0362028035
E0362028035theijes
 
I0363068074
I0363068074I0363068074
I0363068074theijes
 
E0373031039
E0373031039E0373031039
E0373031039theijes
 
The International Journal of Engineering and Science (The IJES)
The International Journal of Engineering and Science (The IJES)The International Journal of Engineering and Science (The IJES)
The International Journal of Engineering and Science (The IJES)theijes
 
B03501007026
B03501007026B03501007026
B03501007026theijes
 
The International Journal of Engineering and Science (The IJES)
The International Journal of Engineering and Science (The IJES)The International Journal of Engineering and Science (The IJES)
The International Journal of Engineering and Science (The IJES)theijes
 
The International Journal of Engineering and Science (The IJES)
The International Journal of Engineering and Science (The IJES)The International Journal of Engineering and Science (The IJES)
The International Journal of Engineering and Science (The IJES)theijes
 
The International Journal of Engineering and Science (The IJES)
The International Journal of Engineering and Science (The IJES)The International Journal of Engineering and Science (The IJES)
The International Journal of Engineering and Science (The IJES)theijes
 
The International Journal of Engineering and Science (The IJES)
The International Journal of Engineering and Science (The IJES)The International Journal of Engineering and Science (The IJES)
The International Journal of Engineering and Science (The IJES)theijes
 

Destaque (19)

D0372027037
D0372027037D0372027037
D0372027037
 
The International Journal of Engineering and Science (The IJES)
The International Journal of Engineering and Science (The IJES)The International Journal of Engineering and Science (The IJES)
The International Journal of Engineering and Science (The IJES)
 
The International Journal of Engineering and Science (The IJES)
The International Journal of Engineering and Science (The IJES)The International Journal of Engineering and Science (The IJES)
The International Journal of Engineering and Science (The IJES)
 
H0373058071
H0373058071H0373058071
H0373058071
 
B02610509
B02610509B02610509
B02610509
 
The International Journal of Engineering and Science (The IJES)
 The International Journal of Engineering and Science (The IJES) The International Journal of Engineering and Science (The IJES)
The International Journal of Engineering and Science (The IJES)
 
G03702054060
G03702054060G03702054060
G03702054060
 
The International Journal of Engineering and Science (The IJES)
The International Journal of Engineering and Science (The IJES)The International Journal of Engineering and Science (The IJES)
The International Journal of Engineering and Science (The IJES)
 
The International Journal of Engineering and Science (The IJES)
The International Journal of Engineering and Science (The IJES)The International Journal of Engineering and Science (The IJES)
The International Journal of Engineering and Science (The IJES)
 
The International Journal of Engineering and Science (The IJES)
The International Journal of Engineering and Science (The IJES)The International Journal of Engineering and Science (The IJES)
The International Journal of Engineering and Science (The IJES)
 
E0362028035
E0362028035E0362028035
E0362028035
 
I0363068074
I0363068074I0363068074
I0363068074
 
E0373031039
E0373031039E0373031039
E0373031039
 
The International Journal of Engineering and Science (The IJES)
The International Journal of Engineering and Science (The IJES)The International Journal of Engineering and Science (The IJES)
The International Journal of Engineering and Science (The IJES)
 
B03501007026
B03501007026B03501007026
B03501007026
 
The International Journal of Engineering and Science (The IJES)
The International Journal of Engineering and Science (The IJES)The International Journal of Engineering and Science (The IJES)
The International Journal of Engineering and Science (The IJES)
 
The International Journal of Engineering and Science (The IJES)
The International Journal of Engineering and Science (The IJES)The International Journal of Engineering and Science (The IJES)
The International Journal of Engineering and Science (The IJES)
 
The International Journal of Engineering and Science (The IJES)
The International Journal of Engineering and Science (The IJES)The International Journal of Engineering and Science (The IJES)
The International Journal of Engineering and Science (The IJES)
 
The International Journal of Engineering and Science (The IJES)
The International Journal of Engineering and Science (The IJES)The International Journal of Engineering and Science (The IJES)
The International Journal of Engineering and Science (The IJES)
 

Semelhante a Automated emotion detection from facial images

IRJET - Emotion Recognising System-Crowd Behavior Analysis
IRJET -  	  Emotion Recognising System-Crowd Behavior AnalysisIRJET -  	  Emotion Recognising System-Crowd Behavior Analysis
IRJET - Emotion Recognising System-Crowd Behavior AnalysisIRJET Journal
 
Emotion Recognition using Image Processing
Emotion Recognition using Image ProcessingEmotion Recognition using Image Processing
Emotion Recognition using Image Processingijtsrd
 
International Journal of Engineering Research and Development (IJERD)
International Journal of Engineering Research and Development (IJERD)International Journal of Engineering Research and Development (IJERD)
International Journal of Engineering Research and Development (IJERD)IJERD Editor
 
IRJET - Emotionalizer : Face Emotion Detection System
IRJET - Emotionalizer : Face Emotion Detection SystemIRJET - Emotionalizer : Face Emotion Detection System
IRJET - Emotionalizer : Face Emotion Detection SystemIRJET Journal
 
IRJET- Emotionalizer : Face Emotion Detection System
IRJET- Emotionalizer : Face Emotion Detection SystemIRJET- Emotionalizer : Face Emotion Detection System
IRJET- Emotionalizer : Face Emotion Detection SystemIRJET Journal
 
IRJET- Detection of Static Body Poses using Skin Detection Techniques to Asse...
IRJET- Detection of Static Body Poses using Skin Detection Techniques to Asse...IRJET- Detection of Static Body Poses using Skin Detection Techniques to Asse...
IRJET- Detection of Static Body Poses using Skin Detection Techniques to Asse...IRJET Journal
 
Real Time Facial Expression Recognition and Imitation
Real Time Facial Expression Recognition and ImitationReal Time Facial Expression Recognition and Imitation
Real Time Facial Expression Recognition and Imitationijtsrd
 
AI Therapist – Emotion Detection using Facial Detection and Recognition and S...
AI Therapist – Emotion Detection using Facial Detection and Recognition and S...AI Therapist – Emotion Detection using Facial Detection and Recognition and S...
AI Therapist – Emotion Detection using Facial Detection and Recognition and S...ijtsrd
 
IRJET- Intelligent Emotion Detection System using Facial Images
IRJET- Intelligent Emotion Detection System using Facial ImagesIRJET- Intelligent Emotion Detection System using Facial Images
IRJET- Intelligent Emotion Detection System using Facial ImagesIRJET Journal
 
International Journal of Engineering Research and Development
International Journal of Engineering Research and DevelopmentInternational Journal of Engineering Research and Development
International Journal of Engineering Research and DevelopmentIJERD Editor
 
IRJET- An Effective System to Detect Face Drowsiness Status using Local F...
IRJET-  	  An Effective System to Detect Face Drowsiness Status using Local F...IRJET-  	  An Effective System to Detect Face Drowsiness Status using Local F...
IRJET- An Effective System to Detect Face Drowsiness Status using Local F...IRJET Journal
 
Facial Emoji Recognition
Facial Emoji RecognitionFacial Emoji Recognition
Facial Emoji Recognitionijtsrd
 
Emotion Detector
Emotion DetectorEmotion Detector
Emotion Detectorijtsrd
 
IRJET- ATM Security using Machine Learning
IRJET- ATM Security using Machine LearningIRJET- ATM Security using Machine Learning
IRJET- ATM Security using Machine LearningIRJET Journal
 
Face detection for video summary using enhancement based fusion strategy
Face detection for video summary using enhancement based fusion strategyFace detection for video summary using enhancement based fusion strategy
Face detection for video summary using enhancement based fusion strategyeSAT Publishing House
 
HUMAN FACE RECOGNITION USING IMAGE PROCESSING PCA AND NEURAL NETWORK
HUMAN FACE RECOGNITION USING IMAGE PROCESSING PCA AND NEURAL NETWORKHUMAN FACE RECOGNITION USING IMAGE PROCESSING PCA AND NEURAL NETWORK
HUMAN FACE RECOGNITION USING IMAGE PROCESSING PCA AND NEURAL NETWORKijiert bestjournal
 
A New Skin Color Based Face Detection Algorithm by Combining Three Color Mode...
A New Skin Color Based Face Detection Algorithm by Combining Three Color Mode...A New Skin Color Based Face Detection Algorithm by Combining Three Color Mode...
A New Skin Color Based Face Detection Algorithm by Combining Three Color Mode...iosrjce
 
EMOTION RECOGNITION SYSTEMS: A REVIEW
EMOTION RECOGNITION SYSTEMS: A REVIEWEMOTION RECOGNITION SYSTEMS: A REVIEW
EMOTION RECOGNITION SYSTEMS: A REVIEWIRJET Journal
 
A Study of Method in Facial Emotional Recognitation
A Study of Method in Facial Emotional RecognitationA Study of Method in Facial Emotional Recognitation
A Study of Method in Facial Emotional Recognitationrahulmonikasharma
 

Semelhante a Automated emotion detection from facial images (20)

IRJET - Emotion Recognising System-Crowd Behavior Analysis
IRJET -  	  Emotion Recognising System-Crowd Behavior AnalysisIRJET -  	  Emotion Recognising System-Crowd Behavior Analysis
IRJET - Emotion Recognising System-Crowd Behavior Analysis
 
Emotion Recognition using Image Processing
Emotion Recognition using Image ProcessingEmotion Recognition using Image Processing
Emotion Recognition using Image Processing
 
International Journal of Engineering Research and Development (IJERD)
International Journal of Engineering Research and Development (IJERD)International Journal of Engineering Research and Development (IJERD)
International Journal of Engineering Research and Development (IJERD)
 
IRJET - Emotionalizer : Face Emotion Detection System
IRJET - Emotionalizer : Face Emotion Detection SystemIRJET - Emotionalizer : Face Emotion Detection System
IRJET - Emotionalizer : Face Emotion Detection System
 
IRJET- Emotionalizer : Face Emotion Detection System
IRJET- Emotionalizer : Face Emotion Detection SystemIRJET- Emotionalizer : Face Emotion Detection System
IRJET- Emotionalizer : Face Emotion Detection System
 
IRJET- Detection of Static Body Poses using Skin Detection Techniques to Asse...
IRJET- Detection of Static Body Poses using Skin Detection Techniques to Asse...IRJET- Detection of Static Body Poses using Skin Detection Techniques to Asse...
IRJET- Detection of Static Body Poses using Skin Detection Techniques to Asse...
 
Real Time Facial Expression Recognition and Imitation
Real Time Facial Expression Recognition and ImitationReal Time Facial Expression Recognition and Imitation
Real Time Facial Expression Recognition and Imitation
 
AI Therapist – Emotion Detection using Facial Detection and Recognition and S...
AI Therapist – Emotion Detection using Facial Detection and Recognition and S...AI Therapist – Emotion Detection using Facial Detection and Recognition and S...
AI Therapist – Emotion Detection using Facial Detection and Recognition and S...
 
IRJET- Intelligent Emotion Detection System using Facial Images
IRJET- Intelligent Emotion Detection System using Facial ImagesIRJET- Intelligent Emotion Detection System using Facial Images
IRJET- Intelligent Emotion Detection System using Facial Images
 
International Journal of Engineering Research and Development
International Journal of Engineering Research and DevelopmentInternational Journal of Engineering Research and Development
International Journal of Engineering Research and Development
 
IRJET- An Effective System to Detect Face Drowsiness Status using Local F...
IRJET-  	  An Effective System to Detect Face Drowsiness Status using Local F...IRJET-  	  An Effective System to Detect Face Drowsiness Status using Local F...
IRJET- An Effective System to Detect Face Drowsiness Status using Local F...
 
Facial Emoji Recognition
Facial Emoji RecognitionFacial Emoji Recognition
Facial Emoji Recognition
 
Emotion Detector
Emotion DetectorEmotion Detector
Emotion Detector
 
IRJET- ATM Security using Machine Learning
IRJET- ATM Security using Machine LearningIRJET- ATM Security using Machine Learning
IRJET- ATM Security using Machine Learning
 
Face detection for video summary using enhancement based fusion strategy
Face detection for video summary using enhancement based fusion strategyFace detection for video summary using enhancement based fusion strategy
Face detection for video summary using enhancement based fusion strategy
 
HUMAN FACE RECOGNITION USING IMAGE PROCESSING PCA AND NEURAL NETWORK
HUMAN FACE RECOGNITION USING IMAGE PROCESSING PCA AND NEURAL NETWORKHUMAN FACE RECOGNITION USING IMAGE PROCESSING PCA AND NEURAL NETWORK
HUMAN FACE RECOGNITION USING IMAGE PROCESSING PCA AND NEURAL NETWORK
 
A New Skin Color Based Face Detection Algorithm by Combining Three Color Mode...
A New Skin Color Based Face Detection Algorithm by Combining Three Color Mode...A New Skin Color Based Face Detection Algorithm by Combining Three Color Mode...
A New Skin Color Based Face Detection Algorithm by Combining Three Color Mode...
 
B017310612
B017310612B017310612
B017310612
 
EMOTION RECOGNITION SYSTEMS: A REVIEW
EMOTION RECOGNITION SYSTEMS: A REVIEWEMOTION RECOGNITION SYSTEMS: A REVIEW
EMOTION RECOGNITION SYSTEMS: A REVIEW
 
A Study of Method in Facial Emotional Recognitation
A Study of Method in Facial Emotional RecognitationA Study of Method in Facial Emotional Recognitation
A Study of Method in Facial Emotional Recognitation
 

Último

Finology Group – Insurtech Innovation Award 2024
Finology Group – Insurtech Innovation Award 2024Finology Group – Insurtech Innovation Award 2024
Finology Group – Insurtech Innovation Award 2024The Digital Insurer
 
Workshop - Best of Both Worlds_ Combine KG and Vector search for enhanced R...
Workshop - Best of Both Worlds_ Combine  KG and Vector search for  enhanced R...Workshop - Best of Both Worlds_ Combine  KG and Vector search for  enhanced R...
Workshop - Best of Both Worlds_ Combine KG and Vector search for enhanced R...Neo4j
 
Tata AIG General Insurance Company - Insurer Innovation Award 2024
Tata AIG General Insurance Company - Insurer Innovation Award 2024Tata AIG General Insurance Company - Insurer Innovation Award 2024
Tata AIG General Insurance Company - Insurer Innovation Award 2024The Digital Insurer
 
Real Time Object Detection Using Open CV
Real Time Object Detection Using Open CVReal Time Object Detection Using Open CV
Real Time Object Detection Using Open CVKhem
 
Boost PC performance: How more available memory can improve productivity
Boost PC performance: How more available memory can improve productivityBoost PC performance: How more available memory can improve productivity
Boost PC performance: How more available memory can improve productivityPrincipled Technologies
 
From Event to Action: Accelerate Your Decision Making with Real-Time Automation
From Event to Action: Accelerate Your Decision Making with Real-Time AutomationFrom Event to Action: Accelerate Your Decision Making with Real-Time Automation
From Event to Action: Accelerate Your Decision Making with Real-Time AutomationSafe Software
 
How to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerHow to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerThousandEyes
 
The 7 Things I Know About Cyber Security After 25 Years | April 2024
The 7 Things I Know About Cyber Security After 25 Years | April 2024The 7 Things I Know About Cyber Security After 25 Years | April 2024
The 7 Things I Know About Cyber Security After 25 Years | April 2024Rafal Los
 
Slack Application Development 101 Slides
Slack Application Development 101 SlidesSlack Application Development 101 Slides
Slack Application Development 101 Slidespraypatel2
 
Scaling API-first – The story of a global engineering organization
Scaling API-first – The story of a global engineering organizationScaling API-first – The story of a global engineering organization
Scaling API-first – The story of a global engineering organizationRadu Cotescu
 
Breaking the Kubernetes Kill Chain: Host Path Mount
Breaking the Kubernetes Kill Chain: Host Path MountBreaking the Kubernetes Kill Chain: Host Path Mount
Breaking the Kubernetes Kill Chain: Host Path MountPuma Security, LLC
 
Data Cloud, More than a CDP by Matt Robison
Data Cloud, More than a CDP by Matt RobisonData Cloud, More than a CDP by Matt Robison
Data Cloud, More than a CDP by Matt RobisonAnna Loughnan Colquhoun
 
Artificial Intelligence: Facts and Myths
Artificial Intelligence: Facts and MythsArtificial Intelligence: Facts and Myths
Artificial Intelligence: Facts and MythsJoaquim Jorge
 
Advantages of Hiring UIUX Design Service Providers for Your Business
Advantages of Hiring UIUX Design Service Providers for Your BusinessAdvantages of Hiring UIUX Design Service Providers for Your Business
Advantages of Hiring UIUX Design Service Providers for Your BusinessPixlogix Infotech
 
[2024]Digital Global Overview Report 2024 Meltwater.pdf
[2024]Digital Global Overview Report 2024 Meltwater.pdf[2024]Digital Global Overview Report 2024 Meltwater.pdf
[2024]Digital Global Overview Report 2024 Meltwater.pdfhans926745
 
A Domino Admins Adventures (Engage 2024)
A Domino Admins Adventures (Engage 2024)A Domino Admins Adventures (Engage 2024)
A Domino Admins Adventures (Engage 2024)Gabriella Davis
 
EIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptx
EIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptxEIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptx
EIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptxEarley Information Science
 
IAC 2024 - IA Fast Track to Search Focused AI Solutions
IAC 2024 - IA Fast Track to Search Focused AI SolutionsIAC 2024 - IA Fast Track to Search Focused AI Solutions
IAC 2024 - IA Fast Track to Search Focused AI SolutionsEnterprise Knowledge
 
Raspberry Pi 5: Challenges and Solutions in Bringing up an OpenGL/Vulkan Driv...
Raspberry Pi 5: Challenges and Solutions in Bringing up an OpenGL/Vulkan Driv...Raspberry Pi 5: Challenges and Solutions in Bringing up an OpenGL/Vulkan Driv...
Raspberry Pi 5: Challenges and Solutions in Bringing up an OpenGL/Vulkan Driv...Igalia
 
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...apidays
 

Último (20)

Finology Group – Insurtech Innovation Award 2024
Finology Group – Insurtech Innovation Award 2024Finology Group – Insurtech Innovation Award 2024
Finology Group – Insurtech Innovation Award 2024
 
Workshop - Best of Both Worlds_ Combine KG and Vector search for enhanced R...
Workshop - Best of Both Worlds_ Combine  KG and Vector search for  enhanced R...Workshop - Best of Both Worlds_ Combine  KG and Vector search for  enhanced R...
Workshop - Best of Both Worlds_ Combine KG and Vector search for enhanced R...
 
Tata AIG General Insurance Company - Insurer Innovation Award 2024
Tata AIG General Insurance Company - Insurer Innovation Award 2024Tata AIG General Insurance Company - Insurer Innovation Award 2024
Tata AIG General Insurance Company - Insurer Innovation Award 2024
 
Real Time Object Detection Using Open CV
Real Time Object Detection Using Open CVReal Time Object Detection Using Open CV
Real Time Object Detection Using Open CV
 
Boost PC performance: How more available memory can improve productivity
Boost PC performance: How more available memory can improve productivityBoost PC performance: How more available memory can improve productivity
Boost PC performance: How more available memory can improve productivity
 
From Event to Action: Accelerate Your Decision Making with Real-Time Automation
From Event to Action: Accelerate Your Decision Making with Real-Time AutomationFrom Event to Action: Accelerate Your Decision Making with Real-Time Automation
From Event to Action: Accelerate Your Decision Making with Real-Time Automation
 
How to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerHow to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected Worker
 
The 7 Things I Know About Cyber Security After 25 Years | April 2024
The 7 Things I Know About Cyber Security After 25 Years | April 2024The 7 Things I Know About Cyber Security After 25 Years | April 2024
The 7 Things I Know About Cyber Security After 25 Years | April 2024
 
Slack Application Development 101 Slides
Slack Application Development 101 SlidesSlack Application Development 101 Slides
Slack Application Development 101 Slides
 
Scaling API-first – The story of a global engineering organization
Scaling API-first – The story of a global engineering organizationScaling API-first – The story of a global engineering organization
Scaling API-first – The story of a global engineering organization
 
Breaking the Kubernetes Kill Chain: Host Path Mount
Breaking the Kubernetes Kill Chain: Host Path MountBreaking the Kubernetes Kill Chain: Host Path Mount
Breaking the Kubernetes Kill Chain: Host Path Mount
 
Data Cloud, More than a CDP by Matt Robison
Data Cloud, More than a CDP by Matt RobisonData Cloud, More than a CDP by Matt Robison
Data Cloud, More than a CDP by Matt Robison
 
Artificial Intelligence: Facts and Myths
Artificial Intelligence: Facts and MythsArtificial Intelligence: Facts and Myths
Artificial Intelligence: Facts and Myths
 
Advantages of Hiring UIUX Design Service Providers for Your Business
Advantages of Hiring UIUX Design Service Providers for Your BusinessAdvantages of Hiring UIUX Design Service Providers for Your Business
Advantages of Hiring UIUX Design Service Providers for Your Business
 
[2024]Digital Global Overview Report 2024 Meltwater.pdf
[2024]Digital Global Overview Report 2024 Meltwater.pdf[2024]Digital Global Overview Report 2024 Meltwater.pdf
[2024]Digital Global Overview Report 2024 Meltwater.pdf
 
A Domino Admins Adventures (Engage 2024)
A Domino Admins Adventures (Engage 2024)A Domino Admins Adventures (Engage 2024)
A Domino Admins Adventures (Engage 2024)
 
EIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptx
EIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptxEIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptx
EIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptx
 
IAC 2024 - IA Fast Track to Search Focused AI Solutions
IAC 2024 - IA Fast Track to Search Focused AI SolutionsIAC 2024 - IA Fast Track to Search Focused AI Solutions
IAC 2024 - IA Fast Track to Search Focused AI Solutions
 
Raspberry Pi 5: Challenges and Solutions in Bringing up an OpenGL/Vulkan Driv...
Raspberry Pi 5: Challenges and Solutions in Bringing up an OpenGL/Vulkan Driv...Raspberry Pi 5: Challenges and Solutions in Bringing up an OpenGL/Vulkan Driv...
Raspberry Pi 5: Challenges and Solutions in Bringing up an OpenGL/Vulkan Driv...
 
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...
 

Automated emotion detection from facial images

  • 1. The International Journal Of Engineering And Science (IJES) || Volume || 3 || Issue || 5 || Pages || 18-22 || 2014 || ISSN (e): 2319 – 1813 ISSN (p): 2319 – 1805 www.theijes.com The IJES Page 18  Process Automation of Human Emotion Detection 1, Namrata Mahajan, 2, Harshad Mahajan 1, Department of Electronics and Communication, VIT Nagpur 2, Department of Electronics and Communication, GCOE Amravati --------------------------------------------------------------ABSTRACT------------------------------------------------------ Human can see emotion as associated with mood, temperament, personality and disposition. To detect emotions is easy for humans but it’s quite difficult for computer as world is three dimensional but Computer has only two dimensions. Computer seeks to emulate the human emotions by digital image analysis. Humans can use vision to identify objects quickly and accurately. Human can detect emotion using voice based on different parameters like tone, pitch, pace and volume; but in case of digital images detecting emotion just by analysing images is a novel way. This algorithm has two major parts. First, Template database generation and another is emotion detection. In this, we first extracts face from an image using some basic image processing operations and color models in it. Here we define thresholds to separate our region of interest. Then we perform lip detection on cropped face and this extracted lips are stored in our database with emotion name in database generation phase while in emotion detection phase this extracted lips are compared with series of stored template in database and on the basis of best correlated template emotion is recognized. This method of detecting emotions is simple and fast as compared to previous methods i.e. using brain activity or by speech. Size of database will affect the effectiveness of the proposed algorithm. Keywords - Emotion Detection, Skin based Segmentation, Automation and Lip Segmentation -------------------------------------------------------------------------------------------------------------------------------------- Date of Submission: 02 May 2014 Date of Publication: 15 May 2014 -------------------------------------------------------------------------------------------------------------------------------------- I. INTRODUCTION Human has ability to see emotions by temperament, disposition, personality and mood. Computer seeks to emulate the human emotions by digital image analysis. The problem with the computer vision is due to the fact that world is three dimensional but Computer has only two dimensions. In our day to day life we interact with each other directly (for e.g. face to face) or indirectly (for e.g. phone calls). In some profession interaction like call centers interaction with people is important. With great advancement in technology in terms of different techniques of people interacting with each other it is quite necessary that one should be aware of current emotions of the person he/she is interacting. It is widely accepted from psychological theory that human emotions can be classified into six archetypal emotions: surprise, fear, disgust, anger, happiness, and sadness. Facial motion and the tone of the speech play a major role in expressing these emotions. The muscles of the face can be changed and the tone and the energy in the production of the speech can be intentionally modified to communicate different feelings. Human beings can recognize these signals even if they are subtly displayed, by simultaneously processing information acquired by ears and eyes. Based on psychological studies, which show that visual information modifies the perception of speech [1], it is possible to assume that human emotion perception follows a similar trend. This paper analyses the use of digital image to recognize four different human emotions: sadness, happiness, anger and neutral state, using a database recorded .the more challenging task of capturing salient visual information directly from conventional videos is a topic for future work but is hoped to be informed by studies such as in this report. The primary purpose of this paper is to define a simple and fast automated emotion recognition system.
  • 2. Process Automation of Human Emotion Detection www.theijes.com The IJES Page 19 II. PROCESS AUTOMATION In order to implement an automated emotion detection system, a series of logical steps have to be developed. The flow of steps in this automated system is defined below. Our project is mainly divided in two parts Database Generation (Fig.1) and Emotion Recognition (Fig.2). In Emotion recognition we are going to match resulting output of input image with our templates which already exist in our database. A. Face Detection We faced many problems related to the face detection. To locate the face image should consists of single face. Hence in this paper our assumption is that an input image is having single face. If this assumption is not satisfied it will take either of any one image from input not any specific one. Although we provide a single face image for face detection but face detection means identifying a region of image which contains face no matter what is the background of image, its lighting conditions or orientation. This task is bit critical as different faces have different colours, shapes and texture. At this time, many approaches involve automatic face detection followed by lip detection .But most of approaches define the scope of upright view of faces such approaches failed to detect faces rotated in plane or out of image plane. Also sometimes it failed while detecting faces from image when only part of face is there in image. Hence in this project we are using thresholding to separate skin region from image in order to detect face. 1. Skin Color Classification Skin color and textures are important cues that people use consciously or unconsciously to infer variety of culture-related aspects about each other. Skin color and texture can be an indication of race, health, age, wealth, beauty, etc. [3]. However, such interpretations vary across cultures and across the history. In images, skin color is an indication of the existence of humans in such media. The color of human skin is created by a combination of blood (red) and melanin (yellow, brown). Skin colors lie between these two extreme hues and are somewhat saturate Facial skin takes on is clearly a subspace of the total color space. The human skin is a fraction of the actual color cube, about 0.25 % of the total colors. Except for extremely hairy subjects, which are rare, skin has only low-amplitude texture. Three color spaces i.e. RGB, HIS and YCrCb are defined in literature and they are useful to achieve our goal. 2. Skin Based Segmentation To detect Human faces, their skin color is unique and efficient feature. Although skin color may vary person to person, major difference lies largely between their intensity rather than their chrominance. Most of the face detection algorithms are on the basis of skin based segmentation. Many of these algorithms uses the algorithms use a range of color values to define skin color.
  • 3. Process Automation of Human Emotion Detection www.theijes.com The IJES Page 20 3. Segmentation Rules Segmentation procedures partition an image into its constituent parts or objects. A rugged segmentation procedure brings the process a long way toward successful solution of imaging problems that require objects to be identified individually. In general, the more accurate the segmentation, the more likely recognition is to succeed. Segmentation rules determine the formation of regions. These rules can be defined on the basis of color and edge properties of region. Using skin based segmentation we can reduce the efforts of first determining faces within image in which faces are out of frame or tilted. 4. Segmentation Rules for different Colour Models After some experiments we found following rules for different color models which give one step closure to our goal. RGB Model: 0.836G – 14 < B < 0.836G + 44 => Skin OR 0.79G -67 < B < 0.78G + 42 => Skin HIS/HSV Model: 19 < H < 240 => Not skin YCrCb Model: 140 < Cr < 165 => Skin OR 140 < Cr < 195 => Skin Fig 3.Orignal Image Fig.4 Image after Skin Thresholding 5. Erosion and Dilation After skin based segmentation, for fast processing we convert our output image into binary image. The noises in this black and white image are removed using suitable window sizes. The pixels in this image which were falsely detected as skin color pixels are removed by erosion and dilation by moving window over image matrix. If it is found that the number of white pixels is more than fixed threshold then whole window is made black. The face detection is performed on skin detected areas. Hence to locate face we always consider the occurrence of white region in output image after erosion and dilation. This process has assumption that image on which we are going to perform process should have single face with minimum skin exposure. Skin colors in this project are using RGB, HIS/HSV and YCrCb models. Values are unique for this project obtained using fresh toned digital photos. For larger skin tone range HIS model gives better results but these models made it difficult to constraint color for flesh-like tones. RGB models are preferred in such situations. Fig.5 Image after Dilation Fig. 6 Image after Erosion
  • 4. Process Automation of Human Emotion Detection www.theijes.com The IJES Page 21 B. Cropping Facial Region After detection of Face we are going to extract lip region .We are aware about the thing that lip region contents maximum of red color content. If red color content is present in the image outside the facial region, it will interfere in our lip extraction process it may also cause increase in time and complexity. Thus we decide to crop the region of our interest only. So focusing on this limited area, results will be more accurate. Also Computational time is directly proportional to the size of image. If processing image size is reduced due to cropping region of interest, computational time will also reduce. Before cropping we used to apply erosion and dilation operations on detected face image. These functions help us to remove noise which is quite helpful for region cropping. Major work in cropping is to find four boundary points. We found first top, bottom, right and left four points. On the basis of these four points height and width of detected region can be calculated. After calculating height and width we create rectangle and then crop required area of facial region. C. Lip Detection After face detection, we have our region of interest; on this image we applied our most important algorithm of lip detection. Human emotions reflect changes in lip region and eyes. But maximum changes around lips are visible rather than eyes. Hence a detected lip gives us accurate results. After face detection and getting our region of interest, we now move towards the most important part of the project i.e. lip detection. As already discussed in the 1st chapter, the variations seen during different emotions are around the lip region as well as the eyes region. But the maximum changes that are visible will be around the lip region, thus giving us more accurate results. Same as we used RGB thresholds in skin based se3gmentations to detect face. Here we used thresholds on red color as lip region contains maximum of red colors. In field of computer vision, automation of the process of locating and tracking person‟s lip is somewhat difficult task. There several methods proposed for lip detection in last 10-12 years. But we found important and effective way of detecting lip based on the segmentation directly from the color space. In this algorithm we can enlarge difference between skin and lip using color filters or color transformations. The processing time in this algorithm is main advantage .Lip segmentation is very sensitive to changes in colors. Hence it may create problem if for some faces lip and skin contrast is lower. Fig. 7 Image after Lip Detection D. Lip Segmentation After completion of lip detection, now our next step is lip segmentation. After this segmentation we will use result of lip segmentation to generate template if we are creating database otherwise it will be used to detect emotion if we are in emotion detection phase of project. Thus this cropped lip region makes us closer to our aim. Again the process of lip segmentation is same as face detection .In which we need to define first four boundary points i.e. top, bottom, left and right points of detected lip. From this point we will find width and height of lip region which is further used to create rectangle and extract lip region. It may possible in case of noise occurs on the image, it will create confusion between lips and any other area on face. To avoid this confusion we used a fact that width of height is 3 times than lips height. But for some worst case we modified this condition as lip width should be greater than the twice of its width, if this condition is true then we create rectangle window and crop lip region. E. Template Database Generation Our project is mainly divided in two parts. 1. Database Generation and 2. Emotion Recognition. In Emotion recognition we are going to match resulting output of input image with our templates which already exist in our database. But for this we need to create database. In order to generate database, templates have to be generated. Roipoly function is used to create template. It is out of box function in MATLAB. Roiply function specifies the polygonal region of interest (ROI) within an image. Its output is in the form of binary image which is used as mask for filtering.
  • 5. Process Automation of Human Emotion Detection www.theijes.com The IJES Page 22 BW = roipoly creates an interactive polygon tool, associated with the image displayed in the current figure, called the target image. With the polygon tool active, the pointer changes to cross hairs,„+ „ when you move the pointer over the image in the figure. Using the mouse, you specify the region by selecting vertices of the polygon or any area of interest. You can move or resize the polygon or figure using the mouse. When you are finished positioning and sizing the polygon/figure, create the mask by double-clicking, or by right-clicking inside the region and selecting Create mask from the context menu. Roipoly returns the mask as a binary image, BW, the same size as I. In the mask image, roipoly sets pixels inside the region to 1 and pixels outside the region to 0. This binary image BW so returned is in logical format and as such cannot be used for implementation. It needs to be converted to a format that can be processed on i.e. double or uint8. After this the masks are cropped in the similar manner used for face cropping and lip cropping i.e. finding four points and creating rectangle and finally cropping.[5] F. Emotion Recognition In Emotion recognition we are going to match resulting output of input image with our templates which already exist in our database. This comparison is made by using cross-correlation. This option calculates the cross correlation function for two images. Each of the images is divided into rectangular blocks. Each block in the first image is correlated with its corresponding block in the second image to produce the cross correlation as a function of position. The cross correlation may be used to determine the degree of similarity between two similar images, or, with the addition of a linear offset to one of the images, the spatial shift or spatial correlation between the images. G. Emotion Detection After finding cross-correlation of the detected and cropped lip region with templates, we have to find the index number on which we get maximum value offered by cross-correlation of two images. Emotion is detected by checking the corresponding emotion across the index number with maximum value of cross- correlation III. FINAL REMARKS We found that this algorithm is faster and easy to implement than existing algorithm based on speech recognition and by ECG in which physical contact with human brain is needed which is quite expensive than this. We worked on five emotions i.e. .happiness, grief, anger, surprise and neutral. These emotions give us successful results on majority of images. Generally no false emotion were found or wrongly interpreted. REFERENCES [1] “Human Emotion Recognition using Frequency & Statistical Measures of EEG Signal” , Monira Islam, Tazrin Ahmed, Sheikh Shanawaz Mostafa*, Md. Salah Uddin Yusuf, and Mohiuddin Ahmad [2] “Human Emotion Recognition using Frequency & Statistical Measures of EEG Signal”, Monira Islam, Tazrin Ahmed, Sheikh Shanawaz Mostafa*, Md. Salah Uddin Yusuf, and Mohiuddin Ahmad [3] Bernhard Fink, K.G., Matts, P.J.: Visible skin color distribution plays a role in the perception of age, attractiveness, and health in female faces. Evolution and Human Behavior 27(6) (2006) 433–442 [4] Jones, M.J., Rehg, and J.M.: Statistical color models with application to skin detection. International Journal of Computer Vision (IJCV) 46(1) (2002) 81–96 [5] http://www.mathworks.in/help/images/ref/roipoly.html [6] A. Yuille, “Feature Extraction from Faces Using Deformable Templates,” Int. Journal of Computer Vision,(1992), 8(2):99-111. [7] Bracewell R.N.: “The Hartley transform”, Oxford University Press, 1986. [8] Gocke R., Millar J.B., Zelinsky A., Robert-Ribes J, “Automatic extraction of lip feature points”. Proc. Austral. Conf. Robotics and Automation, Melbourne, Australia, August 2000, pp. 31–36.