SlideShare uma empresa Scribd logo
1 de 1
Baixar para ler offline
DETECTING THE HUMAN BIOMETRIC SENSOR INTERACTION WITH MICROSOFT KINECT
This project was designed to create a program that can read data that has been captured from a Kinect camera into an excel worksheet.
Then from the excel worksheet, display a linear graph that represents the distance from the camera the user’s head was located. This
can be used to closely look over if an individual was the correct distance from the iris camera. This is important when trying to determine
the cause of a unsuccessful capture of the iris, because you can determine if it was a human error rather than a device error.
Robert Moore, Hongming Cui, Michael Brockly, Stephen Elliott
Overview
The image to the left is of
the iris camera being
used during the test. A
person using this would
be standing about the
same location at which
this picture was taken.
The Kinect camera is
located behind the
individual during this
capture process
This box represent the optimal area to
be standing when using the iris
camera. The red line represents the
farthest you should be when using the
camera and the green line represents
the closest you should be.
The image above shows the Kinect
camera circled in black as well as the red
square in which a person would stand
during the test.
This image shows an area being selected for
zoom. The selected area is represented as the
gray box near the center of the image. This is
selected by holding left click on the graph then
dragging the mouse either right or left.
The excel data above shows a subjects positioning 3 dimensionally
that was captured from the Kinect camera. For this project we used
column A (Timestamp) and P (Head Z) to create the image to the right
which is what the program draws using the Head Z position for the Y-
axis and timestamp for the X-axis. The output is shown here.
The image shows each interaction that occurred
in the total transaction.
This last image represents what is displayed
after selecting an area and releasing left click.
Being able to zoom in on these sections will
allow for a deeper understanding of HBSI
interactions.
The three images below give an example of the zooming process from left to right.

Mais conteúdo relacionado

Mais de International Center for Biometric Research

Advances in testing and evaluation using Human-Biometric sensor interaction m...
Advances in testing and evaluation using Human-Biometric sensor interaction m...Advances in testing and evaluation using Human-Biometric sensor interaction m...
Advances in testing and evaluation using Human-Biometric sensor interaction m...International Center for Biometric Research
 
(2010) Fingerprint recognition performance evaluation for mobile ID applications
(2010) Fingerprint recognition performance evaluation for mobile ID applications(2010) Fingerprint recognition performance evaluation for mobile ID applications
(2010) Fingerprint recognition performance evaluation for mobile ID applicationsInternational Center for Biometric Research
 

Mais de International Center for Biometric Research (20)

Entropy of Fingerprints
Entropy of FingerprintsEntropy of Fingerprints
Entropy of Fingerprints
 
Biometric and usability
Biometric and usabilityBiometric and usability
Biometric and usability
 
Examining Intra-Visit Iris Stability - Visit 4
Examining Intra-Visit Iris Stability - Visit 4Examining Intra-Visit Iris Stability - Visit 4
Examining Intra-Visit Iris Stability - Visit 4
 
Examining Intra-Visit Iris Stability - Visit 6
Examining Intra-Visit Iris Stability - Visit 6Examining Intra-Visit Iris Stability - Visit 6
Examining Intra-Visit Iris Stability - Visit 6
 
Examining Intra-Visit Iris Stability - Visit 2
Examining Intra-Visit Iris Stability - Visit 2Examining Intra-Visit Iris Stability - Visit 2
Examining Intra-Visit Iris Stability - Visit 2
 
Examining Intra-Visit Iris Stability - Visit 1
Examining Intra-Visit Iris Stability - Visit 1Examining Intra-Visit Iris Stability - Visit 1
Examining Intra-Visit Iris Stability - Visit 1
 
Examining Intra-Visit Iris Stability - Visit 3
Examining Intra-Visit Iris Stability - Visit 3Examining Intra-Visit Iris Stability - Visit 3
Examining Intra-Visit Iris Stability - Visit 3
 
Best Practices in Reporting Time Duration in Biometrics
Best Practices in Reporting Time Duration in BiometricsBest Practices in Reporting Time Duration in Biometrics
Best Practices in Reporting Time Duration in Biometrics
 
Examining Intra-Visit Iris Stability - Visit 5
Examining Intra-Visit Iris Stability - Visit 5Examining Intra-Visit Iris Stability - Visit 5
Examining Intra-Visit Iris Stability - Visit 5
 
Standards and Academia
Standards and AcademiaStandards and Academia
Standards and Academia
 
Interoperability and the Stability Score Index
Interoperability and the Stability Score IndexInteroperability and the Stability Score Index
Interoperability and the Stability Score Index
 
Advances in testing and evaluation using Human-Biometric sensor interaction m...
Advances in testing and evaluation using Human-Biometric sensor interaction m...Advances in testing and evaluation using Human-Biometric sensor interaction m...
Advances in testing and evaluation using Human-Biometric sensor interaction m...
 
Cerias talk on testing and evaluation
Cerias talk on testing and evaluationCerias talk on testing and evaluation
Cerias talk on testing and evaluation
 
IT 54500 overview
IT 54500 overviewIT 54500 overview
IT 54500 overview
 
Ben thesis slideshow
Ben thesis slideshowBen thesis slideshow
Ben thesis slideshow
 
(2010) Fingerprint recognition performance evaluation for mobile ID applications
(2010) Fingerprint recognition performance evaluation for mobile ID applications(2010) Fingerprint recognition performance evaluation for mobile ID applications
(2010) Fingerprint recognition performance evaluation for mobile ID applications
 
ICBR Databases
ICBR DatabasesICBR Databases
ICBR Databases
 
Understanding Fingerprint Skin Characteristics and Image Quality
Understanding Fingerprint Skin Characteristics and Image QualityUnderstanding Fingerprint Skin Characteristics and Image Quality
Understanding Fingerprint Skin Characteristics and Image Quality
 
HBSI automation using the kinect
HBSI automation using the kinectHBSI automation using the kinect
HBSI automation using the kinect
 
(2013) Automatic Detection of Biometrics Transaction Times
(2013) Automatic Detection of Biometrics Transaction Times(2013) Automatic Detection of Biometrics Transaction Times
(2013) Automatic Detection of Biometrics Transaction Times
 

(Spring 2013) Detecting the Human Biometric Sensor Interaction with Microsoft Kinect

  • 1. DETECTING THE HUMAN BIOMETRIC SENSOR INTERACTION WITH MICROSOFT KINECT This project was designed to create a program that can read data that has been captured from a Kinect camera into an excel worksheet. Then from the excel worksheet, display a linear graph that represents the distance from the camera the user’s head was located. This can be used to closely look over if an individual was the correct distance from the iris camera. This is important when trying to determine the cause of a unsuccessful capture of the iris, because you can determine if it was a human error rather than a device error. Robert Moore, Hongming Cui, Michael Brockly, Stephen Elliott Overview The image to the left is of the iris camera being used during the test. A person using this would be standing about the same location at which this picture was taken. The Kinect camera is located behind the individual during this capture process This box represent the optimal area to be standing when using the iris camera. The red line represents the farthest you should be when using the camera and the green line represents the closest you should be. The image above shows the Kinect camera circled in black as well as the red square in which a person would stand during the test. This image shows an area being selected for zoom. The selected area is represented as the gray box near the center of the image. This is selected by holding left click on the graph then dragging the mouse either right or left. The excel data above shows a subjects positioning 3 dimensionally that was captured from the Kinect camera. For this project we used column A (Timestamp) and P (Head Z) to create the image to the right which is what the program draws using the Head Z position for the Y- axis and timestamp for the X-axis. The output is shown here. The image shows each interaction that occurred in the total transaction. This last image represents what is displayed after selecting an area and releasing left click. Being able to zoom in on these sections will allow for a deeper understanding of HBSI interactions. The three images below give an example of the zooming process from left to right.