Take a sneak peek at the new NIST.gov and let us know what you think!
(Please note: some content may not be complete on the beta site.).
(Beta Version 2.2)
VASIR (Video-based Automatic System for Iris Recognition) is a NIST-implemented iris recognition system designed for both conventional iris images and iris images collected at a distance.
VASIR is a fully automated system for video-based iris recognition, capable of handling videos that were 1) captured under less-constrained environmental conditions and 2) contained other features (e.g., face, head, hair, neck, shoulder) apart from the eye/iris region, such as a person walking through a portal at a distance. The VASIR system was designed, developed, and optimized to be robust—to address the challenge of recognizing a person using the human iris in less-than-optimal environments.
Although VASIR was developed with less-constrained video-based iris recognition in mind, VASIR also robustly accommodates traditional constrained still-based iris recognition. VASIR hence supports multiple matching scenarios:
1) Less-constrained (non-ideal) video to less constrained (non-ideal) video (VV),
2) Less-constrained (non-ideal) video to constrained (ideal) still image (VS), and
3) Constrained (ideal) still to constrained (ideal) still image (SS) iris recognition.
In VV matching, the extracted iris region of video frames taken at distance is matched to frames from a different video sequence of the same person. The VS matching scenario means that the video frame (e.g., taken at a distance using an IOM system) is compared to classical still-images (e.g., via a LG2200 system)—matching samples captured by a different device. SS matching is the traditional iris recognition scenario in which a classical still image is matched against other classical still images of the same person—all of which were captured by the same device.
VASIR has the capacity to load both still images and video sequences, to automatically detect and extract the eye region, and to subsequently assess and select the best quality iris image from NIR face-visible video captured at a certain distance. After this process, VASIR carries out a comprehensive image comparison analysis that in turn yields a state-of-the-art individual verification.
The VASIR system has into three primary modules:
1) Image Acquisition,
2) Video Processing, and
3) Iris Recognition.
In the Image Acquisition module, VASIR loads the image or video sequence. For evaluation, we use publicly available datasets.
In the Video Processing module, VASIR automatically distinguishes the eye region from face, hair, ear, and shoulder that may be visible in video frames and subsequently extracts the left/right iris sub-images separately. VASIR then automatically calculates an image quality score of the extracted iris sub-images. Based on the quality score, the highest quality iris images—one for left and one for right—are automatically chosen out of all available frames. The Video Processing module consists of four sub-components: Eye Region Detection, Eye Region Image Extraction, Image Quality Measurement, and Best Quality Image Selection.
The Iris Recognition module is then fed either the resulting iris images from the Video Processing module or the loaded still image. For both video and still iris images, VASIR segments the iris region based on VASIR’s segmentation approach. The segmented iris regions are then extracted and normalized based on polar coordinates transformations and bilinear interpolation. Next, VASIR extracts the features from the normalized iris images and encodes the extracted features as binary strings along with a VASIR-developed noise-masking scheme. At the end, VASIR matches the encoded biometric iris template to other existing iris templates. The Iris Recognition module consists of four components: segmentation, normalization, feature extraction/encoding, and similarity distance metric. All three module procedures are completely automatic.
To assess VASIR's performance and practical feasibility, the Iris Challenge Evaluation (ICE) dataset and the Multiple Biometric Grand Challenge (MBGC) dataset were employed, with results as described in the publications below (Please cite the following paper when using the VASIR system).
Yooyoung Lee, Ross J. Micheals, P. J. Phillips, James J. Filliben, “VASIR: An Open-Source Research Platform for Advanced Iris Recognition Technologies", Journal of Research of NIST, Volume 118, p218-259, 2013
The detailed evaluation protocol and performance for each dataset are described in Section 6 of this paper.
Currently, the VASIR source code beta version 2.2 and its user guide are available. This source code is incomplete. Note that you may get a known warning or memory related messages.
We plan to update each component of VASIR incrementally, after evaluating its performance.
Type of software: Open source research software for automated iris recognition.
Yooyoung Lee, Ross J. Micheals, P. Jonathon Phillips, James J. Filliben
This software is released by NIST as a service and is expressly provided "AS IS." NIST MAKES NO WARRANTY OF ANY KIND, EXPRESS, IMPLIED OR STATUTORY, INCLUDING, WITHOUT LIMITATION, THE IMPLIED WARRANTY OF MERCHANTABILITY, AND FITNESS FOR A PARTICULAR PURPOSE, NONINFRINGEMENT AND DATA ACCURACY. NIST DOES NOT REPRESENT OR WARRANT THAT THE OPERATION OF THE SOFTWARE WILL BE UNINTERRUPTED OR ERROR-FREE, OR THAT ANY DEFECTS WILL BE CORRECTED.
Certain trade names and company products are mentioned in the text or identified. In no case does such identification imply recommendation or endorsement by the National Institute of Standards and Technology, nor does it imply that the products are necessarily the best available for the purpose.
Craig Watson / Yooyoung Lee