Next Article in Journal
Adaptive Real-Time Convergence Estimation for Enhancing Reliability of Time Synchronization in Distributed Energy Monitoring System
Previous Article in Journal
Miniaturized Patch Array Antenna Using CSRR Structures for 5G Millimeter-Wave Communication Systems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

VRBiom: A New Periocular Dataset for Biometric Applications of Head-Mounted Display

by
Ketan Kotwal
1,*,
Ibrahim Ulucan
1,
Gökhan Özbulak
1,2,
Janani Selliah
1 and
Sébastien Marcel
1,3
1
Idiap Research Institute, 1920 Martigny, Switzerland
2
Department of Electrical Engineering, École Polytechnique Fédérale de Lausanne, 1015 Lausanne, Switzerland
3
Ecole des Sciences Criminelles, Université de Lausanne, 1015 Lausanne, Switzerland
*
Author to whom correspondence should be addressed.
Electronics 2025, 14(9), 1835; https://doi.org/10.3390/electronics14091835
Submission received: 21 March 2025 / Revised: 18 April 2025 / Accepted: 23 April 2025 / Published: 30 April 2025

Abstract

:
With advancements in hardware, high-quality head-mounted display (HMD) devices are being developed by numerous companies, driving increased consumer interest in AR, VR, and MR applications. This proliferation of HMD devices opens up possibilities for a wide range of applications beyond entertainment. Most commercially available HMD devices are equipped with internal inward-facing cameras to record the periocular areas. Given the nature of these devices and captured data, many applications such as biometric authentication and gaze analysis become feasible. To effectively explore the potential of HMDs for these diverse use-cases and to enhance the corresponding techniques, it is essential to have an HMD dataset that captures realistic scenarios. In this work, we present a new dataset of periocular videos acquired using a virtual reality headset called VRBiom. The VRBiom, targeted at biometric applications, consists of 900 short videos acquired from 25 individuals recorded in the NIR spectrum. These 10 s long videos have been captured using the internal tracking cameras of Meta Quest Pro at 72 FPS. To encompass real-world variations, the dataset includes recordings under three gaze conditions: steady, moving, and partially closed eyes. We have also ensured an equal split of recordings without and with glasses to facilitate the analysis of eye-wear. These videos, characterized by non-frontal views of the eye and relatively low spatial resolutions ( 400 × 400 ), can be instrumental in advancing state-of-the-art research across various biometric applications. The VRBiom dataset can be utilized to evaluate, train, or adapt models for biometric use-cases such as iris and/or periocular recognition and associated sub-tasks such as detection and semantic segmentation. In addition to data from real individuals, we have included around 1100 presentation attacks constructed from 92 PA instruments. These PAIs fall into six categories constructed through combinations of print attacks (real and synthetic identities), fake 3D eyeballs, plastic eyes, and various types of masks and mannequins. These PA videos, combined with genuine (bona fide) data, can be utilized to address concerns related to spoofing, which is a significant threat if these devices are to be used for authentication. The VRBiom dataset is publicly available for research purposes related to biometric applications only.

1. Introduction

The rise in head-mounted displays (HMDs) in recent years has significantly transformed the way we experience digital content. With its Meta Quest series of virtual reality (VR) headsets, Meta (previously Facebook) is the largest headset platform as of 2024. Recently, Apple released its mixed reality (MR) headset named Apple Vision Pro, with specific focus on the spatial computing aspects of the device. Additionally, Sony Interactive Entertainment has been actively working on VR for gaming applications with its PlayStation (PS VR) series. Initially, these devices were primarily designed for immersive entertainment, offering users experiences in augmented, virtual, or mixed reality. However, the applications of HMDs extend far beyond gaming and entertainment, encompassing diverse fields such as education [1,2], professional training [3,4], healthcare [5,6,7], and biometric authentication [8,9]. Extensive reviews on the application of HMDs across various domains can be found in [10,11,12,13].
Besides its core functionality, an interesting feature of HMDs is the integrated multiple internal cameras. These cameras, typically located in the surroundings of the user’s eye region, enhance the immersive experience by tracking user’s eye movements and capturing images and videos of the user’s eyes and surrounding regions. This source of data opens up a wide range of potential applications to be explored, including biometric authentication. In biometrics, the HMDs can be instrumental in user recognition based on their iris and/or periocular traits [9]. Such applications can be employed for both identification and verification purposes, ensuring that the user accessing the device is indeed who they claim to be. In addition to the recognition accuracy, in biometric applications, safeguarding against attacks and ensuring robustness is critical. The presentation attacks (PAs), also known as spoofing, pose a serious challenge to biometric systems. These attacks can involve the use of masks, synthetic eyes, contact lenses, or printed images to deceive the system into granting unauthorized access [14,15]. Therefore, developing effective presentation attack detection (PAD) or anti-spoofing mechanisms is essential for the practical use of biometric authentication systems in HMDs. Beyond biometrics, HMDs have significant potential in other applications, such as semantic segmentation of the eye region [8,16]. This is particularly relevant for industrial, entertainment, and biomedical applications where continuous tracking of the gaze and eye movement is necessary. HMDs can facilitate near-continuous gaze tracking, which has applications ranging from interactive gaming to medical diagnostics [17,18,19]. Figure 1 depicts some of the possible biometric use-cases of HMD data.
The aforementioned applications have been a part of research and commercial deployment for several years. However, most existing research and user applications have focused on data captured by sensors positioned to capture a frontal view of the eye. The design of recent popular HMDs typically places displays in front of the user’s eyes for an immersive experience, and thereby necessitates the eye tracking cameras to be positioned in the surrounding areas. This placement results in an oblique, non-frontal view of the eye region, posing unique challenges for biometric applications. Additionally, the cameras integrated into these HMDs are often of lower resolution and must operate within constrained resources. Moreover, the fit of HMDs varies based on the individual’s inter-eye distance and nose bridge shape, resulting in variations in the captured regions for different users.
These factors indicate the need for specific studies and datasets acquired directly from HMD devices to address the above challenges. Such datasets can be invaluable for multiple purposes, including the following:
  • Benchmarking the performance of existing methods/models on HMD data;
  • Fine-tuning models to address domain shifts (introduced by different capturing angles, devices, resolutions, etc.);
  • Developing new methods tailored to the unique requirements of HMDs.
To address these challenges, in the realm of biometrics, we have created the VRBiom dataset: a new Virtual Reality dataset for Biometric applications acquired from the Meta Quest Pro. One subset of this dataset comprises 900 short videos, each lasting nearly 10 s, collected from 25 individuals under various conditions, including steady gaze, moving gaze, and half-closed eye settings. To ensure evaluation and development of various use-cases in realistic scenarios, half of these videos were recorded with subjects wearing glasses, while the remaining half were recorded without glasses. These videos present potential for a wide range of applications, such as biometric recognition and semantic segmentation. It should be noted that conventional biometric recognition pipelines comprise a sequence of tasks, including region of interest (RoI) detection, segmentation, and orientation. The presented dataset may be harnessed to enhance each of these tasks separately. We have also captured recordings using different types of 3D masks, combined with synthetic eyes such as 2D printouts and 3D fake eyes. These 1104 recordings, also called as presentation attacks (PAs), comprise other subsets of the VRBiom dataset meant for addressing concerns related to PAD and anti-spoofing measures. The overall VRBiom dataset consists of 2004 recordings of periocular (left and right) regions captured in near-infrared (NIR) spectrum for 10 s at 72 FPS.
The present release of the VRBiom dataset consists of videos and subject-level (identity) labels. It also provides details of the scenarios and PA instruments used. We do not provide pixel-level annotations for semantic details. The contributions of our work can be summarized as follows:
  • We have created and publicly released the VRBiom dataset [20]. The VRBiom dataset is a collection of more than 2000 periocular videos acquired from a VR device. To the best of our knowledge, this is the first publicly available dataset featuring realistic, non-frontal views of the periocular region for a variety of biometric applications.
  • The VRBiom dataset offers a range of realistic and challenging variations for the benchmarking and development of biometric applications. It includes recordings under three gaze conditions: steady gaze, moving gaze, and partially closed eyes. Additionally, it provides an equal split of recordings with and without glasses, facilitating the analysis of the impact of eye-wear on iris/periocular recognition.
  • As a part of VRBiom, we have also released more than 1100 PA videos constructed from 92 attack instruments. These attacks, incorporating different combinations of PAIs such as prints, fake 3D eyeballs, and various masks, provide a valuable resource for advancing PAD research in a VR setup.
The organization of this paper is structured as follows: Section 2 provides a brief review of iris/periocular datasets and HMD-based datasets. Section 3 explains details of the VRBiom dataset, including data collection methods and associated challenges. In Section 4, we discuss potential use-cases of the dataset. Finally, we summarize the work in Section 5.

2. Related Datasets

In this section, we briefly present some commonly used datasets for iris/periocular biometrics, followed by a review of recent datasets collected using HMD devices.

2.1. Datasets for Iris/Periocular Biometrics

One of the most prominent and commonly used iris datasets is the CASIA-Iris series, collected by researchers from the Chinese Academy of Sciences, which consists of four main versions denoted with suffixes V1–V4 [21]. The first version, V1, released in 2002, comprised 756 images taken from 108 subjects at a resolution of 320 × 280 . In 2004, CASIA-IrisV2 was released, containing 2400 images from 120 individuals at VGA resolution ( 640 × 480 ). The third version included three different subsets: Interval, Lamp, and Twins. The Casia-Interval subset was captured from 249 subjects in an indoor environment, while the Lamp subset provided 16 k images at VGA resolution. The Twins subset contained recordings of 200 twins, primarily children, in an outdoor environment [22]. The latest version, CASIA-IrisV4, is the most comprehensive, including 54,601 images from 2800 subjects distributed across six subsets namely CASIA-Iris-Interval, Lamp, Twins, Distance, Thousand, and Syn [21,23].
The Iris Liveness Detection Competition (LivDet-Iris) [24,25] dataset combines multiple datasets designed for liveness detection in the context of iris PAD. The 2017 competition utilized four different datasets created by Clarkson, Warsaw, Notre Dame (ND), Indraprastha Institute of Information Technology Delhi (IIIT-Delhi), and West Virginia University (WVU) [24]. The Clarkson and Warsaw datasets mainly focus on print attacks, consisting of 8095 and 12,013 images, respectively. The ND CLD 2015 and IIITD-WVU datasets comprise more than 7 k images each, acquired at VGA resolution. The recent edition of the competition, held in conjunction with IJCB 2023, aimed at creating more challenging attacks [25]. This dataset included attack instruments such as print, contact lenses, electronic displays, fake/prosthetic eyes, and synthetic iris of varying quality.
Researchers at the University of Beira have released a series of UBIRIS datasets of iris images. The UBIRIS v1 dataset provides 1877 images acquired under less constrained imaging conditions [26], whereas UBIRIS v2 offers more than 11 k images captured at a distance [27]. Another version, UBIPr, derived from UBIRIS v2, provides wider regions suitable for periocular recognition. They have also released pixel-level masks of semantic elements such as the iris and sclera, making it a useful resource for segmentation applications. Additionally, the Multimedia University (MMU) released two iris image datasets in 2010 that are publicly available. The first version, MMU-v1, consists of 450 images obtained from 45 subjects, while the subsequently released MMU-v2 provides 995 images from 100 subjects [28,29].
Table 1 summarizes the iris and periocular biometric datasets, providing details related to samples and attacks where applicable.

2.2. HMD Datasets for Iris/ Periocular Biometrics

Due to the inherent challenges in collecting data using HMD devices and the limitations of hardware quality, there have been fewer attempts to gather such datasets. However, with recent advancements in hardware technology and the increasing number of companies releasing HMD headsets, these types of datasets are gaining traction in both research and end-user applications.
Released in 2012, the Point of Gaze (PoG) dataset is one of the HMD datasets created for gaze detection and head pose studies [30]. It comprises data from 20 subjects recorded at a resolution of 768 × 480 . The Labelled Pupils in the Wild (LPW) dataset contains 66 eye-region videos at VGA resolution captured at 95 FPS [31]. As its name suggests, the LPW dataset features several variations in illumination, eye-wear (glasses and contact lenses), and makeup, along with gaze directions. Kim et al. introduced the NVGaze dataset, which was collected from 35 subjects [32]. This dataset focuses on gaze estimation and includes approximately 2.5 million bona fide images at VGA resolution.
The Open Eye Dataset (OpenEDS) has two editions: OpenEDS 2019 and 2020, both released by Facebook Reality Labs, which have significantly advanced the field of HMD-based eye-tracking research. The OpenEDS 2019 dataset [33] comprises videos and images from 152 subjects and is partially annotated for semantic segmentation. The 2020 version [34] includes two sub-datasets (the Gaze Prediction and Eye Segmentation Datasets) from 80 subjects, comprising approximately 580 k images with a resolution of 640 × 400 pixels, captured at 100 FPS.
The creators of OpenEDS also organized a competition (challenge) to improve performance in tracking and segmentation tasks. While OpenEDS is similar to our presented work, it differs in two main aspects: first, OpenEDS data are acquired from eye-facing cameras, yielding frontal eye views, whereas our dataset uses a consumer-level HMD capturing eyes and periocular regions from non-frontal angles. Second, OpenEDS includes recordings only from real (bona fide) subjects, while our dataset incorporates a variety of presentation attacks (PAs), making it valuable for PAD research as well [33,34].
Our dataset, VRBiom, consists of 900 short videos from 25 participants and more than 1100 videos across six categories of attack instruments. Each video, with a spatial resolution of 400 × 400 , is approximately 10-s long and captured at 72 FPS. Table 2 provides a brief comparison of the publicly available HMD datasets for research.

3. The VRBiom Dataset

In this section, we describe the process of acquiring both bona fide and PA data, along with relevant statistics. We also discuss some of the challenges encountered in the creation of the VRBiom dataset.

3.1. Collection Setup

A total of 25 subjects, aged between 18 and 50 and representing a diverse range of skin tones and eye colours, participated in the data collection process. Each participant was briefed on the project objectives and provided their consent through a signed consent form. The data acquisition for each participant was conducted in a single session lasting approximately 20 min. We divided the recordings into two sub-sessions: the first without the subject wearing glasses and the second with glasses. For each sub-session, we captured three recordings involving different gaze variations: steady gaze, moving gaze, and partially closed eyes. Each video had a duration of approximately 10 s, recorded at a frame rate of 72 frames per second (FPS). The videos have a spatial resolution of 400 × 400 pixels and were captured in a single-channel in the NIR spectrum. It was often observed that for nearly first second of most recordings, the frames were over-exposed to the NIR illumination. Therefore, we recommend discarding the first 70–80 frames from processing.
The iris/periocular videos in the VRBiom dataset were captured by the inward-facing cameras located in the Meta Quest Pro headset. Table 3 presents the technical specifications of the Meta Quest Pro, specifically relevant to our data acquisition process [35].
For bona fide recordings, the subjects wore the HMD device, whereas for recordings of PAs, we systematically positioned the Meta Quest Pro on the attack instrument (as described later in this section). The Meta Quest Pro was connected to a workstation via USB-C cable which facilitated both charging and the transfer of recording commands and data. Figure 2a,b depict the recording setups for bona fide and PAs, respectively.
To streamline the data collection, we developed the Meta Quest Pro Data Capture App (app) (refer to Figure 2c for the GUI). This app enabled us to set file names, manage recordings, and transfer files from the Meta Quest Pro to our workstation efficiently. Additionally, the app provided tools to convert the video recordings from the proprietary vrs format to the conventional avi format, which is compatible with most common video players and software. Through the capture app, we also ensured a specific naming convention for each recoded video. The naming convention consisted of the following fields: Type (bona fide or PA), subject ID (or PA ID), Eye (left or right), glasses (without or with), type of gaze (Steady, Moving, or Partially closed), claimed ID (for PAs), details of PAI (for PAs), recording ID, and a random suffix. These fields, with their mapped values, were used to create the file name for each video. More details on this can be found in the actual dataset.

3.2. Bona Fide Recordings

Before starting the recording, we ensured that the HMD (Meta Quest Pro) is securely and comfortably fitted around the subject’s head, providing them with an immersive virtual experience. Considering the diverse array of use-cases, we captured videos under three different gaze variations:
  • Steady Gaze: The subject maintains a nearly fixed gaze position by fixating their eyes on a specific (virtual) object.
  • Moving Gaze: The subject’s gaze moves freely across the scene.
  • Partially Closed Eyes: The subject keeps their eyes partially closed without focusing on any particular gaze.
These variations were recorded under two eye-wear conditions: with glasses and without glasses. If the subjects did not have their own medical glasses, they were provided with a pair of fake glasses. Each gaze variation and glass-wearing condition were repeated three times to ensure a robust dataset.
Figure 3 provides samples of each of the aforementioned variations. The HMD is equipped with cameras that simultaneously record videos of both eyes. Therefore, for each recording, two videos are captured concurrently—one for the left eye and one for the right eye. In total, we collected 900 bona fide videos from 25 subjects in different conditions, as summarized in Table 4.
Thus, the structure and coverage of the VRBiom dataset ensure that each subject’s iris/periocular region is recorded across various realistic scenarios, providing a valuable resource for researchers to analyse and develop biometric applications.

3.3. PA Recordings

Most of the PAs were constructed by combining two categories of attack instruments: those targeting the eyes and those targeting the periocular region. For the eyes, we used a variety of instruments including fake 3D eyes (eyeballs), printouts from synthetic and real identities, and plastic-made synthetic eyes. For the periocular region, we employed mannequins and 3D masks made of different materials. These masks also served as a fake head where the HMD could be securely placed for recording.
Using these combinations, we created six categories of PAIs, each with a variable number of attack instruments. Samples of these categories are provided in Figure 4, with a brief description as follows:
  • Mannequins: We used a collection of seven mannequins made of plastic, each featuring its own eyes. These mannequins are generic (fake heads) as they do not represent any specific bona fide identity. Unlike the requirements of additional components for other PAIs to provide a stable platform or extra equipment to stabilize HMDs, these mannequins are particularly advantageous due to their integrated and stable platform for placing HMDs.
  • Custom Rigid Masks (Type-I): This category includes ten custom rigid masks, comprising five men and five women. The term custom implies that the masks were modelled after real individuals. Manufactured by REAL-f Co. Ltd. (Japan), these masks are made from a mixture of sandstone powder and resin. They represent real persons (though not part of the bona fide subjects in this work) and include eyes made of similar material with synthetic eyelashes attached.
  • Custom Rigid Masks (Type-II): We used another collection of 14 custom rigid masks to construct the PAs for the VRBiom. These masks differ from the previous category in two ways: first, they are made of amorphous powder compacted with resin material by Dig:ED (Germany); second, they have empty spaces at the eye locations, where we inserted fake 3D glass eyeballs to construct an attack.
  • Generic Flexible Masks: This category includes twenty flexible masks made of silicone. These full-head masks do not represent any specific identity, hence termed generic. The masks have empty holes for eyes, where we inserted printouts of periocular regions from synthetic identities. Using synthetic identities alleviates privacy concerns typically associated with creating biometric PAs.
  • Custom Flexible Masks: These silicone masks represent real individuals. Similar to the previous attack categories, these masks have holes at the eye locations, where we inserted fake 3D eyes to create the attacks. We have 16 PAIs in this category.
  • Vulnerability Attacks: For each subject in the bona fide collection, we created a print attack. For both eye-wear conditions (without and with glasses), we manually selected an appropriate frame from each subject’s bona fide videos and printed it at true-scale on a laser printer. The printouts were cut into periocular crops and placed on a mannequin to resemble a realistic appearance when viewed by the tracking cameras of the HMD. The so-obtained 25 print PAs can be used to assess the vulnerability of the corresponding biometric recognition system. These attacks simulate scenarios where an attacker gains access to (unencrypted) data of authorized individuals and constructs simple attacks using printouts.
Figure 4. Samples of PA recordings from VRBiom dataset. The top row represents the PAI captured in RGB (visible) spectrum, while middle and bottom rows depict the NIR recordings without and with glasses, respectively, as acquired by the internal (right) camera of the Meta Quest Pro. From left to right, each column presents a sample of the type of PAIs belonging to the attack series from 2 to 7.
Figure 4. Samples of PA recordings from VRBiom dataset. The top row represents the PAI captured in RGB (visible) spectrum, while middle and bottom rows depict the NIR recordings without and with glasses, respectively, as acquired by the internal (right) camera of the Meta Quest Pro. From left to right, each column presents a sample of the type of PAIs belonging to the attack series from 2 to 7.
Electronics 14 01835 g004
Table 5 summarizes the details of the VRBiom dataset comprising 900 bona fide and 1104 PA videos. It also provides the naming conventions used to indicate the type of PAI.

3.4. Challenges

Here, we briefly discuss some of the challenges encountered during data collection, and the corresponding solutions employed:
  • Eyelashes: During the initial data capture experiments, we observed that the subjects’ eyelashes appeared as a prominent feature, especially when capturing the periocular region. The absence of eyelashes in most PAIs, where prints or 3D eyeballs were used to construct the attack, can be considered as an easy-to-distinguish yet unrealistic feature. To create a more realistic scenario, we decided to use false eyelashes from standard makeup kits.
  • PA Recordings: Recording the PAs was challenging due to the lack of real-time feedback from the device or capture app. To achieve the correct positioning and angle for various PAIs, multiple attempts and recording trials were often required for data collectors to ensure the desired quality of the captured data.
  • NIR Camera’s Over-exposure: The proximity between the attack instrument and the NIR illuminator/receiver often resulted in over-exposed recordings. Due to the fixed distance between the internal light source (illuminator), the subject’s skin (or the material surface in the case of PAs), and the camera, there was limited flexibility to adjust the lighting conditions during acquisition. However, we attempted minor adjustments to the HMD’s orientation and the placement of the PAI to mitigate exposure issues to a considerable extent.
  • Synthetic Eyes: One challenge we faced was the variability in the identity of synthetic eyes. To ensure accurate biometric measurements, we needed to carefully select synthetic eyes that closely resemble human eyes.
  • Print Attack: The printout attacks designed for vulnerability assessment are inherently 2-dimensional, whereas the periocular region’s appearance is 3D. To simulate realistic appearance, we employed small paper balls to roll the printed attacks, thereby providing a structural view akin to 3D representation.
  • Comfort and Safety: Prolonged use of VR devices (as well as smart and wearable devices) has often been associated with user discomfort, including symptoms such as dizziness or uneasiness, and effects of Specific Absorption Rate (SAR) [36,37,38,39]. In our case, since each recording session lasted only a few minutes, these concerns were not encountered. However, we mention this issue for the benefit of future researchers who may intend to conduct longer-duration data collection.

4. Potential Use-Cases of the VRBiom Dataset

This section outlines the potential use-cases of the newly created VRBiom dataset, emphasizing its applicability in biometric applications such as iris and periocular detection/recognition, presentation attack detection (PAD), as well as intriguing tasks (or sub-tasks) associated with biometrics such as semantic segmentation focusing on the eye, iris, and sclera regions. The overview of these use-cases is illustrated in Figure 1.
The use-cases discussed here, although well established in research and development, are highlighted more as possibilities than substantiated claims of direct applicability. The VRBiom dataset can serve as a benchmark to evaluate the effectiveness of current HMD data for these applications. Alternatively, it can be instrumental in developing (or fine-tuning) applications specifically for data obtained from HMD devices. The present data can be characterized by non-frontal views of periocular regions, relatively low spatial resolutions, and limited resources. Thus, some applications discussed in this section may prove to be quite challenging. This, however, presents interesting research problems to transform these data into better-quality ones, as well as to develop robust applications.
Needless to mention, the advances in HMD technology will not only enhance the quality of the data but also expand the scope of applications.

4.1. Iris and Periocular Recognition

The VRBiom dataset comprises 900 video recordings from 25 bona fide subjects. Each subject participated in sessions combining three gaze and two eye-wear variations, thus providing a valuable resource for sampling and analysing iris and periocular regions for detection and recognition tasks. When sampled at the frame level, videos with moving gaze can be treated as individual frames with varied gazes. These individual frames facilitate the localization and enrolment of the eye region, making them suitable for both detection and recognition purposes. Detection involves localizing specific regions of interest (RoIs), such as the iris, eye, or periocular regions, while recognition can be performed in both 1:1 verification and 1:N identification scenarios.
To evaluate recognition performance, two widely adopted metrics are accuracy (Acc) and Equal Error Rate (EER). Here, accuracy represents the proportion of correctly classified instances over the total number of samples. If TP, TN, FP, and FN indicate the number of true positives, true negatives, false positives, and false negatives, respectively, then accuracy is defined as:
Acc = T P + T N T P + F P + F N + T N .
The Equal Error Rate (EER) refers to the error rate at which the False Match Rate (FMR) and the False Non-Match Rate (FNMR) are (almost) equal. Formally, it is expressed as
EER ( τ ) = arg   min τ   FMR ( τ ) FNMR ( τ ) ,
where τ denotes the threshold at which the FMR and FNMR are evaluated. It is necessary to mention that the relatively low resolution and non-frontal capture angles of the samples in the VRBiom dataset may introduce challenges and limitations for achieving high recognition performance, particularly when compared to existing benchmarks. However, it is likely that those benchmarks were collected under controlled conditions using devices with significantly higher resolution. Since our dataset is acquired using a commercially available, real-world VR device, it highlights the need for developing and refining recognition methods that are robust under such practical constraints to advance the current state of the art.
Based on the design of the pipeline, recent iris/periocular recognition methods can be categorized into feature extraction-based or classification-based methods. Feature extraction-based methods employ models, often deep convolutional neural network (CNN) architectures, as feature extractors to obtain a compact representation (embedding) of a preprocessed input (i.e., iris/periocular images). In these methods, preprocessing and matching/scoring are typically separated from feature extraction. Classification-based methods, on the other hand, treat the recognition problem in an end-to-end manner, training the iris dataset in a supervised learning setting.
An overview of iris recognition methods using handcrafted features and classical machine learning techniques can be found in [40]. A recent work by Nguyen et al. provides a systematic survey of deep learning (DL)-based methods for iris recognition [29]. Another comprehensive review of iris recognition methods is presented in [41], which details methods from both categories and various preprocessing techniques for the first category (non-end-to-end methods). Alonso-Fernandez and Bigun discuss several handcrafted feature-based methods for periocular recognition, categorizing them into texture-, shape-, and colour-based methods [42].
Iris recognition remains a topic of significant interest within the biometric community. Over the past decade, several challenges have been organized to advance iris recognition performance [43,44]. The challenge described in [44] was specifically dedicated to mobile iris recognition. Boutros et al. benchmarked the OpenEDS [33]—an HMD dataset—for iris recognition [9,45]. Although this dataset is acquired from an HMD device, it captures nearly frontal view of the eye regions, making it easy to detect, localize, and recognize an individual.
The use of HMD data for biometric recognition is still in its early stages and is yet to gain mainstream attention. A major challenge posed by HMD-based iris detection is the oblique view of the RoI, which often results in localization failures and leads to failure to acquire (FtA). The non-frontal view, even after normalization, may cause feature/data distortion. This, combined with the relatively small size of the RoI, results in the loss of fine, subtle features that may present discriminatory information. Several quality metrics for iris and periocular images, including low-quality images, have been presented in [40]. A study of the quality assessment of HMD-based data for recognition purposes may provide useful insights towards designing new methods.

4.2. Iris and Periocular PAD

For a biometric authentication system to be practically deployable, its resistance to presentation attacks (PAs) is crucial. The importance of PAD for iris and periocular traits has been well recognized by the biometric community. Several works by A. Czajka, K. Bowyer, and colleagues present comprehensive reviews of iris PAD methods and datasets [15,46]. However, these works do not discuss datasets or PAD methods for HMD-based datasets, likely due to a lack of relevant data. With VRBiom, we provide the first PAD dataset acquired using HMD devices, containing approximately 1100 short videos of PAs. Through a combination of attack instruments for the eyes and surrounding regions (refer to Section 3), we have created a variety of attack scenarios for PAD tasks. The inclusion of print attacks of the bona fide samples allows for the vulnerability analysis of different PAD methods.
Several studies have also proposed approaches that combine both iris and periocular regions for PA detection [47,48]. Such approaches can be explored for the VRBiom dataset if the detection and localization of iris do not yield satisfactory results.
LivDet-Iris [49] is a well-known competition regularly organized to compare liveness detection methods (i.e., to differentiate between real human and fake samples). The fifth instalment of this competition was held in conjunction with IJCB in 2023 [25]. Despite the participation of reputable academic institutions, the average classification error rates were as high as 22–37% (the variation refers to the weighing mechanism of different PAIs). Although the quality of data and types of attacks in the competition datasets and VRBiom dataset differ, this highlights the challenging and unresolved nature of the iris and/or periocular PAD problem.

4.3. Semantic Segmentation

Semantic segmentation of the eye to localize components such as the iris and sclera has demonstrated its usefulness in both biometric and non-biometric tasks. This segmentation process is often an integral and crucial stage in recognition- and PAD-related tasks discussed in the previous section. The creators of the OpenEDS dataset have also highlighted segmentation as a key application of interest [33]. Segmentation of individual traits, such as the iris [50] and sclera [51], has received considerable attention in the literature. Recently, multi-class semantic segmentation—which involves the simultaneous detection and localization of various parts of the eye region—has gained interest as well [52,53,54].
The effectiveness of segmentation methods is typically evaluated using the mean Intersection over Union (mIoU). This metric represents the average Intersection over Union (IoU) computed across all classes and is defined as
mIoU = 1 C c = 1 C IoU c ,
where C denotes the total number of classes, and IoU c refers to the Intersection over Union for class c, calculated as
IoU c = T P T P + F P + F N .
Here, T P is the number of true positives, F P is the number of false positives, and F N is the number of false negatives for the given class. A higher mIoU indicates better segmentation accuracy across all classes.
Similar to the previous use-cases, the topic of eye segmentation has been further advanced by numerous competitions. For instance, challenges such as NIR-ISL [55] and SSRBC [56] focus specifically on iris and sclera segmentation, respectively. The VRBiom dataset (after annotations) can also be used to benchmark the performance of existing segmentation methods, with reference to biometric applications, on realistic HMD data captured in the NIR spectrum. Given the variation in the exact angle of capture due to differences in subjects’ face shapes and inter-eye distances, the frames in the VRBiom dataset present a challenging dataset. As typical segmentation methods are usually trained on frontal views of the eye, one may be required to address this domain gap by designing implicit or explicit affine transformations.

5. Summary

In this work, we introduced the VRBiom dataset, a novel dataset of periocular videos captured using the VR device, Meta Quest Pro. This dataset is the first publicly available resource offering realistic, non-frontal views of the periocular region, comprising 900 short videos from 25 subjects and 1104 presentation attack (PA) videos using 92 different attack instruments. The dataset includes diverse and challenging scenarios such as steady gaze, moving gaze, and partially closed eyes, with an equal split of recordings with and without glasses.
The HMD-captured data are relatively novel and present significant potential for various applications that are yet to be fully explored. However, these also pose unique challenges such as non-frontal views, user-specific fitting issues, and low resolutions. With the VRBiom dataset, we hope to provide a valuable resource for understanding and advancing biometric applications associated with AR and VR devices. The inclusion of various attack recordings also provides an opportunity to enhance anti-spoofing measures for VR-based authentication systems.
We have outlined the data collection process and associated challenges, providing useful insights for researchers aiming to acquire similar data from VR devices. The VRBiom dataset is publicly available to support advancements in biometric research, including authentication, semantic segmentation, and presentation attack detection (PAD), along with related sub-tasks.

Author Contributions

Conceptualization, K.K. and S.M.; methodology, K.K.; validation, I.U. and G.Ö.; data curation, G.Ö., J.S. and I.U.; writing—original draft preparation, K.K.; writing—review and editing, K.K., I.U., G.Ö. and J.S.; supervision and funding acquisition, S.M. All authors have read and agreed to the published version of the manuscript.

Funding

We would like to gratefully acknowledge funding support from the Meta program “Towards Trustworthy Products in AR, VR, and Smart Devices” (2022) Meta Research Program and the Swiss Center for Biometrics Research and Testing.

Institutional Review Board Statement

The Idiap Research Institute has an internal body, the Data and Research Ethics Committee (DREC), to ensure that the research projects, and in particular the non-invasive human research projects, submitted for review are designed in compliance with fundamental ethical principles. The project supporting the present research has passed the review by DREC.

Informed Consent Statement

Volunteered data subjects gave their explicit and informed consent by signing a consent form prior to the collection of research data. This consent form complies with Swiss law and the European General Data Protection Regulation (EU GDPR).

Data Availability Statement

The VRBiom dataset, outlined in this article, is accessible for non-commercial research use, specifically in the areas of biometric authentication and presentation attack detection. Researchers can obtain access to the dataset upon signing an EULA. For further information, please visit: https://www.idiap.ch/dataset/vrbiom (accessed on 31 March 2025).

Acknowledgments

The authors would like to thank Yannick Dayer (Idiap Research Institute) for creating a data capture application.

Conflicts of Interest

The authors declare no conflicts of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

References

  1. Radianti, J.; Majchrzak, T.A.; Fromm, J.; Wohlgenannt, I. A systematic review of immersive virtual reality applications for higher education: Design elements, lessons learned, and research agenda. Comput. Educ. 2020, 147, 103778. [Google Scholar] [CrossRef]
  2. Kamińska, D.; Sapiński, T.; Wiak, S.; Tikk, T.; Haamer, R.E.; Avots, E.; Helmi, A.; Ozcinar, C.; Anbarjafari, G. Virtual reality and its applications in education: Survey. Information 2019, 10, 318. [Google Scholar] [CrossRef]
  3. Xie, B.; Liu, H.; Alghofaili, R.; Zhang, Y.; Jiang, Y.; Lobo, F.D.; Li, C.; Li, W.; Huang, H.; Akdere, M.; et al. A review on virtual reality skill training applications. Front. Virtual Real. 2021, 2, 645153. [Google Scholar] [CrossRef]
  4. Chiang, F.K.; Shang, X.; Qiao, L. Augmented reality in vocational training: A systematic review of research and applications. Comput. Hum. Behav. 2022, 129, 107125. [Google Scholar] [CrossRef]
  5. Yeung, A.W.K.; Tosevska, A.; Klager, E.; Eibensteiner, F.; Laxar, D.; Stoyanov, J.; Glisic, M.; Zeiner, S.; Kulnik, S.T.; Crutzen, R.; et al. Virtual and augmented reality applications in medicine: Analysis of the scientific literature. J. Med. Internet Res. 2021, 23, e25499. [Google Scholar] [CrossRef]
  6. Javaid, M.; Haleem, A. Virtual reality applications toward medical field. Clin. Epidemiol. Glob. Health 2020, 8, 600–605. [Google Scholar] [CrossRef]
  7. Suh, I.; McKinney, T.; Siu, K.C. Current perspective of metaverse application in medical education, research and patient care. Virtual Worlds 2023, 2, 115–128. [Google Scholar] [CrossRef]
  8. Boutros, F.; Damer, N.; Raja, K.; Ramachandra, R.; Kirchbuchner, F.; Kuijper, A. Iris and periocular biometrics for head mounted displays: Segmentation, recognition, and synthetic data generation. Image Vis. Comput. 2020, 104, 104007. [Google Scholar] [CrossRef]
  9. Boutros, F.; Damer, N.; Raja, K.; Ramachandra, R.; Kirchbuchner, F.; Kuijper, A. On benchmarking iris recognition within a head-mounted display for ar/vr applications. In Proceedings of the IEEE International Joint Conference on Biometrics (IJCB), Houston, TX, USA, 28 September–1 October 2020; pp. 1–10. [Google Scholar]
  10. Arena, F.; Collotta, M.; Pau, G.; Termine, F. An overview of augmented reality. Computers 2022, 11, 28. [Google Scholar] [CrossRef]
  11. Scavarelli, A.; Arya, A.; Teather, R.J. Virtual reality and augmented reality in social learning spaces: A literature review. Virtual Real. 2021, 25, 257–277. [Google Scholar] [CrossRef]
  12. Zhang, C.; Chen, J.; Li, J.; Peng, Y.; Mao, Z. Large language models for human–robot interaction: A review. Biomim. Intell. Robot. 2023, 3, 100131. [Google Scholar] [CrossRef]
  13. Pfeuffer, K.; Geiger, M.J.; Prange, S.; Mecke, L.; Buschek, D.; Alt, F. Behavioural biometrics in vr: Identifying people from body motion and relations in virtual reality. In Proceedings of the CHI Conference on Human Factors in Computing Systems, Glasgow, UK, 4–9 May 2019; pp. 1–12. [Google Scholar]
  14. Morales, A.; Fierrez, J.; Galbally, J.; Gomez-Barrero, M. Introduction to iris presentation attack detection. In Handbook of Biometric Anti-Spoofing: Presentation Attack Detection; Springer: Berlin/Heidelberg, Germany, 2019; pp. 135–150. [Google Scholar]
  15. Czajka, A.; Bowyer, K.W. Presentation attack detection for iris recognition: An assessment of the state-of-the-art. ACM Comput. Surv. (CSUR) 2018, 51, 1–35. [Google Scholar] [CrossRef]
  16. Wang, Z.; Zhao, Y.; Liu, Y.; Lu, F. Edge-guided near-eye image analysis for head mounted displays. In Proceedings of the IEEE International Symposium on Mixed and Augmented Reality (ISMAR), Bari, Italy, 4–8 October 2021; pp. 11–20. [Google Scholar]
  17. Clay, V.; König, P.; Koenig, S. Eye tracking in virtual reality. J. Eye Mov. Res. 2019, 12, 1–18; [Google Scholar] [CrossRef]
  18. Kapp, S.; Barz, M.; Mukhametov, S.; Sonntag, D.; Kuhn, J. ARETT: Augmented reality eye tracking toolkit for head mounted displays. Sensors 2021, 21, 2234. [Google Scholar] [CrossRef]
  19. Bozkir, E.; Özdel, S.; Wang, M.; David-John, B.; Gao, H.; Butler, K.; Jain, E.; Kasneci, E. Eye-tracked virtual reality: A comprehensive survey on methods and privacy challenges. arXiv 2023, arXiv:2305.14080. [Google Scholar]
  20. Idiap Research Institute. VRBiom Dataset. 2024. Available online: https://www.idiap.ch/dataset/vrbiom (accessed on 31 March 2025).
  21. Chinese Academy of Sciences. CASIA Iris Image Database. 2002. Available online: http://biometrics.idealtest.org (accessed on 31 March 2025).
  22. Zanlorensi, L.A.; Laroca, R.; Luz, E.; Britto, A.S., Jr.; Oliveira, L.S.; Menotti, D. Ocular recognition databases and competitions: A survey. Artif. Intell. Rev. 2022, 55, 129–180. [Google Scholar] [CrossRef]
  23. Omelina, L.; Goga, J.; Pavlovicova, J.; Oravec, M.; Jansen, B. A survey of iris datasets. Image Vis. Comput. 2021, 108, 104109. [Google Scholar] [CrossRef]
  24. Yambay, D.; Becker, B.; Kohli, N.; Yadav, D.; Czajka, A.; Bowyer, K.W.; Schuckers, S.; Singh, R.; Vatsa, M.; Noore, A.; et al. LivDet iris 2017–Iris liveness detection competition 2017. In Proceedings of the IEEE International Joint Conference on Biometrics (IJCB), Denver, CO, USA, 1–4 October 2017; pp. 733–741. [Google Scholar]
  25. Tinsley, P.; Purnapatra, S.; Mitcheff, M.; Boyd, A.; Crum, C.; Bowyer, K.; Flynn, P.; Schuckers, S.; Czajka, A.; Fang, M.; et al. Iris Liveness Detection Competition (LivDet-Iris)–The 2023 Edition. In Proceedings of the IEEE International Joint Conference on Biometrics (IJCB), Ljubljana, Slovenia, 25–28 September 2023; pp. 1–10. [Google Scholar]
  26. Proença, H.; Alexandre, L.A. UBIRIS: A noisy iris image database. In Proceedings of the International Conference on Image Analysis and Processing (ICIAP), Cagliari, Italy, 6–8 September 2005; Springer: Berlin/Heidelberg, Germany, 2005; pp. 970–977. [Google Scholar]
  27. Proença, H.; Filipe, S.; Santos, R.; Oliveira, J.; Alexandre, L.A. The UBIRIS. v2: A database of visible wavelength iris images captured on-the-move and at-a-distance. IEEE Trans. Pattern Anal. Mach. Intell. 2009, 32, 1529–1535. [Google Scholar] [CrossRef]
  28. Teo, C.C.; Neo, H.F.; Michael, G.; Tee, C.; Sim, K. A robust iris segmentation with fuzzy supports. In Proceedings of the International Conference on Neural Information Processing (ICONIP), Sydney, Australia, 21–25 November 2010; Springer: Berlin/Heidelberg, Germany, 2010; pp. 532–539. [Google Scholar]
  29. Nguyen, K.; Proença, H.; Alonso-Fernandez, F. Deep learning for iris recognition: A survey. ACM Comput. Surv. 2024, 56, 1–35. [Google Scholar] [CrossRef]
  30. McMurrough, C.D.; Metsis, V.; Kosmopoulos, D.; Maglogiannis, I.; Makedon, F. A dataset for point of gaze detection using head poses and eye images. J. Multimodal User Interfaces 2013, 7, 207–215. [Google Scholar] [CrossRef]
  31. Tonsen, M.; Zhang, X.; Sugano, Y.; Bulling, A. Labelled pupils in the wild: A dataset for studying pupil detection in unconstrained environments. In Proceedings of the Ninth Biennial ACM Symposium on Eye Tracking Research & Applications, Charleston, SC, USA, 14–17 March 2016; pp. 139–142. [Google Scholar]
  32. Kim, J.; Stengel, M.; Majercik, A.; De Mello, S.; Dunn, D.; Laine, S.; McGuire, M.; Luebke, D. Nvgaze: An anatomically-informed dataset for low-latency, near-eye gaze estimation. In Proceedings of the CHI Conference on Human Factors in Computing Systems, Glasgow, UK, 4–9 May 2019; pp. 1–12. [Google Scholar]
  33. Garbin, S.J.; Shen, Y.; Schuetz, I.; Cavin, R.; Hughes, G.; Talathi, S.S. Openeds: Open eye dataset. arXiv 2019, arXiv:1905.03702. [Google Scholar]
  34. Palmero, C.; Sharma, A.; Behrendt, K.; Krishnakumar, K.; Komogortsev, O.V.; Talathi, S.S. Openeds2020: Open eyes dataset. arXiv 2020, arXiv:2005.03876. [Google Scholar]
  35. Meta Platforms, Inc. Compare Devices. 2025. Available online: https://developers.meta.com/horizon/resources/compare-devices/ (accessed on 31 March 2025).
  36. Chang, E.; Kim, H.T.; Yoo, B. Virtual reality sickness: A review of causes and measurements. Int. J. Hum.- Interact. 2020, 36, 1658–1682. [Google Scholar] [CrossRef]
  37. Chen, Y.; Wang, X.; Xu, H. Human factors/ergonomics evaluation for virtual reality headsets: A review. CCF Trans. Pervasive Comput. Interact. 2021, 3, 99–111. [Google Scholar] [CrossRef]
  38. Laganà, F.; Bibbò, L.; Calcagno, S.; De Carlo, D.; Pullano, S.A.; Pratticò, D.; Angiulli, G. Smart Electronic Device-Based Monitoring of SAR and Temperature Variations in Indoor Human Tissue Interaction. Appl. Sci. 2025, 15, 2439. [Google Scholar] [CrossRef]
  39. Vrellis, I.; Delimitros, M.; Chalki, P.; Gaintatzis, P.; Bellou, I.; Mikropoulos, T.A. Seeing the unseen: User experience and technology acceptance in Augmented Reality science literacy. In Proceedings of the 2020 IEEE 20th International Conference on Advanced Learning Technologies (ICALT), Tartu, Estonia, 6–9 July 2020; pp. 333–337. [Google Scholar]
  40. Bowyer, K.W.; Burge, M.J. Handbook of Iris Recognition; Springer: Berlin/Heidelberg, Germany, 2016. [Google Scholar]
  41. Yin, Y.; He, S.; Zhang, R.; Chang, H.; Han, X.; Zhang, J. Deep learning for iris recognition: A review. arXiv 2023, arXiv:2303.08514. [Google Scholar] [CrossRef]
  42. Alonso-Fernandez, F.; Bigun, J. A survey on periocular biometrics research. Pattern Recognit. Lett. 2016, 82, 92–105. [Google Scholar] [CrossRef]
  43. Zhang, M.; Liu, J.; Sun, Z.; Tan, T. The first ICB* competition on iris recognition. In Proceedings of the IEEE International Joint Conference on Biometrics, Clearwater, FL, USA, 29 September–2 October 2014; pp. 1–6. [Google Scholar]
  44. Zhang, M.; Zhang, Q.; Sun, Z.; Zhou, S.; Ahmed, N.U. The btas competition on mobile iris recognition. In Proceedings of the 8th International Conference on Biometrics Theory, Applications and Systems (BTAS), Niagara Falls, NY, USA, 6–9 September 2016; pp. 1–7. [Google Scholar]
  45. Boutros, F.; Damer, N.; Raja, K.; Ramachandra, R.; Kirchbuchner, F.; Kuijper, A. Periocular Biometrics in Head-Mounted Displays: A Sample Selection Approach for Better Recognition. In Proceedings of the International Workshop on Biometrics and Forensics (IWBF), Porto, Portugal, 29–30 April 2020; pp. 1–6. [Google Scholar]
  46. Boyd, A.; Fang, Z.; Czajka, A.; Bowyer, K.W. Iris presentation attack detection: Where are we now? Pattern Recognit. Lett. 2020, 138, 483–489. [Google Scholar] [CrossRef]
  47. Raja, K.B.; Raghavendra, R.; Braun, J.N.; Busch, C. Presentation attack detection using a generalizable statistical approach for periocular and iris systems. In Proceedings of the International Conference of the Biometrics Special Interest Group (BIOSIG), Darmstadt, Germany, 21–23 September 2016; pp. 1–6. [Google Scholar]
  48. Hoffman, S.; Sharma, R.; Ross, A. Iris + Ocular: Generalized iris presentation attack detection using multiple convolutional neural networks. In Proceedings of the International Conference on Biometrics (ICB), Crete, Greece, 4–7 June 2019; pp. 1–8. [Google Scholar]
  49. Liveness Detection Competitions. LivDet. 2025. Available online: https://livdet.org (accessed on 31 March 2025).
  50. Adegoke, B.; Omidiora, E.; Falohun, S.; Ojo, J. Iris Segmentation: A survey. Int. J. Mod. Eng. Res. (IJMER) 2013, 3, 1885–1889. [Google Scholar]
  51. Rot, P.; Vitek, M.; Grm, K.; Emeršič, Ž.; Peer, P.; Štruc, V. Deep sclera segmentation and recognition. In Handbook of Vascular Biometrics; Springer: Basel, Switzerland, 2020; pp. 395–432. [Google Scholar]
  52. Chaudhary, A.K.; Kothari, R.; Acharya, M.; Dangi, S.; Nair, N.; Bailey, R.; Kanan, C.; Diaz, G.; Pelz, J.B. Ritnet: Real-time semantic segmentation of the eye for gaze tracking. In Proceedings of the IEEE/CVF International Conference on Computer Vision Workshop (ICCVW), Seoul, Republic of Korea, 27–28 October 2019; pp. 3698–3702. [Google Scholar]
  53. Perry, J.; Fernandez, A. Minenet: A dilated cnn for semantic segmentation of eye features. In Proceedings of the IEEE/CVF International Conference on Computer Vision Workshops, Seoul, Republic of Korea, 27–28 October 2019; pp. 3671–3676. [Google Scholar]
  54. Valenzuela, A.; Arellano, C.; Tapia, J.E. Towards an efficient segmentation algorithm for near-infrared eyes images. IEEE Access 2020, 8, 171598–171607. [Google Scholar] [CrossRef]
  55. Wang, C.; Wang, W.; Zhang, K.; Muhammad, J.; Lu, T.; Zhang, Q.; Tian, Q.; He, Z.; Sun, Z.; Zhang, Y.; et al. NIR iris challenge evaluation in non-cooperative environments: Segmentation and localization. In Proceedings of the 2021 IEEE International Joint Conference on Biometrics (IJCB), Shenzhen, China, 4–7 August 2021; pp. 1–10. [Google Scholar]
  56. Das, A.; Atreya, S.; Mukherjee, A.; Vitek, M.; Li, H.; Wang, C.; Zhao, G.; Boutros, F.; Siebke, P.; Kolf, J.N.; et al. Sclera Segmentation and Joint Recognition Benchmarking Competition: SSRBC 2023. In Proceedings of the IEEE International Joint Conference on Biometrics (IJCB), Ljubljana, Slovenia, 25–28 September 2023; pp. 1–10. [Google Scholar]
Figure 1. Examples of biometric use-cases offered by HMD data, in particular by the VRBiom dataset.
Figure 1. Examples of biometric use-cases offered by HMD data, in particular by the VRBiom dataset.
Electronics 14 01835 g001
Figure 2. Setup for the dataset collections: (a) for bona fide recordings, the subjects wore the HMD devices; (b) for PA recordings, the HMD device was carefully placed on the temple region of the attack instrument (mannequin, in this example); and (c) the Meta Quest Pro Data Capture App used for data collection.
Figure 2. Setup for the dataset collections: (a) for bona fide recordings, the subjects wore the HMD devices; (b) for PA recordings, the HMD device was carefully placed on the temple region of the attack instrument (mannequin, in this example); and (c) the Meta Quest Pro Data Capture App used for data collection.
Electronics 14 01835 g002
Figure 3. Samples of bona fide recordings from VRBiom dataset.
Figure 3. Samples of bona fide recordings from VRBiom dataset.
Electronics 14 01835 g003
Table 1. Overview of iris and periocular biometric datasets.
Table 1. Overview of iris and periocular biometric datasets.
Datasets and Subsets#Subjects#ImagesResolutionTypes of AttacksCollection YearInstitution(s)
CASIA-Iris series [21]CASIA-IrisV1108 *756 320 × 280 No2002Chinese Academy of Sciences
CASIA-IrisV2120 **2400 640 × 480 No2004
CASIA-IrisV370022,034VariousNo2005
CASIA-IrisV4280054,601VariousNo2010
LivDetIris 2017 [24]ND CLD 2015-7300 640 × 480 contact lens (live)2015University of Notre Dame
IIITD-WVU-7459 640 × 480 Print2017IIIT Delhi
Clarkson508095 640 × 480 Print and contact lens (live)2017Clarkson University
Warsaw45712,013 640 × 480 Print and contact lens (print, live)2017Warsaw University
UBIRIS [26,27]UBIRIS-V12411877 400 × 300 No2004University of Beira
UBIRIS-V226111,102 800 × 600 No2009
MMU [28,29]MMU-V145450-No2010Multimedia University
MMU-V2100995-No2010
LivDet-Iris 2023 [25]--13,332 ***VariousPrint, contact lens, fake/prosthetic eyes, and synthetic iris2023Multiple Institutions
* Number of classes, ** number of eyes, *** number of test data.
Table 2. Summary of the iris/periocular datasets acquired by the HMDs.
Table 2. Summary of the iris/periocular datasets acquired by the HMDs.
DatasetsCollection Year#Subjects#ImagesResolutionFPS (Hz)Presentation AttacksProfileSynthetic Eye Data
Point of Gaze (PoG) [30]201220 768 × 480 30Frontal
LPW [31]201622130,856 640 × 480 95Frontal
NVGaze [32]2019352,500,000 640 × 480 30Frontal
OpenEDS2019 [33]2019152356,649 640 × 400 200Frontal
OpenEDS2020 [34]202080579,900 640 × 400 100Frontal
VRBiom (this work)2024251,262,520 400 × 400 72Non-Frontal
Table 3. Technical specifications of Meta Quest Pro.
Table 3. Technical specifications of Meta Quest Pro.
FeatureSpecification
Display TypeLCD with local dimming
Pixels Per Degree (PPD)22 PPD
Refresh RateUp to 90 Hz
ProcessorQualcomm Snapdragon XR2 + Gen 1
RAM12 GB LPDDR5
Tracking CapabilitiesHead, hand, face, and eye tracking
Field of View106° horizontal × 90° vertical
Wi-FiWi-Fi 6E capable
Battery Life≈2.5 h
Table 4. Summary of bona fide recordings in the VRBiom dataset.
Table 4. Summary of bona fide recordings in the VRBiom dataset.
#Subjects#Gaze#Glass#Repetitions#Eyes#Total
253232900
Table 5. Details of bona fide and different types of PAs from the VRBiom dataset. Each video was recorded at 72 FPS for approximately 10 s.
Table 5. Details of bona fide and different types of PAs from the VRBiom dataset. Each video was recorded at 72 FPS for approximately 10 s.
TypePA SeriesSubtype# Identities# VideosAttack Types
bona fide[steady gaze, moving gaze, partially closed] × [glass, no glass]25900
Presentation Attacks2Mannequins784Own eyes (same material)
3Custom Rigid Mask (Type I)10120Own eyes (same material)
4Custom Rigid Mask (Type II)14168Fake 3D eyeballs
5Generic Flexible Masks20240Print attacks (synthetic data)
6Custom Silicone Masks16192Fake 3D eyeballs
7Print Attacks25300Print attacks (real data)
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Kotwal, K.; Ulucan, I.; Özbulak, G.; Selliah, J.; Marcel, S. VRBiom: A New Periocular Dataset for Biometric Applications of Head-Mounted Display. Electronics 2025, 14, 1835. https://doi.org/10.3390/electronics14091835

AMA Style

Kotwal K, Ulucan I, Özbulak G, Selliah J, Marcel S. VRBiom: A New Periocular Dataset for Biometric Applications of Head-Mounted Display. Electronics. 2025; 14(9):1835. https://doi.org/10.3390/electronics14091835

Chicago/Turabian Style

Kotwal, Ketan, Ibrahim Ulucan, Gökhan Özbulak, Janani Selliah, and Sébastien Marcel. 2025. "VRBiom: A New Periocular Dataset for Biometric Applications of Head-Mounted Display" Electronics 14, no. 9: 1835. https://doi.org/10.3390/electronics14091835

APA Style

Kotwal, K., Ulucan, I., Özbulak, G., Selliah, J., & Marcel, S. (2025). VRBiom: A New Periocular Dataset for Biometric Applications of Head-Mounted Display. Electronics, 14(9), 1835. https://doi.org/10.3390/electronics14091835

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop