Multimodal MRI–HSI Synthetic Brain Tissue Dataset Based on Agar Phantoms
Abstract
1. Summary
- Extended registration methodology. Each phantom design provides at least one complete pipeline for registering all imaging modalities. In Phantom v1, registration relies exclusively on the tracking system, using camera calibrations and fiducial landmarks to align the MRI, HSI, RGB-D data, and the 3D model. In Phantom v2, the workflow is extended by enabling direct MRI-to-tracking registration through points sampled on the phantom and by adding ArUco markers that allow direct registration between the HSI camera and the 3D model. These complementary pathways reduce cumulative error and allow benchmarking of different multimodal registration strategies.
- Diversity of the data. The phantoms incorporate several agar concentrations, each fully characterized by MRI relaxation times and hyperspectral signatures. This variability mimics inter-patient differences in tissue composition and optical response. Furthermore, the arrangement of tissue layers is intentionally alternated between phantoms (normal and inverted configurations) to prevent learning-based methods from relying solely on fixed structural patterns, thus encouraging more robust segmentation and classification models.
- Usability across research domains. Beyond the core MRI and HSI modalities, the dataset includes multi-camera RGB and depth acquisitions from different sensor technologies (stereo, time-of-flight), extending its applicability to several research directions, including depth estimation from stereo image pairs; MRI segmentation benchmarking with HSI-based validation; standalone HSI classification tasks; multimodal fusion experiments; registration between MRI volumes and 3D container phantom models using surface geometry; and evaluation of tracking-to-camera calibration procedures. This makes the dataset a versatile benchmark for a wide range of multimodal imaging and registration challenges, not limited to MRI–HSI fusion.
2. Data Description
Dataset Organization
- 3D_models: 3D models used to build the phantom.
- Depth: Data captured by the RGB and depth cameras used for each phantom. Each camera has its own subfolder with the following structure:
- –
- captures: Includes the images acquired by the camera in .bin format. Depending on the camera type, the files may include color.bin, depth.bin, left.bin, right.bin (for stereo cameras), and color.bin, depth.bin, ir.bin (for time-of-flight cameras).
- *
- captures_metadata.json: Provides interpretability information about each image, including shape, data size, and number of channels.
- *
- optitrack_metadata.json: Contains tracking system information for the camera, including the last quaternion (q), position (t), rotation–translation matrix (RT), and a buffer with the last 15 RT matrices (RT_buffer).
- –
- cam_params.json: Contains the optical parameters of each sensor, including the intrinsic matrix (k), distortion coefficients (dist), homography matrices relating sensors (H), rotation matrix (R), and translation vector (t) composing H, as well as the resolution at which the calibration was performed using the DLR CalLab framework (German Aerospace Center) [17].
- –
- optitrack_calibration.json: Calibration matrix (E) relating the RGB camera optical center with the tracking system, and matrix G, which relates the chessboard with the tracking system. These matrices are required for the calibration process detailed in [18].
- HSI: Contains the hyperspectral acquisitions. To ensure robust tissue characterization, each scene was captured with eight different exposure times.
- –
- captures: Includes the raw hyperspectral images in .bin format with the naming convention raw_[t_exp]_ms.bin, where exposure times (t_exp) are 20, 40, 70, 90, 100, 120, 150, and 200 ms. This folder also includes captures_metadata.json and optitrack_metadata.json.
- –
- wb: Contains the white-balance images of a homogeneous surface, used to compute reflectance. Files follow the same naming convention as the captures.
- –
- hypercubes: Contains the reflectance hypercubes obtained after preprocessing the raw data [19]. Files follow the format hypercube_[t_exp]_ms.bin. This folder also contains the corresponding ground-truth labels in labelled_cube.png.
- –
- cam_params.json and optitrack_calibration.json: Provide the same information as in the Depth folder.
- MRI: MRI slices for different modalities and orientations.
- –
- T1w, T2w and PD:
- *
- Axial: Axial MRI slices in DICOM format.
- *
- Coronal: Coronal MRI slices in DICOM format.
- –
- labels: Contains the ground-truth for the labeled slices.
- Register: Contains information from the digitizing probe of the tracking system used to record the fiducial markers of the phantom.
- –
- optitrack_points_[n].json: Stores the positions of the landmarks embedded in the phantom 3D models. The index n corresponds to repeated registrations of the same markers.
- –
- optitrack_points_mri.json: Contains the positions of points manually touched on the phantom surface to register the MRI directly with the tracking system. This file is only available in the Phantom v2 dataset.
3. Methods
3.1. Phantom Design
3.2. Phantom Creation
3.3. Data Acquisition
3.4. Data Labeling
3.5. Data Validation
4. Usage Notes
4.1. Code Availability
- For the depth data, each camera is represented as a Camera object that stores intrinsic and extrinsic parameters (the parsed information of the cam_params.json file), tracking information (position, rotation, and calibration), and the captured frames parsed from .bin files to NumPy arrays.
- The HSI data follow a similar structure, with the HSI object containing camera parameters, tracking information, and hyperspectral captures as NumPy arrays.
- MRI data are organized by orientation (axial and coronal) and by sequence (T1, T2, and PD), with each slice stored as a NumPy array. The slices are reconstructed into volumes using the DICOM metadata, including voxel spacing and origin.
- The 3D models are loaded from .stl or .ply files, enabling direct use in analysis or visualization. Finally, the register_points attribute provides access to the fiducial landmark positions collected using the tracking probe on the phantom containers or directly on the MRI volume.
4.2. Usage Example
4.3. Potential Research Applications
- Evaluation of registration methodologies, including clinically inspired approaches using soft landmarks (e.g., patient face points), rigid fiducial landmarks (e.g., bone-based markers) and image-based approaches (e.g., skin-attached ArUco markers).
- Benchmarking data fusion strategies, for example, integrating HSI spectral information with MRI structural context to improve detection or classification.
- Investigation of non-rigid and deformable multimodal registration methods beyond the rigid baselines reported in this work, with performance quantitatively assessed using the provided annotations through segmentation overlap (e.g., DICE) and boundary-based metrics (e.g., ASD).
- Assessment of depth-sensing cameras in confined or small-incision scenarios, using the different depth acquisition systems incorporated in this study.
- Monte Carlo simulations of tissue optical properties, leveraging MRI volumetric information and HSI spectral data.
- Spectral unmixing and analysis, based on the known dyes and agar concentrations used to create the phantom tissues.
- Analysis of tracking system performance and accuracy, comparing different setups across the two dataset versions.
- Characterization of the optical properties of MRI phantom components other than agar using HSI data, following the construction and registration methodologies described in this work.
Author Contributions
Funding
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- Wołk, K.; Wołk, A. Hyperspectral Imaging System Applications in Healthcare. Electronics 2025, 14, 4575. [Google Scholar] [CrossRef]
- Fabelo, H.; Ortega, S.; Szolna, A.; Bulters, D.; Piñeiro, J.F.; Kabwama, S.; J-O’Shanahan, A.; Bulstrode, H.; Bisshopp, S.; Kiran, B.R.; et al. In-Vivo Hyperspectral Human Brain Image Database for Brain Cancer Detection. IEEE Access 2019, 7, 39098–39116. [Google Scholar] [CrossRef]
- Martín-Pérez, A.; Villa, M.; Rosa Olmeda, G.; Sancho, J.; Vazquez, G.; Urbanos, G.; Martinez de Ternero, A.; Chavarrías, M.; Jimenez-Roldan, L.; Perez-Nuñez, A.; et al. SLIMBRAIN database: A multimodal image database of in vivo human brains for tumour detection. Sci. Data 2025, 12, 836. [Google Scholar] [CrossRef] [PubMed]
- Ndu, H.; Sheikh-Akbari, A.; Deng, J.; Mporas, I. HyperVein: A Hyperspectral Image Dataset for Human Vein Detection. Sensors 2024, 24, 1118. [Google Scholar] [CrossRef] [PubMed]
- Lu, G.; Fei, B. Medical hyperspectral imaging: A review. J. Biomed. Opt. 2014, 19, 010901. [Google Scholar] [CrossRef] [PubMed]
- Studier-Fischer, A.; Seidlitz, S.; Sellner, J.; Bressan, M.; Özdemir, B.; Ayala, L.; Odenthal, J.; Knoedler, S.; Kowalewski, K.F.; Haney, C.M.; et al. HeiPorSPECTRAL - the Heidelberg Porcine HyperSPECTRAL Imaging Dataset of 20 Physiological Organs. Sci. Data 2023, 10, 414. [Google Scholar] [CrossRef] [PubMed]
- Yusuff, H.; Chatelin, S.; Dillenseger, J.P. Narrative review of tissue-mimicking materials for MRI phantoms: Composition, fabrication, and relaxation properties. Radiography 2024, 30, 1655–1668. [Google Scholar] [CrossRef] [PubMed]
- Antoniou, A.; Georgiou, L.; Christodoulou, T.; Panayiotou, N.; Ioannides, C.; Zamboglou, N.; Damianou, C. MR relaxation times of agar-based tissue-mimicking phantoms. J. Appl. Clin. Med. Phys. 2022, 23, e13533. [Google Scholar] [CrossRef] [PubMed]
- Chen, Z.J.; Gillies, G.T.; Broaddus, W.C.; Prabhu, S.S.; Fillmore, H.; Mitchell, R.M.; Corwin, F.D.; Fatouros, P.P. A realistic brain tissue phantom for intraparenchymal infusion studies. J. Neurosurg. 2004, 101, 314–322. [Google Scholar] [CrossRef] [PubMed]
- Ntombela, L.; Adeleye, B.; Chetty, N. Low-cost fabrication of optical tissue phantoms for use in biomedical imaging. Heliyon 2020, 6, e03602. [Google Scholar] [CrossRef] [PubMed]
- Crasto, N.; Kirubarajan, A.; Sussman, D. Anthropomorphic brain phantoms for use in MRI systems: A systematic review. Magn. Reson. Mater. Phys. Biol. Med. 2022, 35, 277–289. [Google Scholar] [CrossRef] [PubMed]
- Deng, L.; Wu, T.; Wu, F.; Xiong, L.; Yang, H.; Chen, Q.; Liao, Y. Anthropomorphic Head MRI Phantoms: Technical Development, Brain Imaging Applications, and Future Prospects. J. Magn. Reson. Imaging 2025, 62, 1579–1602. [Google Scholar] [CrossRef] [PubMed]
- de Graaf, R.A.; Brown, P.B.; McIntyre, S.; Nixon, T.W.; Behar, K.L.; Rothman, D.L. High magnetic field water and metabolite proton T1 and T2 relaxation in rat brain in vivo. Magn. Reson. Med. 2006, 56, 386–394. [Google Scholar] [CrossRef] [PubMed]
- Ximea. IMEC SM Range 600–1000 USB3 Hyperspectral Camera. 2025. Available online: https://www.ximea.com/products/hyperspectral-imaging/xispec-hyperspectral-miniature-cameras/imec-sm-range-600-1000-usb3-hyperspectral-camera (accessed on 29 November 2025).
- MR-Solutions. MRS*PET/MR 9.4T Cryogen-Free Preclinical PET/MR Scanner. 2025. Available online: https://www.mrsolutions.com/molecular-imaging/molecular-imaging/pet-mr-cryogen-free/petmr-94t/ (accessed on 29 November 2025).
- NaturalPoint. Optitrack System. 2025. Available online: https://optitrack.com/ (accessed on 29 November 2025).
- Strobl, K.H.; Hirzinger, G. More accurate pinhole camera calibration with imperfect planar target. In Proceedings of the 2011 IEEE International Conference on Computer Vision Workshops (ICCV Workshops), Barcelona, Spain, 6–13 November 2011; pp. 1068–1075. [Google Scholar] [CrossRef]
- Villa, M.; Sancho, J.; Rosa-Olmeda, G.; Chavarrias, M.; Juarez, E.; Sanz, C. Benchmarking commercial depth sensors for intraoperative markerless registration in neurosurgery applications. Int. J. Comput. Assist. Radiol. Surg. 2025, 20, 1759–1769. [Google Scholar] [CrossRef]
- Villa, M.; Sancho, J.; Vazquez, G.; Rosa, G.; Urbanos, G.; Martin-Perez, A.; Sutradhar, P.; Salvador, R.; Chavarrías, M.; Lagares, A.; et al. Data-Type Assessment for Real-Time Hyperspectral Classification in Medical Imaging. In Proceedings of the Design and Architecture for Signal and Image Processing; Desnos, K., Pertuz, S., Eds.; Springer: Cham, Switzerland, 2022; pp. 123–135. [Google Scholar]
- Tkachenko, M.; Malyuk, M.; Holmanyuk, A.; Liubimov, N. Label Studio: Data Labeling Software, 2020–2025. Open Source Software. Available online: https://github.com/HumanSignal/label-studio (accessed on 29 November 2025).
- Kirillov, A.; Mintun, E.; Ravi, N.; Mao, H.; Rolland, C.; Gustafson, L.; Xiao, T.; Whitehead, S.; Berg, A.C.; Lo, W.Y.; et al. Segment Anything. arXiv 2023, arXiv:2304.02643. [Google Scholar] [PubMed]
- Liu, Z.; Mao, H.; Wu, C.Y.; Feichtenhofer, C.; Darrell, T.; Xie, S. A ConvNet for the 2020s. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022. [Google Scholar]
- Terzakis, G.; Lourakis, M. A Consistently Fast and Globally Optimal Solution to the Perspective-n-Point Problem. In Proceedings of the Computer Vision—ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Proceedings, Part I. Springer: Berlin/Heidelberg, Germany, 2020; pp. 478–494. [Google Scholar] [CrossRef]
- Segal, A.V.; Haehnel, D.; Thrun, S. Generalized-ICP. In Robotics; The MIT Press: Cambridge, MA, USA, 2010; pp. 161–168. [Google Scholar] [CrossRef]












| Device | Modality | Resolution | Images | Total |
|---|---|---|---|---|
| Phantom v1 (five phantoms) | ||||
| RGB/Depth | ||||
| Intel D405 | RGBD | 1280 × 720 | 1 | 10 |
| Intel D435f | RGBD | 1280 × 720 | 1 | 10 |
| Azure Kinect v2 | RGB | 1920 × 1080 | 1 | 10 |
| Depth/IR | 640 × 576 | 1 | 10 | |
| HSI | ||||
| Ximea Snapshot v2 | Raw image | 2048 × 1088 | 8 | 80 |
| Hypercube | 409 × 217 × 24 | 8 | 80 | |
| Ground-truth | 409 × 217 | 1 | 10 | |
| MRI | ||||
| Axial | T1 | 512 × 224, 1 mm | 30 | 300 |
| T2 | 512 × 224, 1 mm | 30 | 300 | |
| DP | 512 × 224, 1 mm | 30 | 300 | |
| Ground-truth | 512 × 512 | - | 109 | |
| Coronal | T1 | 224 × 512, 0.5 mm | 30 | 300 |
| T2 | 224 × 512, 0.5 mm | 30 | 300 | |
| DP | 224 × 512, 0.5 mm | 30 | 300 | |
| Ground-truth | 512 × 512 | - | 44 | |
| Tracking | ||||
| OptiTrack Flex 3 (four cameras) | Camera position and rotation, calibration matrices, and pointer locations on 3D model fiducials. | |||
| Miscelaneous | ||||
| 3D models of the container of each phantom and intrinsics parameters of each camera. | ||||
| Phantom v2 (eight phantoms) | ||||
| RGBD | ||||
| Intel D405 | RGBD | 1280 × 720 | 1 | 8 |
| HSI | ||||
| Ximea Snapshot v2 | Raw image | 2048 × 1088 | 8 | 64 |
| White balance | 2048 × 1088 | 8 | 64 | |
| Hypercube | 409 × 217 × 24 | 8 | 64 | |
| Ground-truth | 409 × 217 | 1 | 8 | |
| MRI | ||||
| Axial | T1 | 512 × 224, 1 mm | 40 | 320 |
| T2 | 512 × 224, 1 mm | 40 | 320 | |
| DP | 512 × 224, 1 mm | 40 | 320 | |
| Ground-truth | 512 × 512 | - | 251 | |
| Coronal | T1 | 224 × 512, 0.5 mm | 30 | 240 |
| T2 | 224 × 512, 0.5 mm | 30 | 240 | |
| DP | 224 × 512, 0.5 mm | 30 | 240 | |
| Ground-truth | 512 × 512 | - | 99 | |
| Tracking | ||||
| OptiTrack Prime 22 (10 cameras) | Camera positions and orientations, calibration matrices, and pointer locations on 3D model fiducials and on random points across the phantom. | |||
| Miscelaneous | ||||
| 3D models of the container of each phantom and intrinsics parameters of each camera. | ||||
| Agar Concentration [%] | T1 Time [ms] | T2 Time [ms] | Phantom v1 | Phantom v2 |
|---|---|---|---|---|
| 2 | 2204.2 | 87.4 | Tumor | Tumor |
| 3 | 2040.0 | 43.1 | - | Gray matter |
| 4 | 1816.4 | 42.4 | Gray matter | White matter |
| 5 | 1709.2 | 37.9 | White matter | - |
| 6 | 1276.7 | 43.0 | Blood vessels | Blood vessels |
| Parameter | T1-Weighted | T2-Weighted | Proton Density |
|---|---|---|---|
| Sequence | FSE27 | FSE27 | FSE27 |
| Repetition time | 750 ms | 7000 ms | 7000 ms |
| Echo time | 20 ms | 80 ms | 20 ms |
| Echo train | 4 | 7 | 7 |
| Echo spacing | 20 | 20 | 20 |
| Bandwidth | 33.33 kHz | 33.33 kHz | 33.33 kHz |
| Averages | 2 | 2 | 2 |
| Tissue ID | MRI Label | HSI Label | Classification | Binary Classification |
|---|---|---|---|---|
| White matter | Blue label | Green label | Healthy class | Healthy class |
| Gray matter | White label | |||
| Blood vessels | - | Blue label | Blood vessels class | |
| Tumor | Red label | Red label | Tumoral class | |
| Background | Gray label | - | - | |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2026 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license.
Share and Cite
Villa, M.; Sancho, J.; Rosa-Olmeda, G.; Enkaoua, A.; Moccia, S.; Juarez, E. Multimodal MRI–HSI Synthetic Brain Tissue Dataset Based on Agar Phantoms. Data 2026, 11, 12. https://doi.org/10.3390/data11010012
Villa M, Sancho J, Rosa-Olmeda G, Enkaoua A, Moccia S, Juarez E. Multimodal MRI–HSI Synthetic Brain Tissue Dataset Based on Agar Phantoms. Data. 2026; 11(1):12. https://doi.org/10.3390/data11010012
Chicago/Turabian StyleVilla, Manuel, Jaime Sancho, Gonzalo Rosa-Olmeda, Aure Enkaoua, Sara Moccia, and Eduardo Juarez. 2026. "Multimodal MRI–HSI Synthetic Brain Tissue Dataset Based on Agar Phantoms" Data 11, no. 1: 12. https://doi.org/10.3390/data11010012
APA StyleVilla, M., Sancho, J., Rosa-Olmeda, G., Enkaoua, A., Moccia, S., & Juarez, E. (2026). Multimodal MRI–HSI Synthetic Brain Tissue Dataset Based on Agar Phantoms. Data, 11(1), 12. https://doi.org/10.3390/data11010012

