Previous Article in Journal
A Novel Protocol for Integrated Assessment of Upper Limbs Using the Optoelectronic Motion Analysis System: Validation and Usability in Healthy People
Previous Article in Special Issue
Hyperspectral Imaging for Enhanced Skin Cancer Classification Using Machine Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Review of Linear-Array-Transducer-Based Volumetric Ultrasound Imaging Techniques and Their Biomedical Applications

by
Ninjbadgar Tsedendamba
1,†,
Yuon Song
1,†,
Eun-Yeong Park
2,* and
Jeesu Kim
1,*
1
Departments of Cogno-Mechatronics Engineering, Pusan National University, Busan 46241, Republic of Korea
2
Department of Bio and Brain Engineering, Korea Advanced Institute of Science and Technology, Daejeon 34141, Republic of Korea
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.
Bioengineering 2025, 12(9), 906; https://doi.org/10.3390/bioengineering12090906 (registering DOI)
Submission received: 23 July 2025 / Revised: 13 August 2025 / Accepted: 20 August 2025 / Published: 23 August 2025

Abstract

Ultrasound imaging is one of the most widespread biomedical imaging techniques thanks to its advantages such as being non-invasive, portable, non-ionizing, and cost-effective. Ultrasound imaging generally provides two-dimensional cross-sectional images, but the quality and interpretative ability vary based on the experience of the examiner, leading to a lack of objectivity and accuracy. To address these issues, there is a growing demand for three-dimensional ultrasound imaging. Among the various types of transducers used to obtain three-dimensional ultrasound images, this paper focuses on the most standardized probe, the linear array transducer, and provides an overview of the system implementations, imaging results, and applications of volumetric ultrasound imaging from the perspective of scanning methods. Through this comprehensive review, future researchers will gain insights into the advantages and disadvantages of various approaches to three-dimensional imaging systems using linear arrays, providing direction and applicability for system configuration and application.

1. Introduction

Ultrasound imaging (USI) is a widely utilized biomedical imaging technique that enables the non-invasive and non-ionizing visualization of internal anatomical structures by achieving pulse-echo signals of ultrasound (US) waves propagating through biological tissues [1,2]. Compared to other conventional clinical imaging techniques such as magnetic resonance imaging (MRI), positron emission tomography (PET), and X-ray computed tomography (CT), USI offers distinct advantages of temporal resolution and accessibility for routine monitoring [3,4,5]. By visualizing the structural information of various organs, USI is extensively employed across a broad range of clinical applications, ranging from diagnostic assessments to image-guided interventions [6,7,8,9].
Conventional USI primarily produces two-dimensional (2D) B-mode images by using various types of one-dimensional (1D) array transducers [10,11,12]. These cross-sectional images offer real-time morphology of biological tissues with hand-held operation of the examiner. While highly effective in many clinical scenarios, 2D imaging constrains diagnostic interpretation and reproducibility, especially when anatomical structures extend beyond the imaging plane or exhibit a complex three-dimensional (3D) morphology. These limitations have motivated the development of 3D volumetric USI [13,14,15], which provides more comprehensive and intuitive anatomical information by capturing entire volumes rather than single slices [16,17,18].
For 3D USI, various geometries of US transducers have been utilized [19,20]. The most straightforward approach involves mechanically scanning a single-element transducer, which remains attractive due to its simplicity and low cost. However, it suffers from slow acquisition and motion artifacts, which thus restrict its clinical applicability [21,22,23]. Alternately, 2D matrix array transducers enable real-time volumetric imaging through full electronic steering in both azimuthal and elevational dimensions [24,25,26], and have found applications in fields such as obstetric imaging [27,28,29]. However, their widespread adoption is limited by technological complexity, reduced frame rates, and high fabrication costs. Ultrasound computed tomography (USCT) represents another paradigm, reconstructing 3D images using inverse algorithms from multi-angle transmission data [30]. USCT offers full-angle volumetric reconstructions and quantitative imaging potential, particularly for breast imaging, but typically requires extensive hardware infrastructure and prolonged acquisition, limiting its real-time applicability.
Among these approaches, linear-array-based volumetric USI has emerged as a highly practical and adaptable solution. Linear array transducers (LATs) are already ubiquitous in clinical settings and benefit from mature fabrication processes, standardized interfaces, and broad system compatibility. LATs are composed of numerous small piezoelectric elements (typically 128–256) arranged linearly (Figure 1). The acoustic beamlines generated by LATs produce rectangular-shaped B-mode images with uniform beam density. This configuration offers an excellent near-field resolution, particularly advantageous for imaging superficial anatomical structures. LAT-based USI conventionally generates cross-sectional ( x × z ) B-mode images. To reconstruct 3D volumes, additional scanning in the elevational ( y ) direction is required. By integrating advanced scanning strategies, LATs can be repurposed for 3D image acquisition with relatively modest hardware modifications.
Compared to single-element approaches, LATs offer much faster acquisition and superior image quality thanks to simultaneous multi-line reception and dynamic focusing. Compared to 2D matrix arrays, LAT-based systems require an order of magnitude fewer channels, allowing for simpler electronics and higher native frame rates. While matrix arrays can capture a full volume instantaneously without mechanical motion, they are constrained by reduced temporal resolution from channel multiplexing and higher manufacturing complexity. In contrast, LAT-based volumetric imaging offers a cost-effective and clinically accessible solution, achieving comparable in-plane resolution and adequate volumetric coverage for many applications. These attributes make LAT-based volumetric USI a compelling platform for both translational research and widespread clinical adoption.
In general, the imaging depth and spatial resolution in USI are highly dependent on the acoustic frequency. Acoustic attenuation in biological tissue increases approximately linearly with frequency and can be expressed as follows [31,32]:
α f α 0 f n ,
where α f is the acoustic attenuation coefficient at frequency f . α 0 and n 1 are attenuation constants related to the tissue. The intensity of a US wave I z decreases exponentially from the initial intensity I 0 , according to propagation distance z as follows:
I z = I 0 e a f z .
Higher frequencies reduce the effective imaging depth due to increased attenuation, but their shorter wavelength enables finer axial and lateral resolution. Conversely, lower frequencies allow for deeper penetration by sacrificing imaging resolution. Typical LATs operate in the frequency range of 1–20 MHz, which provides a suitable balance for clinical human imaging, with a penetration depth of several centimeters and spatial resolutions on the order of a few hundred micrometers.
Moreover, LAT platforms exhibit exceptional expandability toward functional and molecular imaging modalities. They can be readily adapted to support photoacoustic imaging (PAI), which leverages light absorption-induced acoustic signals to provide functional and molecular information [33,34,35,36,37]. Based on the similarity in signal detection the and image generation processes of PAI and USI, the two modalities are often configured in a single imaging platform [38,39]. Particularly, when integrated with LAT-based systems, PAI enables simultaneous structural and molecular imaging, offering valuable insights into tissue physiology, angiogenesis, and tumor microenvironments [40,41], showing potential for visualizing the functional information of biological tissues based on the multispectral molecular absorption characteristics [42,43].
Likewise, ultrafast US Doppler (UFD) imaging, which can visualize blood flow dynamics with an exceptionally high temporal resolution, is also readily implemented on LAT platforms by leveraging high-frame-rate plane wave imaging [44,45,46]. Unlike conventional Doppler imaging, which relies on sequential line-by-line scanning and suffers from limited frame rates, UFD achieves frame rates in the order of thousands of frames per second by utilizing plane wave transmission with diverged angles, enabling fast image acquisition across the entire imaging field of view [47]. This enables the detection of slow-flow vessels, microcirculations, and transient hemodynamic changes with high temporal and spatial resolution. When applied in conjunction with LAT-based volumetric acquisition, UFD allows for 3D functional imaging of perfusion, vascular reactivity, and cardiac mechanics.
This review presents a comprehensive overview of volumetric USI techniques based on LATs. The system configurations, volumetric scanning strategies, representative imaging results, and biomedical applications are explored. This review specifically investigates three representative scanning mechanisms for achieving volumetric data acquisition with LATs: linear translational scanning, robot-arm-assisted scanning, and freehand scanning. By summarizing the imaging results, the strengths and limitations of each approach can be explored. This article aims to guide researchers and clinicians toward a deeper understanding of the current state and future directions of LAT-based 3D USI, highlighting its potential for both clinical and preclinical imaging applications.

2. Linear Translational Scanning

Linear translational scanning is one of the most fundamental and widely adopted methods for acquiring volumetric data using LATs [48,49]. In this approach, the LAT is mechanically translated along the elevational ( y ) axis, orthogonal to the imaging plane ( x × z ), to sequentially acquire a series of 2D slices across a predefined cuboidal volume (Figure 2a). Because the scanning trajectory and speed are predefined and controllable, the spatial position of each cross-sectional image can be accurately determined, making the image generation procedure straightforward. The key advantages of linear translational scanning are simplicity and reproducibility. The motorized translation ensures consistent probe orientation throughout the acquisition, minimizing geometric distortion and operator-dependent variability. Furthermore, the method is relatively easy to implement, requiring minimal modification of existing 2D imaging systems. Therefore, this approach has been widely utilized as an initial strategy for volumetric data acquisition, especially for functional imaging modalities, including PAI and UFD (Table 1).
Table 1. Summary of three-dimensional US imaging systems with the linear translational scanning of LATs. US, ultrasound; PA, photoacoustic; UFD, ultrafast US Doppler; LAT, linear array transducer; f C , center frequency of LAT; f B W , bandwidth of LAT; and n E , number of elements.
Table 1. Summary of three-dimensional US imaging systems with the linear translational scanning of LATs. US, ultrasound; PA, photoacoustic; UFD, ultrafast US Doppler; LAT, linear array transducer; f C , center frequency of LAT; f B W , bandwidth of LAT; and n E , number of elements.
ScanningTransducerImaging ModeApplicationRef.
Volume [mm3]Time [s] f C  [MHz] f B W  [MHz] n E
40 × 75 × 30 *30 *8.53–12128US, PARat gastrointestinal tract,
human forearm
[50]
40 × 50 × 24 *25 *8.53–12128US, PAMouse gastrointestinal tract[51]
114 × 160 × 30 *2118.53–12128US, PAHuman foot[52]
38.4 × 40 × 30208.53–12128US, PAHuman neck, wrist, thigh[53]
38.1 × 31.4 × 25 *11.48.53–12128US, PAHuman melanoma[54]
12.8 × 12.8 × 20120015128US, UFDMouse brain[55]
24 × 12.8 × 158401814–22128US, UFDRat kidney[56]
6 × 6 × 61.51814–22128US, UFDMouse brain[57]
100 × 100 × 120 *75 *7.53–10192USHuman residual limbs[58]
128 × 180 × 180 *3.52–4.75128USBreast phantom[59]
128 × 180 × 1801803.52–4.75128USBreast phantom[60]
100 × 100 × 2509.454–7128USHuman hand, wrist, forearm[61]
* Calculated from values provided in the literature. – not reported in the original publication.
Figure 2. Representative 3D US imaging results using the linear translational scanning mechanism. (a) Schematic illustration of the motorized translational scanning for 3D US imaging. (b) Volumetric PA and US overlaid image of the rat abdomen area. Color represents the depth of the PA signals. (c) PA image of human forearm. (d) Photograph of the adaptive scanning module to achieve 3D data of peripheral vasculatures. (e) Structure and blood vessel network acquired from the human foot. (f) Schematic illustration of a 3D UFD imaging system using a linear translation stage. (g) 3D structure of a rat kidney renal vascular network, reconstructed from 3D UFD. 3D, three-dimensional; US, ultrasound; PA, photoacoustic; UFD, ultrafast ultrasound Doppler; TR, transducer; FB, fiber bundle; GP, gel pad; DAQ, data acquisition model; and ILAV, interlobular artery–vein. The images are adapted with permission from [50,52,56].
Figure 2. Representative 3D US imaging results using the linear translational scanning mechanism. (a) Schematic illustration of the motorized translational scanning for 3D US imaging. (b) Volumetric PA and US overlaid image of the rat abdomen area. Color represents the depth of the PA signals. (c) PA image of human forearm. (d) Photograph of the adaptive scanning module to achieve 3D data of peripheral vasculatures. (e) Structure and blood vessel network acquired from the human foot. (f) Schematic illustration of a 3D UFD imaging system using a linear translation stage. (g) 3D structure of a rat kidney renal vascular network, reconstructed from 3D UFD. 3D, three-dimensional; US, ultrasound; PA, photoacoustic; UFD, ultrafast ultrasound Doppler; TR, transducer; FB, fiber bundle; GP, gel pad; DAQ, data acquisition model; and ILAV, interlobular artery–vein. The images are adapted with permission from [50,52,56].
Bioengineering 12 00906 g002
Kim et al. demonstrated a dual-modal PA and US imaging platform by integrating a programmable US machine with a transportable laser source [50]. The system employed a 128-element LAT with a center frequency of 7.5 MHz and a laser operating at a pulse repetition frequency of 10 Hz. During translational scanning, both PA and US data were acquired. Overlaid PA/US images from rat abdomens (Figure 2b) and the human forearm (Figure 2c) successfully visualized both molecular absorption contrast and structural features, demonstrating the potential of the system for clinical and preclinical applications. In a following study, the same platform was further applied to investigate the biodistribution of exogenous contrast agents in small animals, achieving deep tissue imaging at depths of up to 4.6 cm, with theoretical axial resolutions of 0.23 mm and 0.14 mm for PA and US, respectively [51]. The resulting images clearly delineated contrast-enhanced signals from internal organs such as the bladder and gastrointestinal tract.
Choi et al. proposed a translational scanning method that follows the contours of the skin surface using a two-phase adaptive scanning strategy (Figure 2d) [52]. Volumetric PA/US data were acquired from the feet of six healthy volunteers, with each foot scanned in two stages. The first phase rapidly identified the skin surface, while the second phase adaptively adjusted the axial position of the probe for maintaining a consistent probe-to-skin distance during scanning. The total acquisition time per volumetric dataset was approximately 197 s, demonstrating feasibility for clinical use. The repeatability of this method was validated through five repeated scans, which consistently visualized internal structures and multilayered vascular networks. Furthermore, the system enabled the quantitative assessment of vascular morphology and hemoglobin oxygenation, underscoring its potential for evaluating peripheral vascular diseases (Figure 2e).
Despite these advancements, the bulkiness and mechanical complexity of traditional linear translation stages remain a barrier to broader clinical adoption. To address this limitation, Lee et al. developed a hand-held linear translation scanner to achieve volumetric data [53]. This compact imaging device, measuring 100 × 80 × 100 mm3 and weighing 950 g, integrated a 128-element LAT and a miniaturized motorized scanner based on a Scotch yoke mechanism. By achieving an approximately 0.2 mm axial resolution at a depth of 30 mm, the platform was successfully applied to the in vivo imaging of blood vessels in the neck, wrist, and thigh. Its clinical utility was further demonstrated through the non-invasive visualization of cutaneous melanoma lesions in human subjects in vivo [54]. The system achieved depth-resolved PA and US images, enabling the measurement of the depth of lesions with a mean absolute error of 0.36 mm. The measured depth was validated by comparing it to histopathological measurements, supporting its accuracy for tumor thickness assessment.
Linear translational scanning has also been applied to UFD imaging for 3D visualization of the vasculature network. Demené et al. employed a multi-axis scanning system composed of a 3-axis linear stage and a 1-axis yaw stage to acquire volumetric data [55]. By using a 15 MHz LAT that transmitted eight compounded plane waves at a frame rate of 800 Hz, their system was able to achieve UFD images at depths of up to 20 mm. A spatiotemporal singular value decomposition (SVD) filter was applied to suppress tissue signals and enhance microvascular contrast. The approach was validated through whole-brain imaging in rats, successfully revealing cerebral microvasculature and blood flow dynamics relevant to neurovascular studies. However, the relatively long (~20 min for 12.8 mm range) scanning time limits its potential feasibility in biomedical applications. To improve the data acquisition speed, Oh et al. introduced a high-frame-rate system operating at 1 kHz, utilizing seven compounded plane waves to generate volumetric vascular images (Figure 2f) [56]. Volumetric imaging was performed using an LAT with a center frequency of 18 MHz. They implemented accelerated image processing, enabling UFD data acquisition and image generation in 7 s for a single B-mode frame. This reduced the total acquisition time to ~14 min for a 24 mm elevational range with a step size of 0.2 mm (Figure 2g). Despite this improvement, the system remained insufficient for real-time imaging. To further accelerate 3D imaging, Generowicz et al. demonstrated a sweep motion scanning approach for UFD imaging of the mouse brain [57]. Their system transmitted eight compounded plane waves at a frame rate of 4 kHz while performing a continuous back-and-forth sweep across a 6 mm region in the elevational direction. This setup enabled volumetric imaging at a rate of one volume every 1.5 s, significantly enhancing temporal resolution compared to conventional stepwise scanning methods.
While linear translational scanning has demonstrated promising results for 3D volume imaging, it is inherently limited in tomographic image generation due to its unidirectional acquisition geometry. To overcome this constraint, rotational scanning approaches have been introduced to achieve more complete angular sampling and improve spatial resolution. Ranger et al. demonstrated a volumetric US tomography system that utilizes the rotational scanning of a LAT [58]. To achieve full volumetric coverage, the system applied additional linear vertical translation, enabling the acquisition of a stack of 2D slices over a cylindrical imaging volume. A camera-based motion tracking system was integrated to monitor the position and orientation of the probe in real time, allowing for the motion-compensated alignment of the acquired slices. They successfully acquired volumetric US images from the residual limb of human subjects. The accuracy of the resulting 3D ultrasound volumes was validated by showing a good correlation with corresponding MRI scans. Liu et al. also demonstrated the rotational scanning of LATs [59]. Their system is configured with four 128-element LATs arranged at 90° intervals around the target (Figure 3a). This setup reduced the required scanning rotation range to 90°, significantly minimizing the acquisition time. Each pair of opposing transducers enabled the simultaneous acquisition of reflected and transmitted ultrasound signals. This geometry enhanced spatial resolution and contrast by capturing multi-angle perspectives. In a subsequent study, the same group demonstrated the multi-perspective imaging of breast phantoms containing multiple masses of varying sizes [60]. By reconstructing both vertical and horizontal cross-sections, the system successfully visualized the volumetric structure of the phantoms (Figure 3b).
Figure 3. Representative 3D US imaging results using multiple LATs. (a) Schematic illustration of a 3D US imaging system using rotational scanning of four LATs. (b) The resulting 3D volume of a phantom that mimics a human breast. (c) Schematic of a 3D US imaging system using linear scanning of eight LATs. (d) 3D imaging results of the human forearm area, demonstrating volumetric visualization of skin surface, bone, and blood vessels. 3D, three-dimensional; US, ultrasound; and LAT, linear array transducer. The images are adapted with permission from [59,60,61].
Figure 3. Representative 3D US imaging results using multiple LATs. (a) Schematic illustration of a 3D US imaging system using rotational scanning of four LATs. (b) The resulting 3D volume of a phantom that mimics a human breast. (c) Schematic of a 3D US imaging system using linear scanning of eight LATs. (d) 3D imaging results of the human forearm area, demonstrating volumetric visualization of skin surface, bone, and blood vessels. 3D, three-dimensional; US, ultrasound; and LAT, linear array transducer. The images are adapted with permission from [59,60,61].
Bioengineering 12 00906 g003
To further improve imaging speed and eliminate the need for mechanical rotation, Park et al. proposed an octagonal array configuration composed of eight LATs arranged to form a 1024-element circular aperture with a diameter of 162 mm [61]. This geometry enabled full-angle signal acquisition without rotating the probe, providing an isotropic volumetric spatial resolution of less than 0.05 mm. The entire assembly was mounted on a motorized scanner and vertically translated into a water tank to acquire volumetric data (Figure 3c). To address computational challenges, a partial beamforming strategy was implemented. The local beamforming of data from each LAT was performed prior to transmitting data to a central data processing unit. This significantly reduced data transfer and reconstruction latency, enabling real-time volumetric imaging at 25 Hz, which is suitable for in vivo applications. Using this system, they successfully imaged the human forearm, clearly visualizing skin layers, skeletal structures, and vascular networks (Figure 3d).

3. Robot Arm Scanning

While linear translational scanning has been applied for 3D USI, applying this scanning mechanism in the human body may be limited because the skin surface is not flat. With advances in robotic technology, robot arms are increasingly being integrated into medical procedures and diagnostic imaging. In medical imaging, these systems offer consistent and precise control of probe motion, maintaining stable contact with the skin and following predefined trajectories (Figure 4a). By employing more than six degrees of freedom (DOFs), robot arms can provide reproducible and robust scanning while maintaining close contact on the surface of the curved samples (Table 2).
Table 2. Summary of three-dimensional US imaging systems with robot arm scanning of LATs. US, ultrasound; LAT, linear array transducer; f C , center frequency of LAT; f B W , bandwidth of LAT; and n E , number of elements.
Table 2. Summary of three-dimensional US imaging systems with robot arm scanning of LATs. US, ultrasound; LAT, linear array transducer; f C , center frequency of LAT; f B W , bandwidth of LAT; and n E , number of elements.
ScanningTransducerImaging ModeApplicationRef.
Volume [mm3]Time [s] f C  [MHz] f B W  [MHz] n E
7.54–9128USFetus phantom[62]
10 × 100 × 20105–12192USVascular phantom[63]
105–14128USHuman thyroid lobe[64]
4–12192USThyroid phantom[65]
13USForearm phantom[66]
12256USHuman carotid artery[67]
5–9USHuman thyroid[68]
– not reported in the original publication.
Huang et al. demonstrated a robot-assisted system for 3D USI using a 6-DOF robotic arm, which was manually controlled by a joystick with visual feedback from four external cameras (Figure 4b) [62]. The system was first validated on phantoms and then applied to obtain 3D US images of the human forearm (Figure 4c). The system exhibited high computational efficiency, reconstructing 290 frames with a scanning time of 162 s and a reconstruction time of ~8 s for generating a volume image. While the results confirmed the feasibility of robot-assisted 3D imaging, the reliance on manual operation limited consistency across repeated scans. To address this limitation, Janvier et al. proposed a robotic scanning method capable of learning and replaying predefined probe trajectories [63]. In their system, a 6-DOF robot arm equipped with a 10 MHz LAT performed controlled linear scans using a two-phase process: a manual “teach” phase that records the scanning trajectory during manual movement and a “replay” phase that reproduces the recorded scanning motion. This ensured consistent probe orientation and motion throughout the scan. The method was validated using a vascular-mimicking phantom with two stenotic lesions, and the reconstructed 3D volumes clearly visualized both regions of narrowing with high accuracy and repeatability. However, the reliance on fixed trajectories may restrict adaptability to patient-specific anatomies in clinical practice.
Figure 4. Representative 3D US imaging results using the robot arm mechanism. (a) Schematic illustration of the robot arm scanning for 3D US imaging. (b) Schematic illustration of a 3D US imaging system using a remotely controlled robot arm scanner. (c) The resulting 3D image of the human forearm. (d) Photograph and resulting US B-mode images of the human forearm achieved via a robot arm scanning system. (e) Reconstructed volumetric model of the forearm tissues in static contraction. 1, brachioradialis muscle. 2, extensor carpi radialis longus muscle. 3, extensor carpi radialis brevis muscle. 4, extensor digitorum muscle. 5, extensor digiti minimi muscle. 6, extensor carpi ulnaris muscle. 3D, three-dimensional; US, ultrasound; LAT, linear array transducer; and PC, personal computer. The images are adapted with permission from [62,66].
Figure 4. Representative 3D US imaging results using the robot arm mechanism. (a) Schematic illustration of the robot arm scanning for 3D US imaging. (b) Schematic illustration of a 3D US imaging system using a remotely controlled robot arm scanner. (c) The resulting 3D image of the human forearm. (d) Photograph and resulting US B-mode images of the human forearm achieved via a robot arm scanning system. (e) Reconstructed volumetric model of the forearm tissues in static contraction. 1, brachioradialis muscle. 2, extensor carpi radialis longus muscle. 3, extensor carpi radialis brevis muscle. 4, extensor digitorum muscle. 5, extensor digiti minimi muscle. 6, extensor carpi ulnaris muscle. 3D, three-dimensional; US, ultrasound; LAT, linear array transducer; and PC, personal computer. The images are adapted with permission from [62,66].
Bioengineering 12 00906 g004
Kojcev et al. introduced an automated scanning mechanism based on surface detection [64]. Prior to scanning, a 3D camera captured the skin surface, and the user defined the imaging region. The system then automatically planned and executed a sweeping trajectory, maintaining optimal contact and orientation throughout the scan. During acquisition, both US images and tracking data were recorded and compounded into a volumetric dataset. Validation on four healthy volunteers showed reduced variability in thyroid lobe measurements compared to expert-operated scans, highlighting the improved repeatability. Similarly, Zielke et al. demonstrated a robotic scanning system for thyroid imaging that operated without reliance on external tracking systems [65]. Their path control algorithm allowed the robot to autonomously perform linear sweeps over the thyroid region. Using built-in force sensors, the system maintained consistent contact pressure and orientation, enabling the accurate acquisition of coronal, axial, and 3D views with anatomical details. Kapravchuk et al. also reported a sensor-integrated scanning platform that employed a 6-DOF robotic arm equipped with a force–torque sensor to regulate contact pressure during probe motion [66]. The LAT was guided along predefined trajectories (Figure 4d), minimizing tissue deformation artifacts and enhancing spatial accuracy. The system was tested on forearm-mimicking phantoms with known geometries. The resulting 3D reconstructions showed volume errors below 1% and a spatial resolution of 0.2 mm, confirming the accuracy and reliability of the method (Figure 4e).
More recently, Huang et al. proposed a fully autonomous carotid artery imaging system that eliminates the need for human intervention throughout the scanning process [67]. In the pre-scan phase, the robotic system plans and executes a scanning trajectory based on anatomical landmarks. A U-Net-based segmentation model provided real-time guidance for probe positioning and contact force adjustment. The system achieved an 86.1% success rate in capturing volumetric carotid artery images in human subjects. Zhou et al. proposed a fully autonomous thyroid imaging system using a 7-DOF robotic arm guided by multimodal imaging and deep learning [68]. They added flexibility, allowing for better adaptation to curved surfaces. The scanning process begins with RGB-D imaging to localize the neck region and guide the initial probe placement. A deep-learning-based controller then adapts the scanning trajectory using US feedback, actively avoiding acoustic shadows. This strategy ensures safe and stable contact, providing highly accurate operator-free volumetric thyroid imaging with a tracking accuracy of 98.6%.

4. Freehand Scanning

Freehand scanning stacks B-mode images by tracking the position and angles of the probe, allowing the user to freely set a scan path (Figure 5a). Freehand scanning provides maximum DOF during image acquisition by allowing operators to manually control the scanning direction and area. This flexibility enables the adaptive scanning of anatomical variations. However, the accurate reconstruction of 3D volumes from 2D B-mode images requires precise knowledge of the position and orientation of the imaging probe at each acquisition frame. In freehand scanning, the scanning motion is recorded via various types of tracking modules that are attached to the imaging probe (Table 3).
Table 3. Summary of three-dimensional US imaging systems with the freehand scanning of LATs. US, ultrasound; PA, photoacoustic; UFD, ultrafast US Doppler; LAT, linear array transducer; f C , center frequency of LAT; f B W , bandwidth of LAT; and n E , number of elements.
Table 3. Summary of three-dimensional US imaging systems with the freehand scanning of LATs. US, ultrasound; PA, photoacoustic; UFD, ultrafast US Doppler; LAT, linear array transducer; f C , center frequency of LAT; f B W , bandwidth of LAT; and n E , number of elements.
ScanningTransducerImaging ModeApplicationRef.
Volume [mm3]Time [s] f C  [MHz] f B W  [MHz] n E
350 × 50 × 2009.5 *5–14 *128 *USFetus phantom[69]
20 × 15 × 12015–20105–15 *128 *USHuman Achilles tendon[70]
60 *7.5 *5–12 *128 *USCarotid artery phantom[71]
5.3
7.8
192
168
US, UFDHuman brain[72]
12010 *6–14128 *USHuman spine[73]
8 × 8 × 8 *3106–14 *128US
Elastography
Elasticity phantom[74]
908.5 *5–12 *192 *USFetal phantom[75]
100 × 70 × 6014 *13–15USBreast phantom[76]
5128USHuman forearm[77]
46 × 83 × 5598 *4–12 *USHuman carotid artery[78]
100 × 200 × 500120–240USHuman spine[79]
40 × 100 × 40~23.5 *7.6128US, PAHuman forearm[80]
* Calculated from values provided in the literature. – not reported in the original publication.
In one of the initial studies, Huang et al. proposed a system that utilized a linear encoder along a 1-axis sliding track to record probe positions during manual scanning [69]. As the probe moved along the track, 2D B-mode images were acquired and stacked based on their spatial locations to generate 3D volumes. This method demonstrated feasibility by producing volumetric images of the human forearm. However, the fixed linear trajectory limited the scanning flexibility, reducing its suitability for imaging anatomies with curved or irregular surfaces.
To overcome this limitation, optical tracking systems have been widely adopted. Obst et al. developed a setup using five cameras and four reflective markers affixed to the US probe to track its position and orientation in real time [70]. The feasibility of the system was verified by scanning the Achilles tendon in humans. During manual scanning, the B-mode images and positional data were synchronized at 40 Hz, allowing for the 3D reconstruction of tendon morphology over a scanning range exceeding 120 mm (Figure 5b). Chung et al. further enhanced tracking accuracy using eight high-resolution digital cameras and smaller reflective markers attached to the LAT [71]. Their system achieved high spatial and temporal resolution, capturing B-mode images at a frame rate of 26 Hz with real-time positional tracking. Three-dimensional imaging of carotid artery phantoms showed a reconstruction error below 0.2 mm. Verhoef et al. extended this approach to achieve 3D UFD imaging of cerebral and tumor blood flow [72]. Volumetric data were acquired by scanning a LAT, while plane wave US images were acquired with 10 steering angles. A block-wise SVD and normalized convolution algorithms were applied to reconstruct 3D voxel-based images (Figure 5c). While the optical tracking method successfully demonstrated 3D image generation, accurate motion tracking requires line-of-sight visibility, which may limit usability in a clinical environment.
Electromagnetic tracking (EMT) systems offer an alternative that does not require line-of-sight. Cheung et al. employed EMT-based freehand scanning for spinal imaging [73]. During the freehand scanning, EMT recorded spatial data to map 2D frames into 3D space (Figure 5d). The reconstructed volume enabled curvature measurements of the human spine, showing a high correlation with X-ray-based Cobb angles (R2 = 0.86) (Figure 5e). However, slight underestimation of the curvature was observed, possibly due to tissue deformation. Lee et al. proposed an EMT-based 3D imaging system for quasi-static elastography that estimated tissue stiffness by computing displacements under mechanical compression [74]. In their method, a LAT probe was manually scanned across an elasticity phantom while applying repeated axial compression. EMT recorded the position and orientation of each B-mode frame during the scan, and the displacement and strain were calculated using a correlation-based algorithm. The registered 2D strain maps were reconstructed into a 3D elastogram, demonstrating accurate volumetric stiffness estimation for potential elastography applications.
Figure 5. Representative 3D US imaging results using the freehand mechanism. (a) Schematic illustration of the freehand scanning for 3D US imaging. (b) Stacked B-mode images from freehand scanning and the representative cross-section image of the reconstructed 3D US volume of the human Achilles tendon. (c) 3D UFD imaging of a frontoparietal glioblastoma multiforme and the lateral ventricles. (d) Photograph of a freehand scanning tracked via an EMT module. (e) The achieved B-mode images. The purple lines are segmented spinal cords featured in the B-mode images. (f) Schematic diagram of frame-to-frame estimation utilizing the MENN. (g) The comparison between reconstruction methods: black, reference trajectory tracked by an optical tracking module; red, linear motion; blue, speckle decorrelation; and green, proposed MENN. 3D, three-dimensional; US, ultrasound; LAT, linear array transducer; EMT, electromagnetic tracer; Tx, transmitter; Rx, receiver; IMU, inertial measurement unit; MENN, motion estimation neural network; I , image; and T , transformation matrix between images. The images are adapted with permission from [70,72,73,77].
Figure 5. Representative 3D US imaging results using the freehand mechanism. (a) Schematic illustration of the freehand scanning for 3D US imaging. (b) Stacked B-mode images from freehand scanning and the representative cross-section image of the reconstructed 3D US volume of the human Achilles tendon. (c) 3D UFD imaging of a frontoparietal glioblastoma multiforme and the lateral ventricles. (d) Photograph of a freehand scanning tracked via an EMT module. (e) The achieved B-mode images. The purple lines are segmented spinal cords featured in the B-mode images. (f) Schematic diagram of frame-to-frame estimation utilizing the MENN. (g) The comparison between reconstruction methods: black, reference trajectory tracked by an optical tracking module; red, linear motion; blue, speckle decorrelation; and green, proposed MENN. 3D, three-dimensional; US, ultrasound; LAT, linear array transducer; EMT, electromagnetic tracer; Tx, transmitter; Rx, receiver; IMU, inertial measurement unit; MENN, motion estimation neural network; I , image; and T , transformation matrix between images. The images are adapted with permission from [70,72,73,77].
Bioengineering 12 00906 g005
To simplify the hardware setup, inertial measurement units (IMUs) have also been used. IMUs consist of integrated sensors such as accelerometers and gyroscopes to determine probe orientation. Herickhoff et al. employed an IMU to track orientation during the rotational and tilting motions of an imaging probe fixed to a custom-designed fixture [75]. Quaternion-based rotation data were assigned to each B-mode frame to reconstruct 3D orientation. However, this method lacked position tracking and could not compensate for probe or target displacement. Kim et al. addressed this limitation by integrating a distance sensor with the IMU to measure real-time axial displacement [76]. This allowed for the spatial registration of each image frame and significantly reduced misalignment artifacts. Their method enabled smoother surface rendering and improved anatomical continuity in reconstructed images of a breast phantom.
Beyond hardware-based tracking, software-based motion estimation methods have emerged to reduce system complexity. Prevost et al. proposed a machine-learning-based 3D reconstruction method [77]. A convolutional neural network (CNN) was trained to predict inter-frame displacement vectors between consecutive frames (Figure 5f). The model demonstrated robust performance across phantom and in vivo datasets, generating high-quality volumetric reconstructions under varied scanning conditions (Figure 5g). Similarly, Lyu et al. proposed a hybrid approach integrating optical tracking and software interpolation techniques for freehand scanning [78]. They applied Bezier interpolation and pixel nearest neighbor (PNN) methods to reconstruct images from spatially tracked data. Phantom and human experiments yielded mean spatial errors below 0.56 mm, accurately delineating key anatomical structures such as the thyroid and carotid artery. Additionally, Chen et al. enhanced EMT-based 3D ultrasound reconstruction through software techniques such as principal component analysis and discriminative scale space tracking [79]. They applied their method to spine imaging in adolescent idiopathic scoliosis patients, enabling the automatic detection and quantification of spinal deformities. The system achieved a high detection accuracy of 99.5%, highlighting the effectiveness of software-assisted motion correction in clinical applications. More recently, Lee et al. introduced a deep learning framework using global–local self-attention to improve freehand volumetric scanning [80]. They trained a deep learning model using US B-mode images. The resulting 3D volume showed effective motion tracking with strong generalizability. While the model is yet to be validated in the forearm, this technique can be applied to various regions of human applications. In another recent study, Luo et al. proposed an enhanced motion network that combines sensor-based tracking and image-based reconstruction for freehand 3D US volumes [81]. They integrated measurements from IMU and EM positioning systems with a temporal multibranch structure to refine motion estimation. The proposed approach was validated on human forearm, carotid, and thyroid datasets, showcasing its higher reconstruction accuracy compared to the existing methods.

5. Conclusions

LATs are among the most standardized and widely adopted probes in clinical ultrasound, routinely used to visualize anatomical structures in the breast, thyroid, tendons, blood vessels, and other superficial tissues. Although conventionally limited to 2D imaging, LATs can be extended to 3D imaging through various scanning mechanisms, each offering unique trade-offs in acquisition speed, flexibility, and clinical feasibility.
Among these techniques, motorized linear translation provides a straightforward method for volume reconstruction, enabling the accurate spatial stacking of B-mode images due to well-defined probe positioning. However, its rigid scanning path limits adaptability to curved or irregular anatomical surfaces. Robotic arm-based scanning addresses this limitation by providing programmable and repeatable probe control with high spatial precision, allowing scanning over complex surfaces. However, these advantages come with increased system complexity, higher cost, and reduced portability.
Freehand scanning, when combined with position-tracking systems such as optical, electromagnetic, or inertial sensors, provides the greatest operational flexibility for exploring anatomical regions without mechanical constraints. Optical tracking provides high spatial accuracy but requires a clear line-of-sight. Electromagnetic tracking avoids visibility issues but can be affected by metallic interference and environmental noise. IMU is compact and cost-effective, but is prone to drift motion, often requiring sensor fusion or periodic recalibration. All tracking methods can introduce errors such as misalignment, motion artifacts, or resolution loss in the reconstructed volume.
Recently, artificial intelligence (AI)-based reconstruction techniques have shown considerable promise in overcoming these issues by estimating motion directly from image data, correcting spatial errors, reducing noise, and improving volume continuity. These methods could reduce dependence on complex tracking hardware and enable robust volumetric imaging. To ensure reliability across different anatomies and scanning conditions, large and diverse training datasets are required.
Despite the potential of 3D USI, the clinical integration of 3D USI remains limited. One promising application is in automated breast volume scanning [82,83,84,85,86], where high-channel-count LATs (e.g., 768 elements) have been used in motorized scanning modules to acquire breast volumes for tumor detection, especially in dense breasts where mammography is less effective [87,88,89]. These systems are under active investigation for triaging breast nodules in clinical screening workflows. Similarly, 3D USI has shown utility in neuromuscular diagnostics through quantitative muscle ultrasound, with freehand scanning facilitating user-friendly acquisition for tracking disease progression in clinical practice [90,91,92,93].
A particularly promising direction involves integrating functional imaging capabilities with structural 3D USI. PAI, which offers molecular contrast based on optical absorption properties, has been successfully combined with LAT-based probes to enable dual-modal imaging systems. These integrated platforms have been validated in clinical studies across various applications, including thyroid nodules [94,95,96], breast tumors [97,98,99,100,101,102,103], and skin diseases [104,105,106]. Such systems provide not only volumetric anatomical information, but also molecular and hemodynamic biomarkers that may support improved disease characterization and personalized treatment planning. To date, most implementations of 3D PAI have relied on basic linear translational scanning mechanisms. However, by leveraging advanced scanning strategies, such as robotic-assisted, freehand-tracked, or multi-angle tomographic techniques, as reviewed in this paper, the scope and applicability of 3D PAI could be significantly broadened.
In addition to hardware innovations, advancements in reconstruction algorithms would enable the clinical adoption of 3D USI. Particularly in freehand scanning, accurate estimation of the spatial trajectory is essential for precise volume rendering. However, sensor drift, motion artifacts, and environmental interference often degrade image quality. AI-based methods would offer promising solutions for correcting tracking errors, enhancing image reconstruction, and enabling the real-time or automated interpretation of 3D data [107].
Furthermore, the automated tracking and restoration of continuous anatomical structures, such as blood vessels, will be critical for improving the completeness and diagnostic value of both US and PA volumes. Approaches such as vessel centerline extraction [108,109], volumetric segmentation [110,111], and model-driven reconstruction [112,113] can enforce spatial continuity across sequential frames or viewpoints and mitigate artifacts arising from limited-view geometry or acoustic shadowing.
Recent advances in high-frequency thin-film US sensors have expanded the capabilities of volumetric imaging by providing broader bandwidth and higher sensitivity [114]. These devices, often fabricated using micromachined materials, can operate at frequencies exceeding conventional US transducers, enabling higher spatial resolution and the improved detection of small anatomical features [115,116]. However, the reduced penetration depth inherent to high-frequency operation may constrain their widespread clinical application. Hybrid approaches that combine high-frequency thin-film transducers with conventional lower-frequency transducers could help balance resolution and depth, enabling both detailed superficial imaging and broader anatomical coverage [117,118].
Moving forward, several challenges remain for widespread clinical deployment. These include reducing system complexity and cost, improving user interface and ergonomics, minimizing acquisition time, and standardizing quality assurance protocols. Furthermore, robust clinical validation through large-scale studies will be essential to establish diagnostic accuracy, reproducibility, and clinical utility across diverse applications.
In conclusion, this review provided a comprehensive analysis of 3D USI systems based on LATs, highlighting key developments across scanning mechanisms, volume reconstruction strategies, and clinical applications. The reviewed technologies have demonstrated the feasibility of volumetric anatomical and functional data in both preclinical and clinical settings. With continued advancements in image acquisition, motion tracking, and intelligent reconstruction algorithms, 3D USI holds substantial promise as a next-generation tool for the quantitative and multi-dimensional assessment of biological tissues in a broad range of clinical applications.

Author Contributions

Conceptualization, J.K. and E.-Y.P.; investigation, N.T. and Y.S.; resources, N.T. and Y.S.; writing—original draft preparation, N.T. and Y.S.; writing—review and editing, J.K. and E.-Y.P.; visualization, N.T. and Y.S.; supervision, J.K. and E.-Y.P.; project administration, J.K. and E.-Y.P. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the Korean government: the National Research Foundation of Korea (NRF) grant (RS-2021-NR060086, RS-2024-00427153) funded by the Ministry of Science and ICT; the Korea Health Industry Development Institute (KHIDI) grant (HR20C0026) funded by the Ministry of Health & Welfare.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Aldrich, J.E. Basic Physics of Ultrasound Imaging. Crit. Care Med. 2007, 35, S131–S137. [Google Scholar] [CrossRef] [PubMed]
  2. Moran, C.M.; Thomson, A.J. Preclinical Ultrasound Imaging—A Review of Techniques and Imaging Applications. Front. Phys. 2020, 8, 124. [Google Scholar] [CrossRef]
  3. Prinz, C.; Voigt, J.-U. Diagnostic Accuracy of a Hand-Held Ultrasound Scanner in Routine Patients Referred for Echocardiography. J. Am. Soc. Echocardiogr. 2011, 24, 111–116. [Google Scholar] [CrossRef]
  4. Eurenius, K.; Axelsson, O.; Gällstedt-Fransson, I.; Sjöden, P.O. Perception of Information, Expectations and Experiences among Women and Their Partners Attending a Second-Trimester Routine Ultrasound Scan. Ultrasound Obstet. Gynecol. 1997, 9, 86–90. [Google Scholar] [CrossRef]
  5. Luck, C.A. Value of Routine Ultrasound Scanning at 19 Weeks: A Four Year Study of 8849 Deliveries. Br. Med. J. 1992, 304, 1474–1478. [Google Scholar] [CrossRef]
  6. Bax, J.; Cool, D.; Gardi, L.; Knight, K.; Smith, D.; Montreuil, J.; Sherebrin, S.; Romagnoli, C.; Fenster, A. Mechanically Assisted 3D Ultrasound Guided Prostate Biopsy System. Med. Phys. 2008, 35, 5397–5410. [Google Scholar] [CrossRef]
  7. Ferraioli, G.; Monteiro, L.B.S. Ultrasound-based Techniques for the Diagnosis of Liver Steatosis. World J. Gastroenterol. 2019, 25, 6053. [Google Scholar] [CrossRef]
  8. Zhang, J.; Chen, Z.; Anil, G. Ultrasound-Guided Thyroid Nodule Ciopsy: Outcomes and Correlation with Imaging Features. Clin. Imaging 2015, 39, 200–206. [Google Scholar] [CrossRef] [PubMed]
  9. Van Velthoven, V.; Auer, L. Practical Application of Intraoperative Ultrasound Imaging. Acta Neurochir. 1990, 105, 5–13. [Google Scholar] [CrossRef]
  10. Shung, K.K.; Zippuro, M. Ultrasonic Transducers and Arrays. IEEE Eng. Med. Biol. 1996, 15, 20–30. [Google Scholar] [CrossRef]
  11. Lee, W.; Roh, Y. Ultrasonic Transducers for Medical Diagnostic Imaging. Biomed. Eng. Lett. 2017, 7, 91–97. [Google Scholar] [CrossRef] [PubMed]
  12. Johnson, J.; Oralkan, Ö.; Demirci, U.; Ergun, S.; Karaman, M.; Khuri-Yakub, P. Medical Imaging using Capacitive Micromachined Ultrasonic Transducer Arrays. Ultrasonics 2002, 40, 471–476. [Google Scholar] [CrossRef]
  13. Fenster, A.; Downey, D.B. Three-Dimensional Ultrasound Imaging. Annu. Rev. Biomed. Eng. 2000, 2, 457–475. [Google Scholar] [CrossRef] [PubMed]
  14. Arya, S.; Kupesic Plavsic, S. Preimplantation 3D Ultrasound: Current Uses and Challenges. J. Perinat. Med. 2017, 45, 745–758. [Google Scholar] [CrossRef]
  15. Lee, S.; Pretorius, D.; Asfoor, S.; Hull, A.; Newton, R.; Hollenbach, K.; Nelson, T. Prenatal Three-Dimensional Ultrasound: Perception of Sonographers, Sonologists and Undergraduate Students. Ultrasound Obstet. Gynecol. 2007, 30, 77–80. [Google Scholar] [CrossRef]
  16. Merz, E.; Pashaj, S. Advantages of 3D Ultrasound in the Assessment of Fetal Abnormalities. J. Perinat. Med. 2017, 45, 643–650. [Google Scholar] [CrossRef]
  17. Chavignon, A.; Heiles, B.; Hingot, V.; Orset, C.; Vivien, D.; Couture, O. 3D Transcranial Ultrasound Localization Microscopy in the Rat Brain with a Multiplexed Matrix Probe. IEEE Trans. Biomed. Eng. 2021, 69, 2132–2142. [Google Scholar] [CrossRef]
  18. Iommi, D.; Hummel, J.; Figl, M.L. Evaluation of 3D Ultrasound for Image Guidance. PLoS ONE 2020, 15, e0229441. [Google Scholar] [CrossRef] [PubMed]
  19. Huang, Q.; Zeng, Z. A Review on Real-time 3D Ultrasound Imaging Technology. BioMed Res. Int. 2017, 2017, 6027029. [Google Scholar] [CrossRef]
  20. Fenster, A.; Downey, D.B. 3-D Ultrasound Imaging: A Review. IEEE Eng. Med. Biol. 1996, 15, 41–51. [Google Scholar] [CrossRef]
  21. Park, E.-Y.; Park, S.; Lee, H.; Kang, M.; Kim, C.; Kim, J. Simultaneous Dual-Modal Multispectral Photoacoustic and Ultrasound Macroscopy for Three-Dimensional Whole-Body Imaging of Small Animals. Photonics 2021, 8, 13. [Google Scholar] [CrossRef]
  22. He, L.; Wang, B.; Wen, Z.; Li, X.; Wu, D. 3-D High Frequency Ultrasound Imaging by Piezo-Driving a Single-Element Transducer. IEEE Trans. Ultrason. Ferroelectr. Freq. Control 2022, 69, 1932–1942. [Google Scholar] [CrossRef]
  23. Lee, H.; Han, S.; Park, S.; Cho, S.; Yoo, J.; Kim, C.; Kim, J. Ultrasound-Guided Breath-Compensation in Single-Element Photoacoustic Imaging for Three-Dimensional Whole-Body Images of Mice. Front. Phys. 2022, 10, 457. [Google Scholar] [CrossRef]
  24. Smith, S.; Trahey, G.; Von Ramm, O. Two-Dimensional Arrays for Medical Ultrasound. Ultrason. Imaging 1992, 14, 213–233. [Google Scholar] [CrossRef]
  25. Roh, Y. Ultrasonic Transducers for Medical Volumetric Imaging. Jpn. J. Appl. Phys. 2014, 53, 07KA01. [Google Scholar] [CrossRef]
  26. Chen, J.; Cheng, X.; Shen, I.-M.; Liu, J.-H.; Li, P.-C.; Wang, M. A Monolithic Three-Dimensional Ultrasonic Transducer Array for Medical Imaging. J. Microelectromech. Syst. 2007, 16, 1015–1024. [Google Scholar] [CrossRef]
  27. Leung, K.-Y. Applications of Advanced Ultrasound Technology in Obstetrics. Diagnostics 2021, 11, 1217. [Google Scholar] [CrossRef]
  28. Tonni, G.; Martins, W.P.; Guimarães Filho, H.; Júnior, E.A. Role of 3-D Ultrasound in Clinical Obstetric Practice: Evolution over 20 Years. Ultrasound Med. Biol. 2015, 41, 1180–1211. [Google Scholar] [CrossRef]
  29. Steiner, H.; Staudach, A.; Spitzer, D.; Schaffer, H. Diagnostic Techniques: Three-Dimensional Ultrasound in Obstetrics and Gynaecology: Technique, Possibilities and Limitations. Hum. Reprod. 1994, 9, 1773–1778. [Google Scholar] [CrossRef]
  30. Bernard, S.; Monteiller, V.; Komatitsch, D.; Lasaygues, P. Ultrasonic Computed Tomography based on Full-Waveform Inversion for Bone Quantitative Imaging. Phys. Med. Biol. 2017, 62, 7011. [Google Scholar] [CrossRef] [PubMed]
  31. Goss, S.A.; Johnston, R.L.; Dunn, F. Comprehensive Compilation of Empirical Ultrasonic Properties of Mammalian Tissues. J. Acoust. Soc. Am. 1978, 64, 423–457. [Google Scholar] [CrossRef] [PubMed]
  32. Parker, K.J.; Lerner, R.M.; Waag, R.C. Attenuation of Ultrasound: Magnitude and Frequency Dependence for Tissue Characterization. Radiology 1984, 153, 785–788. [Google Scholar] [CrossRef]
  33. Park, E.-Y.; Lee, H.; Han, S.; Kim, C.; Kim, J. Photoacoustic Imaging Systems Based on Clinical Ultrasound Platform. Exp. Biol. Med. 2022, 247, 551–560. [Google Scholar] [CrossRef]
  34. Zafar, H.; Breathnach, A.; Subhash, H.M.; Leahy, M.J. Linear-Array-Based Photoacoustic Imaging of Human Microcirculation with a Range of High Frequency Transducer Probes. J. Biomed. Opt. 2015, 20, 051021. [Google Scholar] [CrossRef]
  35. Gateau, J.; Caballero, M.Á.A.; Dima, A.; Ntziachristos, V. Three-Dimensional Optoacoustic Tomography using a Conventional Ultrasound Linear Detector Array: Whole-Body Tomographic System for Small Animals. Med. Phys. 2013, 40, 013302. [Google Scholar] [CrossRef]
  36. Paul, S.; Lee, S.A.; Zhao, S.; Chen, Y.-S. Model-Informed Deep-Learning Photoacoustic Reconstruction for Low-Element Linear Array. Photoacoustics 2025, 44, 100732. [Google Scholar] [CrossRef] [PubMed]
  37. Yang, M.; Qu, Z.; Amjadian, M.; Tang, X.; Chen, J.; Wang, L. All-Fiber Three-Wavelength Laser for Functional Photoacoustic Microscopy. Photoacoustics 2025, 42, 100703. [Google Scholar] [CrossRef]
  38. Lee, H.; Choi, W.; Kim, C.; Park, B.; Kim, J. Review on Ultrasound-Guided Photoacoustic Imaging for Complementary Analyses of Biological Systems in Vivo. Exp. Biol. Med. 2023, 248, 762–774. [Google Scholar] [CrossRef]
  39. Huang, Z.; Liu, D.; Mo, S.; Hong, X.; Xie, J.; Chen, Y.; Liu, L.; Song, D.; Tang, S.; Wu, H.; et al. Multimodal PA/US Imaging in Rheumatoid Arthritis: Enhanced Correlation with Clinical Scores. Photoacoustics 2024, 38, 100615. [Google Scholar] [CrossRef] [PubMed]
  40. Menezes, G.L.; Pijnappel, R.M.; Meeuwis, C.; Bisschops, R.; Veltman, J.; Lavin, P.T.; Van De Vijver, M.J.; Mann, R.M. Downgrading of Breast Masses Suspicious for Cancer by using Optoacoustic Breast Imaging. Radiology 2018, 288, 355–365. [Google Scholar] [CrossRef]
  41. Zhou, J.; Ou, M.; Yuan, B.; Yan, B.; Wang, X.; Qiao, S.; Huang, Y.; Feng, L.; Huang, L.; Luo, Y. Dual-Modality Ultrasound/Photoacoustic Tomography for Mapping Tissue Oxygen Saturation Distribution in Intestinal Strangulation. Photoacoustics 2025, 43, 100721. [Google Scholar] [CrossRef]
  42. Bak, S.; Park, S.M.; Song, Y.; Kim, J.; Nam, T.W.; Han, D.-W.; Kim, C.-S.; Cho, S.-W.; Bouma, B.E.; Lee, H. High Spectral Energy Density All-Fiber Nanosecond Pulsed 1.7 μm Light Source for Photoacoustic Microscopy. Photoacoustics 2025, 44, 100744. [Google Scholar]
  43. Tan, Y.; Zhang, M.; Wu, Z.; Chen, J.; Ren, Y.; Liu, C.; Gu, Y. Local Oxygen Concentration Reversal from Hyperoxia to Hypoxia Monitored by Optical-Resolution Photoacoustic Microscopy in Inflammation-Resolution Process. Photoacoustics 2025, 44, 100730. [Google Scholar] [CrossRef]
  44. Bercoff, J.; Montaldo, G.; Loupas, T.; Savery, D.; Mézière, F.; Fink, M.; Tanter, M. Ultrafast Compound Doppler imaging: Providing full Blood Flow Characterization. IEEE Trans. Ultrason. Ferroelectr. Freq. Control 2011, 58, 134–147. [Google Scholar] [CrossRef]
  45. Tsedendamba, N.; Vial, J.-C.A.; Taylor, R.; Kim, J.; Choi, W. In Vivo Photoacoustic and Ultrafast Ultrasound Doppler Assessment of Vascularity for Potential Thyroid Cancer Diagnosis: A Comprehensive Review. J. Phys. Photonics 2025, 7, 022002. [Google Scholar] [CrossRef]
  46. Chen, Y.; Fang, B.; Meng, F.; Luo, J.; Luo, X. Competitive Swarm Optimized SVD Clutter Filtering for Ultrafast Power Doppler Imaging. IEEE Trans. Ultrason. Ferroelectr. Freq. Control 2024, 71, 459–473. [Google Scholar] [CrossRef] [PubMed]
  47. Osmanski, B.-F.; Pernot, M.; Montaldo, G.; Bel, A.; Messas, E.; Tanter, M. Ultrafast Doppler Imaging of Blood Flow Dynamics in the Myocardium. IEEE Trans. Med. Imaging 2012, 31, 1661–1668. [Google Scholar] [CrossRef] [PubMed]
  48. Karlita, T.; Yuniarno, E.M.; Purnama, I.; Purnomo, M.H. Design and Development of a Mechanical Linear Scanning Device for the Three-Dimensional Ultrasound Imaging System. Int. J. Electr. Eng. Inform. 2019, 11, 272–295. [Google Scholar] [CrossRef]
  49. Chen, C.; Hendriks, G.A.; Fekkes, S.; Mann, R.M.; Menssen, J.; Siebers, C.C.; de Korte, C.L.; Hansen, H.H. In vivo 3D Power Doppler Imaging Using Continuous Translation and Ultrafast Ultrasound. IEEE Trans. Biomed. Eng. 2021, 69, 1042–1051. [Google Scholar] [CrossRef]
  50. Kim, J.; Park, S.; Jung, Y.; Chang, S.; Park, J.; Zhang, Y.; Lovell, J.F.; Kim, C. Programmable Real-time Clinical Photoacoustic and Ultrasound Imaging System. Sci. Rep. 2016, 6, 35137. [Google Scholar] [CrossRef]
  51. Park, S.; Park, G.; Kim, J.; Choi, W.; Jeong, U.; Kim, C. Bi2Se3 Nanoplates for Contrast-Enhanced Photoacoustic Imaging at 1064 nm. Nanoscale 2018, 10, 20548–20558. [Google Scholar] [CrossRef]
  52. Choi, W.; Park, E.-Y.; Jeon, S.; Yang, Y.; Park, B.; Ahn, J.; Cho, S.; Lee, C.; Seo, D.-K.; Cho, J.-H.; et al. Three-Dimensional Multistructural Quantitative Photoacoustic and US Imaging of Human Feet In Vivo. Radiology 2022, 303, 467–473. [Google Scholar] [CrossRef]
  53. Lee, C.; Choi, W.; Kim, J.; Kim, C. Three-Dimensional Clinical Handheld Photoacoustic/Ultrasound Scanner. Photoacoustics 2020, 18, 100173. [Google Scholar] [CrossRef]
  54. Park, B.; Bang, C.H.; Lee, C.; Han, J.H.; Choi, W.; Kim, J.; Park, G.S.; Rhie, J.W.; Lee, J.H.; Kim, C. 3D Wide-Field Multispectral Photoacoustic Imaging of Human Melanomas In Vivo: A Pilot Study. J. Eur. Acad. Dermatol. Venereol. 2020, 35, 669–676. [Google Scholar] [CrossRef] [PubMed]
  55. Demené, C.; Tiran, E.; Sieu, L.-A.; Bergel, A.; Gennisson, J.L.; Pernot, M.; Deffieux, T.; Cohen, I.; Tanter, M. 4D Microvascular Imaging based on Ultrafast Doppler Tomography. NeuroImage 2016, 127, 472–483. [Google Scholar] [CrossRef] [PubMed]
  56. Oh, D.; Lee, D.; Heo, J.; Kweon, J.; Yong, U.; Jang, J.; Ahn, Y.J.; Kim, C. Contrast Agent-Free 3D Renal Ultrafast Doppler Imaging Reveals Vascular Dysfunction in Acute and Diabetic Kidney Diseases. Adv. Sci. 2023, 10, 2303966. [Google Scholar] [CrossRef] [PubMed]
  57. Generowicz, B.S.; Dijkhuizen, S.; Bosman, L.W.; De Zeeuw, C.I.; Koekkoek, S.K.; Kruizinga, P. Swept-3D Ultrasound Imaging of the Mouse Brain Using a Continuously Moving 1D-Array Part II: Functional Imaging. IEEE Trans. Ultrason. Ferroelectr. Freq. Control 2023, 70, 1726–1738. [Google Scholar] [CrossRef]
  58. Ranger, B.J.; Feigin, M.; Zhang, X.; Moerman, K.M.; Herr, H.; Anthony, B.W. 3D Ultrasound Imaging of Residual Limbs with Camera-based Motion Compensation. IEEE Trans. Neural Syst. Rehabil. Eng. 2019, 27, 207–217. [Google Scholar] [CrossRef]
  59. Liu, C.; Xue, C.; Zhang, B.; Zhang, G.; He, C. The Application of an Ultrasound Tomography Algorithm in a Novel Ring 3D Ultrasound Imaging System. Sensors 2018, 18, 1332. [Google Scholar] [CrossRef]
  60. Liu, C.; Zhang, B.; Xue, C.; Zhang, W.; Zhang, G.; Cheng, Y. Multi-Perspective Ultrasound Imaging Technology of the Breast with Cylindrical Motion of Linear Arrays. Appl. Sci. 2019, 9, 419. [Google Scholar] [CrossRef]
  61. Park, E.-Y.; Cai, X.; Foiret, J.; Bendjador, H.; Hyun, D.; Fite, B.Z.; Wodnicki, R.; Dahl, J.J.; Boutin, R.D.; Ferrara, K.W. Fast Volumetric Ultrasound Facilitates High-Resolution 3D Mapping of Tissue Compartments. Sci. Adv. 2023, 9, eadg8176. [Google Scholar] [CrossRef]
  62. Huang, Q.; Lan, J. Remote Control of a Robotic Prosthesis Arm with Six-Degree-of-Freedom for Ultrasonic Scanning and Three-Dimensional Imaging. Biomed. Signal Process. Control 2019, 54, 101606. [Google Scholar] [CrossRef]
  63. Janvier, M.-A.; Durand, L.-G.; Cardinal, M.-H.R.; Renaud, I.; Chayer, B.; Bigras, P.; De Guise, J.; Soulez, G.; Cloutier, G. Performance Evaluation of a Medical Robotic 3D-Ultrasound Imaging System. Med. Image Anal. 2008, 12, 275–290. [Google Scholar] [CrossRef]
  64. Kojcev, R.; Khakzar, A.; Fuerst, B.; Zettinig, O.; Fahkry, C.; DeJong, R.; Richmon, J.; Taylor, R.; Sinibaldi, E.; Navab, N. On the Reproducibility of Expert-Operated and Robotic Ultrasound Acquisitions. Int. J. Comput. Assist. Radiol. Surg. 2017, 12, 1003–1011. [Google Scholar] [CrossRef]
  65. Zielke, J.; Eilers, C.; Busam, B.; Weber, W.; Navab, N.; Wendler, T. RSV: Robotic Sonography for Thyroid Volumetry. IEEE Robot. Autom. Lett. 2022, 7, 3342–3348. [Google Scholar] [CrossRef]
  66. Kapravchuk, V.; Ishkildin, A.; Briko, A.; Borde, A.; Kodenko, M.; Nasibullina, A.; Shchukin, S. Method of Forearm Muscles 3D Modeling Using Robotic Ultrasound Scanning. Sensors 2025, 25, 2298. [Google Scholar] [CrossRef]
  67. Huang, Q.; Gao, B.; Wang, M. Robot-Assisted Autonomous Ultrasound Imaging for Carotid Artery. IEEE Trans. Instrum. Meas. 2024, 73, 1–9. [Google Scholar] [CrossRef]
  68. Zhou, J.; Tian, H.; Wang, W. Fully Automated Thyroid Ultrasound Screening Utilizing Multi-Modality Image and Anatomical Prior. Biomed. Signal Process. Control 2024, 87, 105430. [Google Scholar] [CrossRef]
  69. Huang, Q.-H.; Yang, Z.; Hu, W.; Jin, L.-W.; Wei, G.; Li, X. Linear Tracking for 3-D Medical Ultrasound Imaging. IEEE Trans. Cybern. 2013, 43, 1747–1754. [Google Scholar] [CrossRef]
  70. Obst, S.J.; Newsham-West, R.; Barrett, R.S. In vivo Measurement of Human Achilles Tendon Morphology using Freehand 3-D Ultrasound. Ultrasound Med. Biol. 2014, 40, 62–70. [Google Scholar] [CrossRef]
  71. Chung, S.W.; Shih, C.C.; Huang, C.C. Freehand Three-Dimensional Ultrasound Imaging of Carotid Artery using Motion Tracking Technology. Ultrasonics 2017, 74, 11–20. [Google Scholar] [CrossRef] [PubMed]
  72. Verhoef, L.; Soloukey, S.; Mastik, F.; Generowicz, B.S.; Bos, E.M.; Schouten, J.W.; Koekkoek, S.K.; Vincent, A.J.; Klein, S.; Kruizinga, P. Freehand Ultrafast Doppler Ultrasound Imaging with Optical Tracking Allows for Detailed 3D Reconstruction of Blood Flow in the Human Brain. IEEE Trans. Med. Imaging 2025, 44, 3125–3138. [Google Scholar] [CrossRef]
  73. Cheung, C.-W.J.; Zhou, G.-Q.; Law, S.-Y.; Lai, K.-L.; Jiang, W.-W.; Zheng, Y.-P. Freehand Three-Dimensional Ultrasound System for Assessment of Scoliosis. J. Orthop. Transl. 2015, 3, 123–133. [Google Scholar] [CrossRef]
  74. Lee, F.F.; He, Q.; Luo, J. Electromagnetic Tracking-based Freehand 3D Quasi-Static Elastography with 1D Linear Array: A Phantom Study. Phys. Med. Biol. 2018, 63, 245006. [Google Scholar] [CrossRef] [PubMed]
  75. Herickhoff, C.D.; Morgan, M.R.; Broder, J.S.; Dahl, J.J. Low-Cost Volumetric Ultrasound by Augmentation of 2D Systems: Design and Prototype. Ultrason. Imaging 2017, 40, 35–48. [Google Scholar] [CrossRef]
  76. Kim, T.; Kang, D.H.; Shim, S.; Im, M.; Seo, B.K.; Kim, H.; Lee, B.C. Versatile Low-Cost Volumetric 3D Ultrasound Imaging Using Gimbal-Assisted Distance Sensors and an Inertial Measurement Unit. Sensors 2020, 20, 6613. [Google Scholar] [CrossRef]
  77. Prevost, R.; Salehi, M.; Jagoda, S.; Kumar, N.; Sprung, J.; Ladikos, A.; Bauer, R.; Zettinig, O.; Wein, W. 3D Freehand Ultrasound without External Tracking using Deep Learning. Med. Image Anal. 2018, 48, 187–202. [Google Scholar] [CrossRef]
  78. Lyu, Y.; Shen, Y.; Zhang, M.; Wang, J. Real-Time 3D Ultrasound Imaging System based on a Hybrid Reconstruction Algorithm. Chin. J. Electron. 2024, 33, 245–255. [Google Scholar] [CrossRef]
  79. Chen, H.; Qian, L.; Gao, Y.; Zhao, J.; Tang, Y.; Li, J.; Le, L.H.; Lou, E.; Zheng, R. Development of Automatic Assessment Framework for Spine Deformity using Freehand 3-D Ultrasound Imaging System. IEEE Trans. Ultrason. Ferroelectr. Freq. Control 2024, 71, 408–422. [Google Scholar] [CrossRef] [PubMed]
  80. Lee, S.; Kim, S.; Seo, M.; Park, S.; Imrus, S.; Ashok, K.; Lee, D.; Park, C.; Lee, S.; Kim, J.; et al. Enhancing Free-hand 3D Photoacoustic and Ultrasound Reconstruction using Deep Learning. IEEE Trans. Med. Imaging 2025. [Google Scholar] [CrossRef]
  81. Luo, M.; Yang, X.; Yan, Z.; Cao, Y.; Zhang, Y.; Hu, X.; Wang, J.; Ding, H.; Han, W.; Sun, L.; et al. MoNetV2: Enhanced Motion Network for Freehand 3-D Ultrasound Reconstruction. IEEE Trans. Neural Netw. 2025. [Google Scholar] [CrossRef]
  82. Vourtsis, A. Three-Dimensional Automated Breast Ultrasound: Technical Aspects and First Results. Diagn. Interv. Imaging 2019, 100, 579–592. [Google Scholar] [CrossRef]
  83. Chang, J.M.; Moon, W.K.; Cho, N.; Park, J.S.; Kim, S.J. Radiologists’ Performance in the Detection of Benign and Malignant Masses with 3D Automated Breast Ultrasound (ABUS). Eur. J. Radiol. 2011, 78, 99–103. [Google Scholar] [CrossRef] [PubMed]
  84. Giuliano, V.; Giuliano, C. Improved Breast Cancer Detection in Asymptomatic Women using 3D-Automated Breast Ultrasound in Mammographically Dense Breasts. Clin. Imaging 2013, 37, 480–486. [Google Scholar] [CrossRef] [PubMed]
  85. Wojcinski, S.; Farrokh, A.; Hille, U.; Wiskirchen, J.; Gyapong, S.; Soliman, A.A.; Degenhardt, F.; Hillemanns, P. The Automated Breast Volume Scanner (ABVS): Initial Experiences in Lesion Detection Compared with Conventional Handheld B-mode Ultrasound: A Pilot Study of 50 Cases. Int. J. Women’s Health 2011, 3, 337–346. [Google Scholar] [CrossRef]
  86. Wojcinski, S.; Farrokh, A.; Gyapong, S.; Hille, U.; Hillemanns, P.; Degenhardt, F. The Automated Breast Volume Scanner (ABVS): Initial Experiences in Lesion Detection and Interobserver Concordance of a New Method. Ultrasound Med. Biol. 2011, 37, S46–S47. [Google Scholar] [CrossRef]
  87. Arleo, E.K.; Saleh, M.; Ionescu, D.; Drotman, M.; Min, R.J.; Hentel, K. Recall Rate of Screening Ultrasound with Automated Breast Volumetric Scanning (ABVS) in Women with Dense Breasts: A First Quarter Experience. Clin. Imaging 2014, 38, 439–444. [Google Scholar] [CrossRef]
  88. Mundinger, A. 3D Supine Automated Ultrasound (SAUS, ABUS, ABVS) for Supplemental Screening Women with Dense Breasts. J. Breast Health 2016, 12, 52. [Google Scholar] [CrossRef]
  89. Tutar, B.; Esen Icten, G.; Guldogan, N.; Kara, H.; Arıkan, A.E.; Tutar, O.; Uras, C. Comparison of Automated Versus Hand-Held Breast US in Supplemental Screening in Asymptomatic Women with Dense Breasts: Is There a Difference Regarding Woman Preference, Lesion Detection and Lesion Characterization? Arch. Gynecol. Obstet. 2020, 301, 1257–1265. [Google Scholar] [CrossRef]
  90. De Jong, L.; Nikolaev, A.; Greco, A.; Weijers, G.; de Korte, C.L.; Fütterer, J.J. Three-Dimensional Quantitative Muscle Ultrasound in a Healthy Population. Muscle Nerve 2021, 64, 199–205. [Google Scholar] [CrossRef] [PubMed]
  91. Ashir, A.; Jerban, S.; Barrère, V.; Wu, Y.; Shah, S.B.; Andre, M.P.; Chang, E.Y. Skeletal Muscle Assessment using Quantitative Ultrasound: A Narrative Review. Sensors 2023, 23, 4763. [Google Scholar] [CrossRef]
  92. De Jong, L.; Greco, A.; Nikolaev, A.; Weijers, G.; van Engelen, B.G.; de Korte, C.L.; Fütterer, J.J. Three-Dimensional Quantitative Muscle Ultrasound in Patients with Facioscapulohumeral Dystrophy and Myotonic Dystrophy. Muscle Nerve 2023, 68, 432–438. [Google Scholar] [CrossRef]
  93. MacGillivray, T.J.; Ross, E.; Simpson, H.A.; Greig, C.A. 3D Freehand Ultrasound for in vivo Determination of Human Skeletal Muscle Volume. Ultrasound Med. Biol. 2009, 35, 928–935. [Google Scholar] [CrossRef]
  94. Noltes, M.E.; Bader, M.; Metman, M.J.; Vonk, J.; Steinkamp, P.J.; Kukačka, J.; Westerlaan, H.E.; Dierckx, R.A.; van Hemel, B.M.; Brouwers, A.H.; et al. Towards In Vivo Characterization of Thyroid Nodules Suspicious for Malignancy using Multispectral Optoacoustic Tomography. Eur. J. Nucl. Med. Mol. Imaging 2023, 50, 2736–2750. [Google Scholar] [CrossRef]
  95. Roll, W.; Markwardt, N.A.; Masthoff, M.; Helfen, A.; Claussen, J.; Eisenblätter, M.; Hasenbach, A.; Hermann, S.; Karlas, A.; Wildgruber, M.; et al. Multispectral Optoacoustic Tomography of Benign and Malignant Thyroid Disorders—A Pilot Study. J. Nucl. Med. 2019, 60, 1461–1466. [Google Scholar] [CrossRef]
  96. Kim, J.; Park, B.; Ha, J.; Steinberg, I.; Hooper, S.M.; Jeong, C.; Park, E.-Y.; Choi, W.; Liang, T.; Bae, J.-S.; et al. Multiparametric Photoacoustic Analysis of Human Thyroid Cancers In Vivo. Cancer Res. 2021, 81, 4849–4860. [Google Scholar] [CrossRef]
  97. Schoustra, S.M.; Piras, D.; Huijink, R.; Op’t Root, T.J.; Alink, L.; Kobold, W.M.F.; Steenbergen, W.; Manohar, S. Twente Photoacoustic Mammoscope 2: System Overview and Three-Dimensional Vascular Network Images in Healthy Breasts. J. Biomed. Opt. 2019, 24, 121909. [Google Scholar] [CrossRef]
  98. Nyayapathi, N.; Xia, J. Photoacoustic Imaging of Breast Cancer: A Mini Review of System Design and Image Features. J. Biomed. Opt. 2019, 24, 121911. [Google Scholar] [CrossRef]
  99. Becker, A.; Masthoff, M.; Claussen, J.; Ford, S.J.; Roll, W.; Burg, M.; Barth, P.J.; Heindel, W.; Schaefers, M.; Eisenblaetter, M.; et al. Multispectral Optoacoustic Tomography of the Human Breast: Characterisation of Healthy Tissue and Malignant Lesions using a Hybrid Ultrasound-Optoacoustic Approach. Eur. Radiol. 2018, 28, 602–609. [Google Scholar] [CrossRef]
  100. Neuschler, E.I.; Butler, R.; Young, C.A.; Barke, L.D.; Bertrand, M.L.; Böhm-Vélez, M.; Destounis, S.; Donlan, P.; Grobmyer, S.R.; Katzen, J.; et al. A Pivotal Study of Optoacoustic Imaging to Diagnose Benign and Malignant Breast Masses: A New Evaluation Tool for Radiologists. Radiology 2017, 287, 398–412. [Google Scholar] [CrossRef]
  101. Diot, G.; Metz, S.; Noske, A.; Liapis, E.; Schroeder, B.; Ovsepian, S.V.; Meier, R.; Rummeny, E.; Ntziachristos, V. Multispectral Optoacoustic Tomography (MSOT) of Human Breast Cancer. Clin. Cancer. Res. 2017, 23, 6912–6922. [Google Scholar] [CrossRef]
  102. Riksen, J.J.; Schurink, A.W.; Francis, K.J.; Verhoef, C.; Grünhagen, D.J.; van Soest, G. Dual-Wavelength Photoacoustic Imaging of Sentinel Lymph Nodes in Patients with Melanoma and Breast Cancer. Photoacoustics 2025, 45, 100747. [Google Scholar] [CrossRef]
  103. Han, M.; Lee, Y.J.; Ahn, J.; Nam, S.; Kim, M.; Park, J.; Ahn, J.; Ryu, H.; Seo, Y.; Park, B.; et al. A Clinical Feasibility Study of a Photoacoustic Finder for Sentinel Lymph Node Biopsy in Breast Cancer Patients: A Prospective Cross-Sectional Study. Photoacoustics 2025, 43, 100716. [Google Scholar] [CrossRef]
  104. Zhou, Y.; Tripathi, S.V.; Rosman, I.; Ma, J.; Hai, P.; Linette, G.P.; Council, M.L.; Fields, R.C.; Wang, L.V.; Cornelius, L.A. Noninvasive Determination of Melanoma Depth Using a Handheld Photoacoustic Probe. J. Investig. Dermatol. 2017, 137, 1370–1372. [Google Scholar] [CrossRef]
  105. Attia, A.B.E.; Chuah, S.Y.; Razansky, D.; Ho, C.J.H.; Malempati, P.; Dinish, U.; Bi, R.; Fu, C.Y.; Ford, S.J.; Lee, J.S.-S.; et al. Noninvasive Real-Time Characterization of Non-Melanoma Skin Cancers with Handheld Optoacoustic Probes. Photoacoustics 2017, 7, 20–26. [Google Scholar] [CrossRef]
  106. Kim, M.; Han, J.H.; Ahn, J.; Kim, E.; Bang, C.H.; Kim, C.; Lee, J.H.; Choi, W. In Vivo 3D Photoacoustic and Ultrasound Analysis of Hypopigmented Skin Llesions: A Pilot Study. Photoacoustics 2025, 43, 100705. [Google Scholar] [CrossRef]
  107. Shen, Y.-T.; Chen, L.; Yue, W.-W.; Xu, H.-X. Artificial Intelligence in Ultrasound. Eur. J. Radiol. 2021, 139, 109717. [Google Scholar] [CrossRef]
  108. Guerrero, J.; Salcudean, S.E.; McEwen, J.A.; Masri, B.A.; Nicolaou, S. Real-Time Vessel Segmentation and Tracking for Ultrasound Imaging Applications. IEEE Trans. Med. Imaging 2007, 26, 1079–1090. [Google Scholar] [CrossRef]
  109. Zhao, H.; Wang, G.; Lin, R.; Gong, X.; Song, L.; Li, T.; Wang, W.; Zhang, K.; Qian, X.; Zhang, H.; et al. Three-Dimensional Hessian Matrix-based Quantitative Vascular Imaging of Rat Iris with Optical-Resolution Photoacoustic Microscopy in Vivo. J. Biomed. Opt. 2018, 23, 046006. [Google Scholar] [CrossRef]
  110. Akbari, H.; Fei, B. 3D Ultrasound Image Segmentation using Wavelet Support Vector Machines. Med. Phys. 2012, 39, 2972–2984. [Google Scholar] [CrossRef]
  111. Gu, P.; Lee, W.-M.; Roubidoux, M.A.; Yuan, J.; Wang, X.; Carson, P.L. Automated 3D Ultrasound Image Segmentation to Aid Breast Cancer Image Interpretation. Ultrasonics 2016, 65, 51–58. [Google Scholar] [CrossRef] [PubMed]
  112. Hauptmann, A.; Lucka, F.; Betcke, M.; Huynh, N.; Adler, J.; Cox, B.; Beard, P.; Ourselin, S.; Arridge, S. Model-based Learning for Accelerated, Limited-View 3-D Photoacoustic Tomography. IEEE Trans. Med. Imaging 2018, 37, 1382–1393. [Google Scholar] [CrossRef]
  113. Zhu, J.; Huynh, N.; Ogunlade, O.; Ansari, R.; Lucka, F.; Cox, B.; Beard, P. Mitigating the Limited View Problem in Photoacoustic Tomography for a Planar Detection Geometry by Regularized Iterative Reconstruction. IEEE Trans. Med. Imaging 2023, 42, 2603–2615. [Google Scholar] [CrossRef] [PubMed]
  114. Roy, K.; Lee, J.E.-Y.; Lee, C. Thin-Film PMUTs: A Review of Over 40 Years of Research. Microsyst. Nanoeng. 2023, 9, 95. [Google Scholar] [CrossRef] [PubMed]
  115. Atheeth, S.; Krishnan, K.; Arora, M. Review of pMUTs for Medical Imaging: Towards High Frequency Arrays. Biomed. Phys. Eng. Express 2023, 9, 022001. [Google Scholar] [CrossRef]
  116. Qiu, Y.; Gigliotti, J.V.; Wallace, M.; Griggio, F.; Demore, C.E.; Cochran, S.; Trolier-McKinstry, S. Piezoelectric Micromachined Ultrasound Transducer (PMUT) Arrays for Integrated Sensing, Actuation and Imaging. Sensors 2015, 15, 8020–8041. [Google Scholar] [CrossRef]
  117. Zheng, Q.; Wang, H.; Yang, H.; Jiang, H.; Chen, Z.; Lu, Y.; Feng, P.X.-L.; Xie, H. Thin Ceramic PZT Dual-and Multi-Frequency pMUT Arrays for Photoacoustic Imaging. Microsyst. Nanoeng. 2022, 8, 122. [Google Scholar] [CrossRef]
  118. Demi, L.; Demi, M.; Prediletto, R.; Soldati, G. Real-Time Multi-Frequency Ultrasound Imaging for Quantitative Lung Ultrasound–First Clinical Results. J. Acoust. Soc. Am. 2020, 148, 998–1006. [Google Scholar] [CrossRef]
Figure 1. Schematic illustration for 3D ultrasound image generation using LAT. LAT, linear array transducer; 3D, three-dimensional.
Figure 1. Schematic illustration for 3D ultrasound image generation using LAT. LAT, linear array transducer; 3D, three-dimensional.
Bioengineering 12 00906 g001
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Tsedendamba, N.; Song, Y.; Park, E.-Y.; Kim, J. Review of Linear-Array-Transducer-Based Volumetric Ultrasound Imaging Techniques and Their Biomedical Applications. Bioengineering 2025, 12, 906. https://doi.org/10.3390/bioengineering12090906

AMA Style

Tsedendamba N, Song Y, Park E-Y, Kim J. Review of Linear-Array-Transducer-Based Volumetric Ultrasound Imaging Techniques and Their Biomedical Applications. Bioengineering. 2025; 12(9):906. https://doi.org/10.3390/bioengineering12090906

Chicago/Turabian Style

Tsedendamba, Ninjbadgar, Yuon Song, Eun-Yeong Park, and Jeesu Kim. 2025. "Review of Linear-Array-Transducer-Based Volumetric Ultrasound Imaging Techniques and Their Biomedical Applications" Bioengineering 12, no. 9: 906. https://doi.org/10.3390/bioengineering12090906

APA Style

Tsedendamba, N., Song, Y., Park, E.-Y., & Kim, J. (2025). Review of Linear-Array-Transducer-Based Volumetric Ultrasound Imaging Techniques and Their Biomedical Applications. Bioengineering, 12(9), 906. https://doi.org/10.3390/bioengineering12090906

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop