AR-HUD Optical System Design and Its Multiple Conﬁgurations Analysis

: The use of augmented reality head-up displays (AR-HUD) in automobile safety driving has drawn more and more interest in recent years. An AR-HUD display system should be developed to ﬁt the vehicle and the complicated trafﬁc environment in order to increase the driver’s driving concentration and improve the man–vehicle synchronization. In this article, we suggest an AR-HUD display system with dual-layer virtual-image displays for the near ﬁeld and far ﬁeld, as well as further research and design of the adjustment system for multi-depth displays of far-ﬁeld images. It also examines the EYEBOX horizontal adjustment margin of the dual light path. The analysis results show that the scale of EYEBOX is 120 × 60 mm 2 , the modulation transfer function (MTF) of near-ﬁeld light path > 0.2 @ 6.7 lp/mm, and the MTF of far-ﬁeld optical path > 0.4 @ 6.7 lp/mm. The distortion of the near-ﬁeld optical path is less than 0.86%, and that of the far-ﬁeld optical path is less than 2.2%. By modifying the folding mirror, the far-ﬁeld optical path creates an 8 m to 24 m multi-depth virtual picture display. Image quality can be maintained when the near-ﬁeld and far-ﬁeld optical paths are moved horizontally by 25 mm and 100 mm, respectively. This study offers guidelines for the multi-depth display, EYEBOX horizontal adjustment, and optical layout of augmented reality head-up displays.


Introduction
Head-up display (HUD) is a display technology based on vehicle-machine interactions that reflects vehicle driving information into the human eye through optical components so that the driver can read the information in a horizontal view to improve driving safety, such as speed, fuel consumption, navigation information, driver assistance information, warning information, etc.The first use of HUD was in the 1950s, for aiming and shooting [1].In 1960, the Hawker Siddeley Buccaneer first used heads-up display technology in military aircraft [2].HUD was first deployed in commercial aviation in 1970 [3].In 1988, HUD technology was initially made available in Oldsmobile and Pontiac models by General Motors (GM) [4].At present, the Boeing 787 Dreamliner is the first mainstream commercial aircraft to use HUD as the standard model [5].The benefit of HUD is that it allows the driver to avoid moving between the flat (0-5 • ) road conditions and the overhead (20-25 • ) dashboard, which eliminates the need for constant switching between dark and light settings.The threshold for traffic accidents is taking one's eyes off the road for longer than two seconds, which is a primary contributor to traffic accidents [6].Equipped with HUD, system drivers can read vehicle driving information in a flat perspective, especially to reduce the risk of traffic accidents for senior drivers, enhance driving coordination, and improve man-machine efficiency.
Combiner HUD (C-HUD), Windshield HUD (W-HUD), and Augmented Reality HUD (AR-HUD) are some of the existing head-up display systems for cars.W-HUD and AR-HUD both reflect the image source into the driver's eyes through the front windshield, and AR-HUD can be seen as an upgraded version of W-HUD.Rear-mounted HUDs like the C-HUD, which also includes instrument panel and suspension types, pose a safety risk to drivers in the event of an accident.Front-mounted embedded HUD including such image generation units (PGUs), flat reflector, freeform reflector, wedge-shaped front windscreen, etc., are called W-HUD and AR-HUD.W-HUD is the primary HUD product at the moment, with virtual image dimensions of 7~12 inches, a field of view (FOV) of fewer than 10 degrees, and a virtual image distance (VID) of 1.8~2.5 m.The inevitable trend of HUD development is toward AR-HUD, which has a VID of more than 7 m, a FOV of more than 10 • , and a larger virtual image size.It can display information from advanced driver assistance systems and realize true human, vehicle, and road collaboration [7,8].The global installed base of HUDs will reach 15 million units in 2025, with W-HUD serving as the primary product, and close to 35 million units in 2030, with AR-share HUD's expanding quickly, according to Continental's head-up display market development forecast data [9].
According to the type of imaging, current AR-HUD optical systems can be divided into two categories: geometric optical imaging and physical optical imaging.Geometric optical imaging include the author-designed off-axis three-mirror HUD optical system, where the PGU, plane mirror, and freeform reflector were laid out in the same horizontal plane [10][11][12].By using Solidworks' basic mechanical assembly structure to determine the coordinates of the plane mirror and PGU, the light constraint equations were developed to guarantee unhindered light transmission [11].The study could not have an augmented reality influence on the Advanced Driving Assistance System information that should have been given in the far-field optical path because it was only designed and analyzed for the near-field light path.To accomplish infinity imaging with a FOV of 24 • × 15 • , the authors created an off-axis quad-mirror HUD optical system for airplanes [12].The initial spatial architecture of the four-mirror system was acquired by restricting the edge beam of the maximum FOV.In the end, the biological lens parallax was assessed and adjusted to produce infinity imaging with a FOV of 24 • × 15 • .The design of this aircraft head-up display system is only designed and analyzed for the virtual image at infinity, which does not have the effect of layered display.Using lens eccentricity and tilt to correct for the aberrations and off-axis aberrations caused by the asymmetric FOV and the tilted and bent combiner, physical optical imaging researchers like the authors designed a holographic HUD optical system with an asymmetric large FOV [13][14][15][16].However, physical optical imaging technology is still in its infancy and cannot currently handle the demand for batch use.Other AR-HUD optical systems exist that use two PGUs and a dual optical route separation design method, but these systems require more room to install.At the same time, we have yet to see an AR-HUD system that performs a continuous multi-depth display with a large range of far-field light paths, which will not maximize the augmented reality effect of the AR system.Therefore, it is essential to suggest a new AR-HUD optical system with dual-layer display functionality that uses a coupling architecture of near-field and far-field dual optical routes.Based on this, the optimum augmented reality impact requires its far-field optical channel to have a wide range of continuous multi-depth virtual image display capabilities.This article is devoted to exploring this problem.
In this paper, we develop a single PGU approach for linking near-field and far-field dual optical routes in order to solve the issue of huge space occupation while designing nearfield and far-field split optical channels with dual PGUs for AR-HUD [13,14].The design difference of the upper and lower FOV deviation allows the near-field and far-field dual optical paths to deviate in position on the free surface reflector, while the far-field optical path is connected to the common image plane by a folding mirror and the near-field optical path.The optical distortion of the near and far field is less than 0.86% and 2.2%, respectively, and this design achieves an MTF of the near-field optical path > 0.2 @ 6.7 lp/mm and an MTF of the far-field optical path > 0.4 @ 6.7 lp/mm.The spot diagrams of the near-and far-field light paths are also within the Airy disk.In addition, taking into account the requirement for far-field image distance in various road scenes, a design analysis for a multi-depth variable VID for far-field images from 8 to 24 m was carried out.This design analysis shows that the requirement for the multi-depth virtual image quality can be met only by adjusting the distance of three folding mirrors, while also guaranteeing the constant PGU position at various VIDs.It was discovered that there was no appreciable loss of image quality when the near-field optical path was adjusted horizontally by 25 mm in the preset EYEBOX plane and by 100 mm in the far-field optical path in the preset EYEBOX plane within the EYEBOX horizontal adjustment range (120 × 60 mm 2 ) in the near-field and far-field optical paths.

Design Considerations and System Parameters
The AR-HUD optical system consists of two optical routes: near field and far field, with the single PGU scheme and the double PGUs scheme being used for the two optical paths, respectively.The single PGU scheme divides a projection screen into near-field and far-field sections and creates two virtual images of various sizes and distances through two optical paths.This method has clear volumetric advantages and is the current development trend for AR-HUD systems.From the virtual image end to the human eye, then to the windshield, freeform reflector, folding mirror, and PGU, the AR-HUD optical path system uses an inverse optical path design.The AR-HUD optical route is intended to generate virtual images of various horizontal distances through various object distance settings and virtual images of various vertical heights through various up and down viewing angles.The windshield 3D model can be used to fit the optical surface shape, and an optical reflection model can be built.Different models' data for the windshield surface and inclination angle varies slightly.The design does the design analysis at a 45 • tilt angle condition and cites the windshield face pattern data from the study [17].As an off-axis, unobstructed reflective optical system, the HUD introduces a number of aberrations like a spherical aberration, coma, astigmatism, field curvature, distortion, etc., due to the system architecture and hyperbolic features of the windshield glass surface.When designing an aberration, balance and correction are accomplished using a freeform reflector with strong aberration correction capability.Zernike polynomial and Seidel aberration have a direct mapping relationship that allows for the effective and individual correction of each aberration while also allowing for the conversion to a more processing-friendly XY polynomial surface type.The design of the freeform reflector uses the Zernike standard sag surface type.The Zernike Standard Sag surface is defined by the same polynomial as the Even Aspheric surface (which supports planes, spheres, conics, and polynomial aspheres) plus additional aspheric terms defined by the Zernike Standard coefficients.The surface sag is of the form: where c denotes the basal curvature, k is the aspheric coefficient, r is the radial ray coordinate in lens units, α i is the coefficient on the ith Even Aspheric polynomial, N is the number of Zernike coefficients in the series, A i is the coefficient on the ith Zernike Standard polynomial, ρ is the normalized radial ray coordinate, and ϕ is the angular ray coordinate.The cut-off frequency of the modulation transfer function (MTF), which corresponds to the PGU pixel size of 75 µm in the design, is 6.7 lp/mm.The driver's EYEBOX, which is where their eyes are focused when they are operating a vehicle, is where the virtual image that PGU projects may be seen.Its front and back depth can be adjusted, and its design size is 120 × 60 mm 2 .The preset distance of far-field virtual image is not less than 7 m, the FOV is not less than 10 • , the distance of near-field image is around 2 m, and the FOV is not less than 6 • .The authors used two freeform reflectors to design an AR-HUD optical system including only the far-field optical path, with a horizontal FOV of 10 • and a vertical FOV of 5 • [8].The authors used freeform reflector to design an AR system including only the near-field optical path, with a horizontal FOV of 5 • and a vertical FOV of 2.5 • [11].This study suggests designing a dual-field AR-HUD optical system with a 6 • × 2 • FOV for the near-field optical path and a 10 • × 3 • FOV for the far-field optical path, taking into account the challenge of dual-optical path coupling design and the upper and lower viewing angle deviation of the near-field and far-field optical paths.The design indexes of AR-HUD optical system are shown in Table 1.
Table 1.AR-HUD optical system design indexes.

Parameters
Near-Field Image Far-Field Image The fully reflective optical system, which has flexible design forms and fewer optical components than the conventional refractive optical system, does not cause chromatic aberration and is suitable for broad-spectrum and multispectral imaging, especially for large FOV and large aperture optical system design.Due to the fact that off-axis characteristics still necessitate up-and down-view deviations for AR-HUD near-field and far-field optical routes, freeform reflectors may provide two optical layouts, as shown in Figure 1a,b, depending on the reflectors' placement angles.Three folding mirrors fold the far-field optical path and connect it to a shared image plane with the near-field optical path, providing an adjustment margin to maintain the image plane position constant throughout the ensuing 8~24 m continuous multi-depth virtual image adjustment.
m, and the FOV is not less than 6°.The authors used two freeform reflectors to design an AR-HUD optical system including only the far-field optical path, with a horizontal FOV of 10° and a vertical FOV of 5° [8].The authors used freeform reflector to design an AR system including only the near-field optical path, with a horizontal FOV of 5° and a vertical FOV of 2.5° [11].This study suggests designing a dual-field AR-HUD optical system with a 6° × 2° FOV for the near-field optical path and a 10° × 3° FOV for the far-field optical path, taking into account the challenge of dual-optical path coupling design and the upper and lower viewing angle deviation of the near-field and far-field optical paths.The design indexes of AR-HUD optical system are shown in Table 1.The fully reflective optical system, which has flexible design forms and fewer optical components than the conventional refractive optical system, does not cause chromatic aberration and is suitable for broad-spectrum and multispectral imaging, especially for large FOV and large aperture optical system design.Due to the fact that off-axis characteristics still necessitate up-and down-view deviations for AR-HUD near-field and far-field optical routes, freeform reflectors may provide two optical layouts, as shown in Figure 1a,b, depending on the reflectors' placement angles.Three folding mirrors fold the far-field optical path and connect it to a shared image plane with the nearfield optical path, providing an adjustment margin to maintain the image plane position constant throughout the ensuing 8~24 m continuous multi-depth virtual image adjustment.

Design Method
The optical arrangement of the dual optical channel is one of the challenges in the original design of the AR-HUD optical system.Several simulations were used to tweak the design to obtain the two optical configurations shown in Figure 1.The optical arrangement in Figure 1a produces the effect of multiple mirrors without blocking and dual optical routes with a common image plane.The design and analysis work of this study is focused on this optical layout.The coupling of the dual optical routes' initial structures must be realized via ZEMAX's multi-configuration function since the two optical paths' fields of view, VIDs, and upper and lower viewing angles are all out of alignment.The freeform reflector is shared by the two optical pathways, and because the near-field optical path is relatively straightforward and heavily dependent on the placement angle of the freeform reflector, the priority design of the near-field optical path can be taken into account.By repositioning several mirrors to create a double optical route without blocking coupling, the far-field optical way is based on the near-field optical path, and the multi-stage folding mirror is optimized to meet the requirements of the far-field optical path design.The radius of curvature, rotation angle, and distance of each reflector, as well as the image plane coupling of the two optical channels, are the parameters that must be constrained during the design process.
Although the design process should consider the F-number of the near-field and far-field optical paths, the design should be optimized using the full aperture of 134 mm.However, the image quality is evaluated for the size of the human eye pupil, considering the human eye pupil maximum.The system F-number will increase 16.75 times in the case of 8 mm.Using the diffraction limit equation as a guide [18]: where ∆θ is the minimum resolution angle of the optical system, λ is the wavelength, and D is the pupil diameter.When ∆θ is very small, sin∆θ ≈ ∆θ, which is approximately equal to D/f, D is the minimum resolution size, and f is the focal length, we can obtain: The diffraction limit aperture is 104.7 when the center wavelength is 0.587 um and the image element size is 75 µm.Consequently, when maximizing for an optical system with a human eye pupil of 8 mm at a full aperture of 134 mm, the system F-number should be close to 6.3 or even lower.

Design Results and Analysis
The varied amounts of aberrations in the dual optical routes are corrected after the synergistic optimization design of the near-field and far-field optical paths, and the superior image quality design under the dual structure is obtained.Figure 2 displays the outcomes of the dual optical route design, with the near-field optical path illustrated in green and the far-field optical path shown in blue.A freeform reflector and folding reflector are both present in the blue box.The far-field optical path likewise has two spherical foldable mirrors and one aspheric foldable mirror, and the two optical channels share a freeform reflector.The entire EYEBOX domain can use the design results because the design procedure adopts a 134 mm pupil diameter.The human eye's pupil can grow to 8 mm in diameter in complete darkness, and the optical system exhibits significant aberration.The visual quality that the human eye sees in these circumstances is suitable for situations with smaller pupils.The near-field and far-field MTF curves are depicted in Figure 3a,b, respectively, with the spatial frequency acting as the horizontal coordinate and the value of the diffraction modulation transfer function acting as the vertical coordinate.The MTF at the Nyquist frequency of 6.7 lp/mm is greater than 0.2 in the virtual image range of 210 × 70 mm 2 for the near-field optical path and is greater than 0.4 in the virtual image range of 1400 × 420 mm 2 for the far-field optical path.The imaging quality of the dual optical path is excellent.The image quality of the near-field optical path did not change significantly when the image plane moved in the range of about ±3 mm, and the image quality of the far-field optical path did not change significantly when image plane moved in the range of about ±10 mm.The optical system's depth of focus is also larger, and the axial adjustment margin of the PGU projection screen is larger, as shown in Figure 4 s MTF out-of-focus curve.Figure 5a,b depict the spot diagrams of the near-field and far-field optical paths, respectively.F1~F9 correspond to the spot diagrams of nine FOV points.The near-field optical path Airy spot's radius is 83.55 µm larger than its maximum RMS radius of 58.752 µm, and the far-field optical path Airy spot's radius is 71.26 µm larger than its maximum RMS radius of 16.631µm, indicating that the dual optical path imaging resolution reaches the diffraction limit.The near-field and far-field MTF curves are depicted in Figure 3a,b, respectively, with the spatial frequency acting as the horizontal coordinate and the value of the diffraction modulation transfer function acting as the vertical coordinate.The MTF at the Nyquist frequency of 6.7 lp/mm is greater than 0.2 in the virtual image range of 210 × 70 mm 2 for the near-field optical path and is greater than 0.4 in the virtual image range of 1400 × 420 mm 2 for the far-field optical path.The imaging quality of the dual optical path is excellent.The image quality of the near-field optical path did not change significantly when the image plane moved in the range of about ±3 mm, and the image quality of the far-field optical path did not change significantly when image plane moved in the range of about ±10 mm.The optical system's depth of focus is also larger, and the axial adjustment margin of the PGU projection screen is larger, as shown in Figure 4′s MTF out-of-focus curve.Figure 5a,b depict the spot diagrams of the near-field and far-field optical paths, respectively.F1~F9 correspond to the spot diagrams of nine FOV points.The near-field optical path Airy spot's radius is 83.55 μm larger than its maximum RMS radius of 58.752 μm, and the far-field optical path Airy spot's radius is 71.26 μm larger than its maximum RMS radius of 16.631μm, indicating that the dual optical path imaging resolution reaches the diffraction limit.The near-field and far-field optical routes have optical distortions that are less than 0.86% and 2.2%, respectively.Figure 6a,b illustrate the grid distortion effect, where the horizontal and vertical coordinates stand in for the horizontal FOV and the vertical FOV, respectively.The actual images have a slight distortion effect in comparison to the ideal image, but this does not affect how the images are used.The human eye's pupil distance is 65 mm, so when two people are observing the same object, there will be a parallax between their two eyes.Figure 7 displays the binocular parallax of the 2 m near-field image system and the 8 m far-field image system.As shown in Figure 7a, the near-field system's binocular divergence parallax is less than 1 mrad and its convergence parallax is less than 2 mrad.As shown in Figure 7b, the far-field system's binocular divergence parallax is around 1.5 mrad and its convergence parallax is less than 2 mrad, which meets the standard value of clause 4.2.10 in SAE AS8055 [19].It avoids the physiological discomfort caused by binocular parallax when observing images.The near-field and far-field optical routes have optical distortions that are less than 0.86% and 2.2%, respectively.Figure 6a,b illustrate the grid distortion effect, where the horizontal and vertical coordinates stand in for the horizontal FOV and the vertical FOV, respectively.The actual images have a slight distortion effect in comparison to the ideal image, but this does not affect how the images are used.The human eye's pupil distance is 65 mm, so when two people are observing the same object, there will be a parallax between their two eyes.Figure 7 displays the binocular parallax of the 2 m near-field image system and the 8 m far-field image system.As shown in Figure 7a, the near-field system's binocular divergence parallax is less than 1 mrad and its convergence parallax is less than 2 mrad.As shown in Figure 7b, the far-field system's binocular divergence parallax is around 1.5 mrad and its convergence parallax is less than 2 mrad, which meets the standard value of clause 4.2.10 in SAE AS8055 [19].It avoids the physiological discomfort caused by binocular parallax when observing images.

Multi-Structure Analysis of Far-Field Image
For better augmented reality, the AR-HUD far-field virtual image must be merged

Multi-Structure Analysis of Far-Field Image
For better augmented reality, the AR-HUD far-field virtual image must be merged with markers placed across the real world at various distances, which suggests that the

Multi-Structure Analysis of Far-Field Image
For better augmented reality, the AR-HUD far-field virtual image must be merged with markers placed across the real world at various distances, which suggests that the far-field virtual image is multi-depth adjustable.Two mirrors-a zoom mirror and a compensating mirror-make up the reflective zoom optical system.It is challenging to implement zoom in AR-HUD since it takes up a lot of room and is tough to design.The review mentions a variety of multi-focus display methods, and similarly, the design creates a multi-depth virtual image display by modifying the optics of configuration 2 under the prototype of AR-HUD geometric imaging [20].With reference to the coaxial zoom optical system, the zooming process maintains the size of the image plane by flipping the field-of-view and focal distance, as well as coordinating the change in image distance.In this case, a smaller field-of-view, a larger VID, and coordinated focal distance change can be used to maintain the size of the virtual image plane.The center coordinates of the PGU image plane are constrained by the design optimization process to remain constant, and multi-depth virtual image distribution is only achieved by adjusting the total 4 distances (L2-L5) of the freeform reflector-folding mirror 1 (F1)-folding mirror 2 (F2)-folding mirror 3 (F3)-image plane.To achieve unhindered light during the adjustment under this optical structure, the distance L2 should rise as the VID increases.Throughout the design phase, the distance L2 is provided as an empirical number, along with the optimized distances L3, L4, and L5. Figure 8 depicts the isometric perspective, and Table 2 displays the specifications of the multi-depth virtual image system.The coordinated shift in focal length makes optical design less challenging, and the decrease in FOV is accompanied by a decrease in the size of the PGU image and an increase in optical magnification.The design ensures that the fundamental position of the center of the PGU picture plane remains intact while allowing for an increase in VID by altering the distance between folding mirrors (EYEBOX plane is the global coordinate reference plane).Using samples for three depths of 8 m, 16 m, and 24 m, the design achieves continuous depth adjustment of the virtual picture from 8 m to 24 m.The three colors correspond to the positions of the folding mirror at three image depths, as shown in the adjustment schematic for the folding mirror at three virtual image depths in the bottom right of Figure 8.As seen in Figure 9, where Figure 9a depicts the MTF curve for the first sample depth (8 m) as well as the far-field optical path in Section 2.2.3.The MTF curves for the second sampling depth (16 m) and third sampling depth (24 m), respectively, are shown in Figure 9b,c.It can be seen that the MTF values of each FOV at the second and third sampling depths are slightly decreased by the adjustment of the three reflective elements.However, the overall image quality is still excellent and has little impact on how the virtual image is commonly used.In the actual use of the PGU far-field projection screen should be added to the frame scaling function, in order to match the PGU frame reduction characteristics generated by the variable VIDs of the system.The coordinated shift in focal length makes optical design less challenging, and the decrease in FOV is accompanied by a decrease in the size of the PGU image and an increase in optical magnification.The design ensures that the fundamental position of the center of the PGU picture plane remains intact while allowing for an increase in VID by altering the distance between folding mirrors (EYEBOX plane is the global coordinate reference plane).Using samples for three depths of 8 m, 16 m, and 24 m, the design achieves continuous depth adjustment of the virtual picture from 8 m to 24 m.The three colors correspond to the positions of the folding mirror at three image depths, as shown in the adjustment schematic for the folding mirror at three virtual image depths in the bottom right of Figure 8.As seen in Figure 9, where Figure 9a   reflective elements.However, the overall image quality is still excellent and has little impact on how the virtual image is commonly used.In the actual use of the PGU farfield projection screen should be added to the frame scaling function, in order to match the PGU frame reduction characteristics generated by the variable VIDs of the system.

EYEBOX Multiple Structure Analysis
The intended AR-HUD optical system EYEBOX is at a 900 mm distance from the front windshield; nevertheless, during real use, the driver will adjust their head slightly back and forth, which will interfere somewhat with their ability to observe the virtual image.In the system design of the retro-optical path, it is equivalent to the impact of the micro-adjustment changes in the spatial distance of the object on the image quality of the PGU image plane.Chapter 2 conducted a coupling design analysis on the close range and distant dual optical paths, which involved many variables.On the basis of Chapter 2, Section 3.1 implements the multi-depth virtual image analysis of the prospective light path only by adjusting the distance of the folding mirror.Only the EYEBOX horizontal adjustment margin of the near-field and far-field light path created by ZEMAX's multi-

EYEBOX Multiple Structure Analysis
The intended AR-HUD optical system EYEBOX is at a 900 mm distance from the front windshield; nevertheless, during real use, the driver will adjust their head slightly back and forth, which will interfere somewhat with their ability to observe the virtual image.In the system design of the retro-optical path, it is equivalent to the impact of the micro-adjustment changes in the spatial distance of the object on the image quality of the PGU image plane.Section 2 conducted a coupling design analysis on the close range and distant dual optical paths, which involved many variables.On the basis of Section 2, Section 3.1 implements the multi-depth virtual image analysis of the prospective light path only by adjusting the distance of the folding mirror.Only the EYEBOX horizontal adjustment margin of the nearfield and far-field light path created by ZEMAX's multi-configuration functional analysis will be employed in this case because other optical components were not optimized.On the basis of Section 2, the distance between configuration 1 and configuration 2's EYEBOX surfaces was changed in a triple structure, and the EYEBOX placements were discovered to have a negligible impact on the quality of the images.The EYEBOX surface with one step added or deleted for analysis is represented by the level 1 and level −1 structures in Figure 10, whereas the level 0 structure reflects the distance already designed in the previous section.Among them, the step size for configuration 1 is 25 mm, and the step size for configuration 2 is 100 mm. Figure 11 shows the MTF curves of the near-field optical pat EYEBOX structures of 925 mm, 900 mm, and 875 mm.Among the is the MTF curve of the near-field optical path in 2.2.3, and Figur curves of the near-field optical path for increasing and decreasing by 25 mm, respectively.The full-field MTF value of the level 1 greater than 0.17@6.7 lp/mm.The full-field MTF value of level − better than 2.1@6.7 lp/mm.It is known that the near-field optica horizontal distance adjustment based on the default EYEBOX sur to no impact on the virtual image's quality.Figure 12b (level 0) for the far-field optical path in Section 2.2.3, whereas Figure 12a,c for the far-field optical path by altering the EYEBOX distance by The full-field MTF value of the level 1 EYEBOX structure is greate The full-field MTF value of level −1 EYEBOX structure is better th the basis of the predetermined EYEBOX surface (level 0), it can be Figure 11 shows the MTF curves of the near-field optical path under the three-level EYEBOX structures of 925 mm, 900 mm, and 875 mm.Among them, Figure 11b (level 0) is the MTF curve of the near-field optical path in 2.2.3, and Figure 11a,c show the MTF curves of the near-field optical path for increasing and decreasing the EYEBOX distance by 25 mm, respectively.The full-field MTF value of the level 1 EYEBOX structure is greater than 0.17 @ 6.7 lp/mm.The full-field MTF value of level −1 EYEBOX structure is better than 2.1 @ 6.7 lp/mm.It is known that the near-field optical path with a ±25 mm horizontal distance adjustment based on the default EYEBOX surface (level 0) has little to no impact on the virtual image's quality.Figure 12b (level 0) depicts the MTF curve for the far-field optical path in Section 2.2.3, whereas Figure 12a,c depict the MTF curves for the far-field optical path by altering the EYEBOX distance by 100 mm, respectively.The full-field MTF value of the level 1 EYEBOX structure is greater than 0.35 @ 6.7 lp/mm.The full-field MTF value of level −1 EYEBOX structure is better than 0.42 @ 6.7 lp/mm.On the basis of the predetermined EYEBOX surface (level 0), it can be seen that the far-field light path is modified horizontally by ±100 mm, however, the quality of the virtual image will not vary greatly.
for the far-field optical path by altering the EYEBOX distance by 100 mm, respectiv The full-field MTF value of the level 1 EYEBOX structure is greater than 0.35@6.7 lp/m The full-field MTF value of level −1 EYEBOX structure is better than 0.42@6.7 lp/mm the basis of the predetermined EYEBOX surface (level 0), it can be seen that the far-f light path is modified horizontally by ±100 mm, however, the quality of the virtual age will not vary greatly.

Conclusions
The prior concepts of single optical path design for near-field HUDs and far-field AR-HUDs are integrated in this study to create a dual optical path coupled adjustable AR-HUD optical system.To actualize the categorized display of in-vehicle information at 2 m and 8 m, a single PGU is separated into near-field and far-field display zones, with the far-field optical route being relayed using the folding mirrors.Only the axial movement of the three folding mirrors is required to show the ADAS information and navigation information of the far-field optical path at any distance between 8 m and 24

Conclusions
The prior concepts of single optical path design for near-field HUDs and far-field AR-HUDs are integrated in this study to create a dual optical path coupled adjustable AR-HUD optical system.To actualize the categorized display of in-vehicle information at

Figure 2 .
Figure 2. Structure diagram for two configurations of AR-HUD.

Figure 2 .
Figure 2. Structure diagram for two configurations of AR-HUD.

Figure 8 .
Figure 8. Multi-depth virtual images of far-field optical path.
depicts the MTF curve for the first sample depth (8 m) as well as the far-field optical path in Section 2.2.3.The MTF curves for the second sampling depth (16 m) and third sampling depth (24 m), respectively, are shown in Figure9b,c.It can be seen that the MTF values of each FOV at the second and third sampling depths are slightly decreased by the adjustment of the three

Figure 8 .
Figure 8. Multi-depth virtual images of far-field optical path.

Figure 9 .
Figure 9. MTF curves: (a) MTF curve for VID of 8 m; (b) MTF curve for VID of 16 m; and (c) MTF curve for VID of 24 m.

Figure 9 .
Figure 9. MTF curves: (a) MTF curve for VID of 8 m; (b) MTF curve for VID of 16 m; and (c) MTF curve for VID of 24 m.

Figure 11 .
Figure 11.MTF curves for the near-field optical path: (a) MTF curve of the level 1 EYEBOX MTF curve of the level 0 EYEBOX; and (c) MTF curve of the level −1 EYEBOX.

Figure 11 .
Figure 11.MTF curves for the near-field optical path: (a) MTF curve of the level 1 EYEBOX; (b) MTF curve of the level 0 EYEBOX; and (c) MTF curve of the level −1 EYEBOX.Photonics 2023, 10, x FOR PEER REVIEW 12 of 13

Figure 12 .
Figure 12.MTF curves for the far-field optical path: (a) MTF curve of the level 1 EYEBOX; (b) MTF curve of the level 0 EYEBOX; and (c) MTF curve of the level −1 EYEBOX.

Figure 12 .
Figure 12.MTF curves for the far-field optical path: (a) MTF curve of the level 1 EYEBOX; (b) MTF curve of the level 0 EYEBOX; and (c) MTF curve of the level −1 EYEBOX.

Table 1 .
AR-HUD optical system design indexes.

Table 2 .
Multiple structure parameters for multi-depth VID.