Abstract
Due to the complex marine environment, side-scan sonar signals are unstable, resulting in random non-rigid distortion in side-scan sonar strip images. To reduce the influence of resolution difference of common areas on strip image mosaicking, we proposed a mosaic method for side-scan sonar strip images based on curvelet transform and resolution constraints. First, image registration was carried out to eliminate dislocation and distortion of the strip images. Then, the resolution vector of the common area in two strip images were calculated, and a resolution model was created. Curvelet transform was then performed for the images, the resolution fusion rules were used for Coarse layer coefficients, and the maximum coefficient integration was applied to the Detail layer and Fine layer to calculate the fusion coefficients. Last, inverse Curvelet transform was carried out on the fusion coefficients to obtain images in the fusion area. The fusion images in multiple areas were then combined in the registered images to obtain the final image. The experiment results showed that the proposed method had better mosaicking performance than some conventional fusion algorithms.
1. Introduction
As the depth of global ocean exploration continues to increase, understanding the seafloor surface and near-surface is of great significance in the “digital ocean” and “transparent ocean” era. Currently, side-scan sonar is an important means to explore seafloor geomorphology [1], and side-scan sonar images provide important data for seafloor object identification, classification of seafloor sediments, and exploration of marine resources [2,3]. In order to obtain side-scan sonar images in the entire testing zone, the most important task is to mosaic the strip images, in addition to seafloor tracking, slant-range correction, gain correction, and geocoding [4]. Side-scan sonar is generally operated using towing cables, which leads to inaccurate location information. If the coordinate information is directly used to mosaic the images, there will be distortion in the images [5,6,7,8]. Currently, a large number of studies have been carried out to achieve a mosaic of object-level strip images that produce images with complete information and of high quality.
By dividing sonar strip images into paired objects and shadows, Daniel et al. [9] realized rigid registration of side-scan sonar images using a decision tree. Through region segmentation, Thisen et al. [10] extracted shadow areas from side-scan sonar images and calculated the displacement between two images using the cross-correlative method, thereby achieving rigid registration. Vandrish et al. [11] showed that the scale invariant feature transform (SIFT) algorithm can be used for registration of sonar images, although the accuracy was not ideal. Using correlation coefficients and mutual information as similarity parameters, Chailloux et al. [12] extracted a series of significantly correlated areas on adjacent strip images and calculated the global rigid transformation parameters and local elastic transformation parameters, thereby eventually realizing mosaic of adjacent strip images. Wang et al. [13] improved the pre-processing method of side-scan sonar images to extract feature points more accurately and effectively after preprocessing; they also proposed a sped up robust feature (SURF)-based elastic mosaic algorithm to achieve feature-level conformal mosaic of the images. Moreover, Cao et al. [14] used wavelet transform in a strip image mosaic, yet it required the 3D posture information of the side-scan sonar. Zhao et al. [15] extracted SURF features of the pre-processed strip images and then performed block registration, which achieved good mosaic results. To obtain sonar images of large-area seafloor surface, Zhao et al. [16] also proposed a side-scan image mosaicking method based on the coupling feature points of position constraints. In addition, He et al. [17] used the unsharp masking (USM) algorithm to enhance the side-scan images and the SURF algorithm for image mosaicking; experiments showed that their method effectively enhanced image features and increased the amount of image information, but the average gray values of the images were affected.
The above image mosaic algorithms primarily focused on the extraction and registration of features points of adjacent strip images, and most adopted the wavelet fusion algorithm after image registration, without further exploration for alternative image fusion algorithms. Due to the complex marine environment during ocean exploration, it is nearly impossible to ensure that the sonar images on one survey line are always better than those of an adjacent strip image. Therefore, it is necessary to take into account the differences in image resolution during strip image mosaicking and retain clear image information while screening necessary information in blurred images. To address this problem, we performed image fusion using curvelet transform, which can reveal more detailed information of strip images than wavelet transform. Then, the resolution of strip images was evaluated using a resolution weight model to constrain the curvelet transform, thereby achieving mosaicked strip images with better quality. The contents of this paper were arranged as follows: Section 2 mainly introduces seven different methods of resolution assessment, which would all be used in the calculation of resolution weight model; Section 3 mainly introduces the specific process of strip Mosaic method proposed in this paper; Section 4 uses the measured data to verify the feasibility of this method; and Section 5 contains the summary and prospects.
2. Image Resolution Assessment Methods
As an important data source of seafloor geomorphology, the resolution of side-scan sonar images directly determines the accuracy of target identification and seafloor sediments classification. The assessment of image quality can be divided into two types: subjective assessment and objective assessment [18,19]. Subjective assessment is mainly performed by trained professionals, whereas objective assessment uses mathematical models to measure the image resolution based on different indices. Thus, it is imperative to develop an objective assessment method that is in consistency with subjective assessment. Currently, common objective assessment methods can be divided into three categories according to the degree of use of reference images, i.e., full-reference quality assessment, reduced-reference quality assessment, and no-reference quality assessment [20]. Since there is no original reference image for side-scan sonar images, the no-reference quality assessment method was adopted in this study.
Image resolution is one of the most important image quality evaluation indexes and is the most important image parameter of sonar image. Therefore, the resolution of image became the main research object. A total of seven resolution assessment methods from four aspects will be introduced in this section. As the more classical parameter indexes in the assessment method, they measure the sharpness of the image from different aspects. Additionally, they will all be used in the calculation of resolution vector in Section 3, making the evaluation result more accurate and perfect.
2.1. Assessment Method Based on Image Gradient
Image gradient reflects the marginal information of images. The greater the gradient value is, the sharper the image edge and the clearer the image will be. Common gradient functions for evaluating image resolution include the following three types [21].
2.1.1. Energy Gradient Function
The energy gradient of an image is the quadratic sum of the difference in grayscale value of adjacent pixels in the horizontal and vertical direction. The summation of energy gradient values of all pixels in the image is then taken as the function value. The function is shown in Equation (1):
where x and y are pixel coordinates, and is the grayscale value of the pixel.
2.1.2. Brenner Gradient Function
Brenner gradient function is relatively the easiest gradient assessment function [22]. It calculates the quadratic sum of the grayscale difference of two adjacent pixels, meaning a small calculation amount. Yet, it is sensitive to noise. The function is shown in Equation (2):
2.1.3. Tenengrad Gradient Function
Krotkv et al. [23] used the Tenengrad gradient function as one of the assessment indexes of image resolution, the results of which were close to objective assessment results. In this method, the Sobel operator was first used to extract the horizontal and vertical gradient values of pixels, then the quadratic sum was compared with a threshold T. The gradient values of pixels greater than T were added to obtain the Tenengrad gradient function value. The function is shown in Equation (3):
where is the gradient calculated by the Sobel operator, as shown in Equation (4):
where and represent the horizontal and vertical gradient values, respectively.
where is the convolution operator, and and represent the horizontal and vertical templates of the Sobel operator, respectively:
2.2. Assessment Method Based on Image Transform Domain
It is generally believed that a clear image contains more high-frequency components than a blurry image. Thus, some studies have attempted to transform the image to the frequency domain to perform image quality assessment [24].
2.2.1. Discrete Fourier Transform (DFT)
As the most basic time–frequency transformation methods, DFT is widely used in resolution assessment. Specifically, 2D DFT is first performed on the image, and then the zero-frequency component is shifted to the matrix center, such that the frequency diffuses from the center to the periphery and from low frequency to high frequency. The spectrum values of corresponding pixels are weighted based on the distance to the central pixel, and the resolution assessment value is the weighted average of the spectrum values of all pixels [25,26]. The function of DFT-based image resolution assessment is shown in Equation (7) [27]:
where M and N are the image dimensions, represents the distance of a pixel to the central pixel, and is the spectrum value of a pixel after DFT.
2.2.2. Discrete Cosine Transform (DCT)
DFT-based resolution assessment methods have high sensitivity; however, they are computationally more demanding than DCT-based methods. In comparison, DCT has a general orthogonal transform property, and the base vector of DCT matrix could describe image features very well [28,29]. Therefore, by replacing DFT with DCT, the transform coefficient is changed into a real number, which reduces the computation while still obtaining the distribution of image frequency. The resolution assessment function based on DCT is shown in Equation (8):
where is the spectrum value of a pixel after DCT.
2.3. Assessment Method Based on Entropy Function
The entropy of an image is an important index to measure the richness of image information. Shannon believed that the greater the entropy value, the richer information the image contains. During image resolution assessment, the clearer the image is, the more abundant grayscale distribution it has, and thus, the greater the entropy value is [30]. The definition of entropy function is shown in Equation (9):
where is the probability of occurrence of every grayscale value.
2.4. Assessment Method Based on Variance Function
The variance function can represent the dispersion degree of the image grayscales. The smaller the range of grayscale, the smaller the variance is and the blurrier the image is, and vice versa [31]. The definition of variance function is shown in Equation (10):
where is the average grayscale value of the image, the definition of which is in Equation (11):
3. Strip Mosaic Method Based on Curvelet Transform and Resolution Constraints
3.1. Image Fusion Algorithm Based on Curvelet Transform
To obtain a clear and continuous image that can reflect complete information of the entire testing zone, image fusion in the overlapping area of side-scan sonar strip images is required. Currently, there are three common image fusion methods, namely weighted average method, image pyramid method, and wavelet fusion method [32]. The wavelet fusion method is the most common side-scan sonar strip image mosaicking method. However, due to the limitations in algorithms, the wavelet transform can only obtain edge features in the horizontal and vertical directions, and the wavelet basis does not have the anisotropy property. Hence, it is unable to get close to the image texture features. To overcome the limitations in the wavelet transform and improve the quality of strip image mosaicking, the Curvelet transform was introduced in the current study.
The Curvetlet transform was first proposed by Candes and Donoho in 1999 [33] based on the Ridgelet transform. As a multi-resolution, band-pass, and directional multi-scale image analysis method, Curvelet transform has the three characteristics of an optimal image representation method proposed by the National Institute for Physiological Science, Japan [34]. Similar to wavelet transform, Curvelet transform calculates the correlation of spatial images using a group of base functions, thereby characterizing edges and curves at different angles. The main steps of image fusion based on Curvelet transform are as follows: Curvelet coefficients are first obtained from Curvelet decomposition of the image, the coefficients are then processed based on specific fusion rules, and lastly, inverse Curvelet transform is carried out on the fused coefficient to obtain the final fusion image [35,36].
The Curvelet coefficients are obtained using the equation below:
where is the input image, are the image dimensions, is the scale, is the direction, is the spatial location of Curvelet, and represents the Curvelet function, which includes a group of base functions described by parameters .
Different from the wavelet coefficients, the Curvelet coefficients include the low-frequency coefficient in the innermost layer (i.e., the Coarse layer), the mid-to-high frequency coefficient in the Detail layer, and the high-frequency coefficient in the outermost Fine layer. As the number of layers increases, the scale of the corresponding base function turns smaller, and there are more directions. Figure 1 shows a frequency-domain base division method. Each square in Figure 1 represents a scale, and there are five scales. The bigger the square, the higher the frequency, and the smaller the scale is; hence, more detailed information will be reflected. The radial lines represent the angles. At each scale, the angle division is different, and the higher the frequency is, the smaller the angle is.
Figure 1.
Frequency-Domain Base of Curvelet Transform.
From Jia et al. [37], the energy of coefficients is mainly concentrated in the low-frequency coefficient, and the energy gradually declines as the frequency increases. In other words, the low-frequency coefficient reflects the general trend of the image, whereas high-frequency coefficient reflects the outline and texture details of an image. By fusing the coefficients at various layers using different fusion rules, the fusion image coefficient can be obtained, and by performing inverse Curvelet transform of the fusion image coefficient, the fusion image is obtained.
3.2. Strip Image Mosaicking Based on Curvelet Transform and Resolution Constraints
Due to uncertainties in the marine environment during exploration, common areas in adjacent strip images might have large differences during actual measurement. Both strip images might have good quality, or one or both of them may not be good at all. The traditional side-scan strip image mosaicking algorithms do not take the image resolution into account. In order to ensure good mosaic results, a Curvelet coefficient fusion criterion based on the resolution weight model was proposed in the present study.
In Section 2, we have introduced seven different image resolution assessment methods, including energy gradient function, Brenner gradient function, Tenengrad gradient function, DFT, DCT, entropy function, and variance function. According to Li et al. [38] and Xie et al. [39], different resolution assessment methods may have different results for the same group of images. In other words, a single method is not able to assess the resolution of an image accurately. Hence, these seven resolution assessment methods were integrated in this study to build a resolution vector, and the image resolution was obtained based on probability and given weights.
The resolution vector , created based on the resolution value of the above seven methods, is shown in Equation (13):
Since the resolution index in each method has a positive relationship with the image resolution, the resolution weight is obtained by comparing the resolution vectors of image 1 and image 2, , respectively.
where the resolution weight represents the probability of an image having better resolution than the other image. Thus, it was taken as the fusion rule in the Coarse layer of Curvelet transform, as shown in Equation (15).
where , and represent the coefficient in the Coarse layer after fusion and that of image 1 and image 2, respectively.
In order to fully show the texture and details of the image, the maximum coefficient fusion approach was adopted to process the Detail layer and Fine layer coefficients, as shown in Equation (16):
where , and represent the coefficient in the Detail layer after fusion and that of image 1 and image 2, respectively. , and represent the coefficient of the Fine layer after fusion and that of image 1 and image 2, respectively. Figure 2 shows the flowchart of the proposed mosaic method based on Curvelet transform and resolution constraints.
Figure 2.
Flowchart of the proposed method.
- Extract and match feature points of adjacent strip images and obtain registered mosaic strips using the affine transformation.
- Select the common area A from two strip images.
- Perform Curvelet transform for two images to obtain the coefficients in the Coarse layer, Detail layer, and Fine layer.
- Calculate the resolution vectors of the two images to obtain the corresponding resolution weight.
- Fuse the Coarse layer coefficients using resolution fusion rules to obtain the low-frequency coefficients. Fuse the Detail layer and Fine layer coefficients using the maximum coefficient fusion rules to obtain the high-frequency coefficients.
- Perform inverse Curvelet transform on the fusion coefficients to obtain the fusion image in area A, which is then mosaicked to the registered strip images.
- Repeat steps 2–6 until the whole mosaic image is obtained.
In traditional mosaic algorithms for strip images, there are various problems, such as inconsistent resolution of adjacent strip images and image distortion. In this study, we proposed a mosaic method for strip images based on Curvelet transform and resolution constraints, which produced mosaic images with complete information and high quality.
4. Experiment and Results
To verify the effectiveness of the proposed image mosaicking method, image data collected in 2019 using the Klein4000 side-scan sonar in Jiaozhou Bay, Qingdao, Shandong Province, China was used in the experiment. The water depth of the survey area is approximately 30–40 m. The overlapping rate of adjacent strip images is 50%. After preprocessing, such as seafloor tracking, slant-range correction, gray level equalization, noise suppression, gain correction, and geocoding, a group of strip image pairs with obvious common features were selected, as shown in Figure 3.
Figure 3.
Two strips used for verification. Four image pairs with obvious common features were selected.
Figure 4a shows a mosaic image calculated based on geographic coordinate information. As can be seen, there is obvious dislocation and distortion. According to the steps of our method, the feature points in the strip images were extracted and matched, as shown in Figure 4b. Figure 4c shows a registered strip image after affine transformation. Based on the results, the distortion and dislocation were eliminated after image registration, resulting in good visual effects and laying a solid foundation for image fusion in the next step.
Figure 4.
Strip image registration. A–D and E–H are four areas selected from (a) and (c) respectively. (b) shows the registration process of strips. It can be seen that there was significant dislocation in A–D. After strip registration, the dislocation effect largely disappeared in E–H.
To effectively select the fusion area and ensure the integrity of the selected features, the whole survey area was first rotated counterclockwise for a certain angle, such that the survey line was approximately along the vertical direction [40]. Another reason to rotate the strips is that a series of subsequent steps, such as Curvelet transform and image fusion, require regular rectangles. After image mosaicking, it was rotated back to the original direction. Areas 1–3 were selected, and the sonar images of two strips in these areas are shown in Figure 5.
Figure 5.
Selected fusion areas.
Taking Area 1 as an example, the proposed algorithm was used to process two strips in the area. First, the coefficients in the Coarse, Detail, and Fine layers were extracted using Curvelet transform. The coefficient structure is shown in Table 1.
Table 1.
Structure of Curvelet transform coefficients.
In both strips, Area 1 has the same dimensions of . Five layer decomposition was carried out. As shown in Table 1, the dimension of the coefficient matrix increases with the increase in scale. The larger the scale in spatial domain, the smaller the scale in frequency domain, and the more detailed the description of high frequency information.
Then, the resolution vectors of two strips in Area 1 were calculated, and the results are shown in Table 2.
Table 2.
Resolution vectors of two strips in Area 1.
Then, using the proposed algorithm, the coefficients in the Coarse layer of the two images were fused based on the resolution fusion rule, and the coefficients in the Detail and Fine layers of the two images were fused using the maximum coefficient fusion approach, thereby obtaining the low-frequency and high-frequency coefficients of the fused image. Lastly, the fused image of Area 1 was obtained via inverse Curvelet transform.
In order to verify the rationality of the resolution fusion rule proposed in this paper, the resolution fusion rule, the mean fusion rule, and the maximum fusion rule are used to combine the five layer coefficients of the two images obtained by the Curvelet decomposition, respectively. As shown in Figure 6, 19 combinations of fusion coefficients were obtained.
Figure 6.
Combination diagram of fusion rules about Curvelet transform coefficients.
Then, the fusion coefficients of each group were inversely transformed to obtain fusion images.
The information entropy, average gradient, and spatial frequency were used as evaluation indices of the fusion results. The information entropy reveals the amount of information contained in the image, and the greater the entropy, the better the fusion result; the average gradient reflects the image’s contrast expression of small details, and the greater the average gradient, the higher the image fusion quality; the spatial frequency represents the overall activity of the image in spatial domain, and the higher the spatial frequency, the better the fusion result. Table 3 shows the three indices of each combination, and Figure 7 shows the line chart of the analysis results.
Table 3.
Comparison of fusion effects in different combinations.
Figure 7.
Line chart of the fusion coefficients analysis results.
As shown in Table 2 and Figure 7, the Curvelet coefficient fusion strategy proposed in this paper, namely the resolution fusion rule used in the Coarse layer and the maximum coefficient fusion rule used in the Detail layer and Fine layer, has the best image fusion effect.
To further demonstrate the effectiveness of the proposed algorithm, the images were fused using different algorithms, including simple average, traditional wavelet fusion and wavelet fusion with resolution constraints. The fusion results were compared with that of the proposed algorithm. The traditional wavelet fusion algorithm applies the mean fusion rule to the low-frequency information of wavelet transform and the maximum coefficient fusion rule to the high-frequency information. In the wavelet fusion with resolution constraints, the resolution fusion rule is applied to the low-frequency information of wavelet transform and the maximum coefficient fusion rule is applied to the high-frequency information.
Table 4 shows the three indices of the four fusion methods, and Figure 8 shows the fusion strip images.
Table 4.
Comparison of fusion results of different methods in Area 1.
Figure 8.
Fusion results of the four different methods.
As shown in Table 4, the information entropy, average gradient, and spatial frequency of the proposed algorithm are much greater than those of the other three methods, indicating that the fusion result of the proposed method is the best. By comparing the results of wavelet fusion with resolution constraints and our method, it can be seen that Curvelet fusion achieved better fusion results than wavelet fusion. In addition, based on the value of indices of traditional wavelet fusion and wavelet fusion with resolution constraints, the effectiveness of the resolution fusion rule proposed in this study was demonstrated. It can also be seen intuitively from Figure 8 that the fusion image obtained by our method has better clarity and can show more details.
To further verify the effectiveness of the proposed method, the same experiments were repeated for Areas 2 and 3. Figure 9 shows the fusion strip images in Area 2 and Area 3. The evaluation results are shown in Table 5.

Figure 9.
(a) shows the strip images and fusion strip images in Area 2. (b) shows the strip images and fusion strip images in Area 3.
Table 5.
Comparison of fusion results of different methods in Areas 2 and 3.
As shown in Table 5, the fused images in Areas 2 and 3 of the proposed method have the highest information entropy, average gradient, and spatial frequency, suggesting the best performance in image fusion and validating the effectiveness and stability of the proposed algorithm.
Then, the fused images in the three areas were mosaicked onto the registered strip, which was then rotated clockwise to the original orientation, as shown in Figure 10. Compared with Figure 4c, it can be seen that Figure 9 better reflects the overall characteristics of the features by enhancing detail texture information while retaining the overall trend of the overlapping areas.
Figure 10.
Results of strip image mosaicking.
5. Conclusions
Current strip image mosaicking algorithms do not consider the influence of the resolution difference of common objects in adjacent images on the results of mosaicking. Moreover, a traditional wavelet fusion algorithm is not able to fully describe the image details. To address these problems, in this study, we proposed an image mosaic method based on Curvelet transform and resolution constraints. Experimental verification using actual measurement data showed that the proposed method can greatly improve the fusion results, which provides high-quality image data for subsequent submarine target recognition and sediment classification, thereby greatly benefiting ocean exploration. However, there are still a lot of improvements to be made in this method, such as human involvement in the process. In view of this, target recognition and other technologies in deep learning can be introduced in the future. Thus, it can automatically identify and extract the areas that need to be fused and achieve full automation.
Author Contributions
Conceptualization, S.J. and N.Z.; methodology, S.J. and N.Z.; software, G.B. and Y.C.; validation, L.C., Y.C. and G.B.; formal analysis, S.J.; investigation, N.Z. and L.C.; resources, S.J.; data curation, L.C.; writing—original draft preparation, N.Z.; writing—review and editing, N.Z., S.J. and G.B.; visualization, G.B.; supervision, Y.C.; project administration, S.J.; funding acquisition, S.J. All authors have read and agreed to the published version of the manuscript.
Funding
This research was funded by National Natural Science Foundation of China, grant number: 41876103.
Institutional Review Board Statement
Not applicable.
Informed Consent Statement
Not applicable.
Data Availability Statement
The data is kept in_house and not available online.
Conflicts of Interest
The authors declare no conflict of interest.
References
- Philippe, B. The Handbook of Sidescan Sonar; Springer: Berlin/Heidelberg, Germany; New York, NY, USA, 2009; ISBN 978-3-540-42641-7. [Google Scholar]
- Fakiris, E.; Papatheodorou, G.; Geraga, M.; Ferentinos, G. An Automatic Target Detection Algorithm for Swath Sonar Backscatter Imagery, Using Image Texture and Independent Component Analysis. Remote Sens. 2016, 8, 373. [Google Scholar] [CrossRef] [Green Version]
- Wang, X.; Zhao, J.; Zhu, B.; Jiang, T.; Qin, T. A Side Scan Sonar Image Target Detection Algorithm Based on a Neutrosophic Set and Diffusion Maps. Remote Sens. 2018, 10, 295. [Google Scholar] [CrossRef] [Green Version]
- Cervenka, P.; Moustier, C.D. Sidescan sonar image processing techniques. IEEE J. Ocean. Eng. 1993, 18, 108–122. [Google Scholar] [CrossRef] [Green Version]
- Chavez, P.S.C., Jr.; Isbrecht, J.A.; Galanis, P.; Gabel, G.L.; Sides, S.C.; Soltesz, D.L.; Ross, S.L.; Velasco, M.G. Processing, mosaicking and management of the monterey bay digital sidescan-sonar images. Mar. Geol. 2002, 181, 305–315. [Google Scholar] [CrossRef]
- Wang, A. Research on 3D Seafloor Terrian Recovery from the Side Scan Sonar Image; Wuhan University: Wuhan, China, 2014. [Google Scholar]
- Cervenka, P.; Moustier, C.; Lonsdale, P. Geometric corrections on sidescan sonar images based on bathymetry: Application with SeaMARC II and Sea Beam data. Mar. Geophys. Res. 1995, 17, 217–219. [Google Scholar] [CrossRef] [Green Version]
- Cervenka, P.; Moustier, C. Postprocessing and corrections of bathymetry derived from sidescan sonar systems: Application with SeaMARC II. IEEE J. Ocean. Eng. 1994, 19, 619–629. [Google Scholar] [CrossRef]
- Daniel, S.; FL, L.; Roux, C. Side-scan sonar image matching. IEEE J. Ocean. Eng. 1998, 23, 245–259. [Google Scholar] [CrossRef]
- Thisen, E.; Sorensen, H.B.; Stage, B. Sidescan Sonar Image Matching Using Cross Correlation; International Society for Optics and Photonics: Bellingham, WA, USA, 2003; Volume 5089. [Google Scholar]
- Vandrish, P.; Vardy, A.; Walker, D. Side-scan sonar image registration for AUV navigation. In Proceedings of the 2011 IEEE Symposium on Underwater Technology and Workshop on Scientific Use of Submarine Cables and Related Technologies, Tokyo, Japan, 5–8 April 2011; IEEE: Piscataway, NJ, USA, 2011. [Google Scholar]
- Chailloux, C.; Le Caillec, J.M.; Gueriot, D.; Zerr, B. Intensity-Based Block Matching Algorithm for Mosaicing Sonar Images. IEEE J. Ocean. Eng. 2011, 36, 627–645. [Google Scholar] [CrossRef]
- Wang, A.; Zhao, J.; Guo, J.; Wang, X. Elastic Mosaic Method in Block for Side-Scan Sonar Image Based on Speeded-Up Robust Features. J. Wuhan Univ. Inf. Sci. Ed. 2018, 43, 697–703. [Google Scholar]
- Cao, M.; Guo, J. Research on Side Scan Sonar Stripe Image Mosaic Method Based on Common Vision Target. Geomat. Spat. Inf. Technol. 2014, 46–49, 52. [Google Scholar]
- Zhao, J.; Wang, A.; Wang, X. Research on Segment Mosaic Method of Side Scan Sonar Stripe Image. Geomat. Inf. Sci. Wuhan Univ. 2013, 9, 29–33. [Google Scholar]
- Zhao, J.; Shang, X.; Zhang, H. Side-Scan Sonar Image Mosaic Using Couple Feature Points with Constraint of Track Line Positions. Remote Sens. 2018, 10, 953. [Google Scholar] [CrossRef] [Green Version]
- He, F.; Wu, M.; Long, R.; Chen, Z. Accurate Mosaic of side scan sonar image based on SURF feature. Chin. J. Ocean. Technol. 2020, 39, 35–41. [Google Scholar]
- Dharmishtha, P.; Jaliya, U.; Vasava, H. A review: No-reference/blind image quality assessment. Int. Res. J. Eng. Technol. 2017, 4, 339–343. [Google Scholar]
- Yang, X.; Li, F.; Liu, H. A survey of DNN methods for blind image quality assessment. IEEE Access 2019, 7, 123788–123806. [Google Scholar] [CrossRef]
- Wang, Z. Review of no-reference image quality assessment. Acta Autom. Sin. 2015, 41, 1062–1079. [Google Scholar]
- Chern, N.N.K.; Neow, P.A.; Ang, M.H. Practical issues in pixel-based autofocusing for machine vision. In Proceedings of the 2001 ICRA. IEEE International Conference on Robotics and Automation (Cat. No. 01CH37164), Seoul, Korea, 21–26 May 2001; IEEE: Piscataway, NJ, USA, 2001; pp. 2791–2796. [Google Scholar]
- Subbarao, M.; Tyan, J. Selecting the optimal focus measure for autofocusing and depth-from-focus. IEEE Trans. Pattern Anal. Mach. Intell. 1998, 20, 864–870. [Google Scholar] [CrossRef] [Green Version]
- Krotkov, E.P. Active Computer Vision by Cooperative Focus and Stereo; Springer: New York, NY, USA, 1989; pp. 1–17. ISBN 9781461396659. [Google Scholar]
- Chetouani, A.; Beghdadi, A.; Deriche, M. A new reference-free image quality index for blur estimation in the frequency domain. In Proceedings of the 2009 IEEE International Symposium on Signal Processing and Information Technology (ISSPIT), Ajman, United Arab Emirates, 14–17 December 2009; IEEE: Piscataway, NJ, USA, 2009; pp. 155–159. [Google Scholar]
- Vu, C.; Phan, T.; Chandler, D. S3: A spectral and spatial measure of local perceived sharpness in natural images. IEEE Trans. Image Process. 2012, 21, 934–945. [Google Scholar] [CrossRef]
- Lu, Y.; Zhang, T.; Zheng, J.; LI, M.; Zhang, C. No-reference blurring image quality assessment based on local standard deviation and saliency map. J. Jilin Univ. Eng. Technol. Ed. 2016, 46, 1337–1343. [Google Scholar]
- Li, Z. Study on Large-Scale Image Registration Algorithm in Fourier Transform Domain; Nanjing University of Information Science & Technology: Nanjing, China, 2013. [Google Scholar]
- Marichal, X.; Ma, W.; Zhang, H. Blur determination in the compressed domain using DCT information. In Proceedings of the 1999 International Conference on Image Processing (Cat. 99CH36348), Kobe, Japan, 24–28 October 1999; IEEE: Piscataway, NJ, USA, 1999; pp. 386–390. [Google Scholar]
- Caviedes, J.; Oberti, F. A new sharpness metric based on local kurtosis, edge and energy information. Signal Process. Image Commun. 2004, 19, 147–161. [Google Scholar] [CrossRef]
- Zhu, Z.; Li, S.; Chen, H. Research on Automatic Focusing Function Based on Image Entropy. Opt. Precis. Eng. 2004, 5, 537–542. [Google Scholar]
- Wang, C.; Cui, L.; Yan, B. Research on the evaluation algorithm of microscopic image sharpness based on Variance-Brenner function. Equip. Manuf. Technol. 2020, 10, 78–82. [Google Scholar]
- Xu, J. Research on Key Technologies of Side-Scan Sonar Image Mosaic and Segmentation; East China University of Technology: Fuzhou, China, 2017. [Google Scholar]
- Candes, E. Curvelets—A Surprisingly Effective Nonadaptive Representation for Objects with Edges. 2020. Available online: http://citeseerx.ist.psu.edu/viewdoc/summary;jsessionid=0B25E5104FB2CB688F212E84E4F2BD75?doi=10.1.1.161.9294 (accessed on 1 September 2021).
- Chen, Z. Curving Transform and Information Extraction Application of Potential Field Data; China University of Geosciences: Beijing, China, 2012. [Google Scholar]
- Vijayalakshmi, G.; Shanthakumar, M.; Muthukumar, K. A Contrast Adjusted Wavelet and Curvelet Transform Based Fusion of Breast Images. IOP Conf. Ser. Mater. Sci. Eng. 2020, 994, 012027. [Google Scholar] [CrossRef]
- Tian, Y.; Zhang, H.; Ma, X. An Image Fusion Method Based on Curvelet Transform and Guided Filter Enhancement. Math. Probl. Eng. 2020, 2020, 9821715. [Google Scholar] [CrossRef]
- Jia, X. Research on Face Recognition Algorithm Based on Curvwave Transform; Harbin Engineering University: Harbin, China, 2011. [Google Scholar]
- Li, Z.; Li, X.; Ma, L. Research on Definition Assessment based on No-reference Digital Image Quality. Remote. Sens. Technol. Appl. 2011, 26, 239–246. [Google Scholar]
- Xie, X.; Zhou, J.; Wu, Q. A Nonreference Quality Evaluation Index for Image Blur. Comput. Appl. 2010, 4, 921–924. [Google Scholar]
- Gao, F.; Wang, X.; Yang, J.; Zhang, B.; Zhou, H.; Chen, J. Research on Multi-strip Side Scan Sonar Image Fine Mosaic Method. Sci. Technol. Innov. Appl. 2021, 5, 1–4. [Google Scholar]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2021 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).