Next Article in Journal
Cost-Efficient Two-Level Modeling of Microwave Passives Using Feature-Based Surrogates and Domain Confinement
Previous Article in Journal
A Deep Learning Framework for Adaptive Beamforming in Massive MIMO Millimeter Wave 5G Multicellular Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Self-Supervised Sound Promotion Method of Sound Localization from Video

School of Electronicand Electrical Engineering, Shanghai University of Engineering Science, 333 Longteng Road, Shanghai 201620, China
*
Author to whom correspondence should be addressed.
Electronics 2023, 12(17), 3558; https://doi.org/10.3390/electronics12173558
Submission received: 13 July 2023 / Revised: 14 August 2023 / Accepted: 21 August 2023 / Published: 23 August 2023
(This article belongs to the Topic Computer Vision and Image Processing)

Abstract

:
Compared to traditional unimodal methods, multimodal audio-visual correspondence learning has many advantages in the field of video understanding, but it also faces significant challenges. In order to fully utilize the feature information from both modalities, we needs to ensure accurate alignment of the semantic information from each modality, rather than simply concatenating them together. This requires consideration of how to design fusion networks that can better perform this task. Current algorithms heavily rely on the network’s output results for sound-object localization while neglecting the potential issue of suppressed feature information due to the internal structure of the network. Thus, we propose a sound promotion method (SPM), a self-supervised framework that aims to increase the contribution of voices to produce better performance of the audiovisual learning. We first cluster the audio separately to generate pseudo-labels and then use the clusters to train the backbone of audio. Finally, we explore the impact of our method to several existing approaches on MUSIC datasets and the results prove that our proposed method is able to produce better performance.

1. Introduction

As humans, we not only perceive the world through vision, but also use hearing, tasting, touching, etc. If only an image is provided, the existing algorithm can already meet most of people’s needs, such as object category recognition, action understanding, text translation, etc. But if a video is provided, the original research in the field of single-modal images cannot cover the understanding of the video well, because the video has not only one more time dimension than the image, but also additional sound information, and the image and sound in the video are naturally corresponding, which is the key to join them in our network. Therefore, audio-visual correspondence learning is gradually separated from the original image field and sound field, and combined into a new research direction. Now, with the influx of a large number of researchers, this direction also has many branches, such as speaker separation [1,2,3,4], sound source localization [5,6,7], speech recognition [8,9,10,11], audio-visual retrieval [12,13], etc.
In this paper, we focus on the problem of sound source localization [14] which has recently become one of the mainstream research projects of Audio-Visual Learning (AVL), because we believe that the accuracy of the sound localization reflects the network’s learning of the video. In addition, since there are various unlabeled videos on the Internet, the common starting point of recent work is to learn the position information of the sound in the video in a self-supervised or weakly supervised way. The earlier works feed information from both modalities into the same network [5,6,7,15], like, for example, Arandjelovic et al. [5], who utilized the audiovisual correspondence (AVC) to find the sound localization in 2018. Then, Zhao et al. [16] proposed a mix-and-separate approach to compute the sound of each pixel by an audio synthesizer network, and Senocak et al. [15] fused the information of two modalities by using the attention mechanism which can be applied to supervised or un-supervised learning. In 2019, Hu et al. [17] clustered audio and visual representations within each modality to calculate similarity through contrastive learning. In 2021, Lin et al. [18] extended this method by using an iterative contrastive learning algorithm. Based on this, Chen et al. [19] also proposed a tri-map method to divide a picture into more detailed positive and negative samples, and in 2022, Song et al. [20] discarded negative samples directly to form a negative-free method and propose a predictive coding module (PCM) for feature alignment. At the same time, Senocak et al. [21] used multi-task classification to train the backbone of the audiovisual network.
These methods did localize the sounding object, but they just focus on the image or the image-to-audio relationship and ignore the audio. According to the conclusion in [21], the feature extractor obtained through a classification task is applicable to the research on sound source localization algorithms. Therefore, it can be inferred that the classification accuracy of the two-modal feature extractor followed by a classifier reflects its contribution to the final results. Based on this observation, we connected two classifiers after the image and audio feature extractors, respectively. Through a series of training experiments, it was found that the classification accuracy of the image channel is approximately 65%, while the accuracy of the audio classifier is only about 27%. It turns out that the sound itself was not fully exploited and utilized, which lead to the ambiguity of sound source location. Thus, different from the aforementioned articles, we propose our sound promotion method (SPM) which makes the network pay more attention to the audio and obtains a more precise localization of the sound source. The procedure is as follows: First, since we want to improve the contribution of audio, we cluster audio information features separately. Then, we use the pseudo-label of sound to train the backbone of the audio (Figure 1). We freeze this part and train the whole network, because we find that with the network training, the entire network focuses on the image level, which makes the audio contribution decline. Finally, the results indicate that our method greatly improves the accuracy, which proves the effectiveness of our method.
As Figure 1 shows, our method mainly includes three key parts: the audio image feature extraction network, the sound promotion framework and the multi-modal fusion framework (AV localization). The audio and image feature extraction networks are designed using the first five layers of a ResNet-18 model with a residual structure. Considering that the audio input is a mel-log spectrum, i.e., a one-dimensional signal, the number of audio channels is set to one in the first layer of the audio extractor. Meanwhile, for image input, the image frames in the video are randomly selected as input. At the same time, the image feature extraction network is obtained by pre-training on the ImageNet Datasets, while the audio feature extraction network uses the sound promotion method (SPM) proposed in this study which uses sound clustering to generate pseudo-label categories for training, freezing its parameters and then performing transfer learning. Through these two feature extraction networks, high-level semantic information of images and sounds can be obtained. Subsequently, this information is input into the multi-modal fusion framework for settlement. In this framework, by calculating the cosine similarity between the image and the sound, and through a sigmoid function, a mask consistent with the size of the image is generated to obtain the information of the sounding object in the image. Since images and audio come from the same video, there is a natural correspondence between them, so they are used as pseudo-labels to supervise the learning network to achieve the goal of self-supervised learning. In the end, the mask generated by the network is able to accurately localize the location of the sounding object.
Our main contributions of this work are summarized as follows: (1) We mine the contribution of audio and image to the results. (2) We propose a novel sound promotion method (SPM). It clusters the audio separately to generate pseudo-labels and then uses the clusters to train the backbone of audio. (3) We explore the impact of our method on several existing approaches on MUSIC datasets [18] which show that our method can improve the sound source localization.

2. Materials and Methods

The goal of our model is to localize sound sources by increasing the contribution of voice to the whole network. Most of the existing works [17,19,20,22] are all about improvement in the field of image, and they do not consider optimizing audio as one of the modalities to improve network performance.
Different from the aforementioned methods, our method is shown in Figure 1 in the dashed frame and can be explained as follows. Because our method is validated on DSOL [23] as the baseline, we first introduce the model of the baseline, and then introduce our method.

2.1. Audiovisual Model of the Baseline

We obtain visual frames v i and audio spectrogram a j from video clips X = { v i , a j | v i R C × H × W , a j R C } ; H and W are the spatial size. Then, we obtain their visual feature representation V i and audio feature representation A j through vision embedding  f v · and audio embedding f a · .
V i = f v v i , v i R C × H × W ,
A j = f a a j , a j R C .
Note that it is regarded as a positive sample that the audio and image are from the same video, i.e.,  i = j . Otherwise, when they come from different videos, i.e.,  i j , the value is perceived as a negative sample.
In addition, the cosine similarities are computed in Equation (3) by feeding the vision and audio features, V i and A j .
c o s i n e i j u v = A j , V i : u v A j V i : u v , u v h × w .
Then, we use the sigmoid function σ to obtain the mask of sounding object m i j as follows:
m i j = σ c o s i n e i j ϵ / τ ,
where ϵ refers to the thresholding parameter and τ is the temperature. Hence, we can extract potential object representation O i which represents the degree of correlation between audio and video,
O i = G A P f v v i m i j ,
where G A P is the Global Average Pooling operation and ∘ is the Hadamard product.

2.2. Sound Similar Mining

As we have lots of audio information, clustering is the most efficient way for self-supervision. We cluster the audio representation A j obtained through audio embedding f a · to generate pseudo-labels. Then, we use Equation (6) as the criterion for classification and find the argmax D i μ which denotes the most comparable sound.
D i μ = A i T A μ | i 1 , n , μ 1 , k , a r g m a x D i μ ,
where A μ are the centroids of k categories, and n refers to the mounts of audio clips.

2.3. Audio Backbone Training

Inspired by [22], where it is demonstrated that the backbone trained by the classifier can perform well on the sound localization task, we utilize the pseudo-labels of sound to train our audio embedding layers and freeze this part of parameters in the following training.
And The overall training process of the algorithm is as Algorithm 1 shows.
Algorithm 1: Sound Promotion Method (SPM)
Electronics 12 03558 i001

2.4. Dataset

The MUSIC (Multimodal Sources of Instrument Combinations) dataset [18] contains 714 untrimmed videos of musical solos and duets, covering 11 classes of musical instruments. To demonstrate the effectiveness of our approach, we perform the same job as that in the Hu et al. [23] work, where the first five/two videos of each instrument category in solo/duet are used for testing, and the rest is used for training. Note that some videos are now not available on YouTube. We finally obtained 422 videos.

3. Results

3.1. Evaluation Metric

We adopt Consensus Intersection over Union (CIoU) and Area Under Curve (AUC) as the evaluation metrics to quantitatively analyze the results. The annotated bounding boxes are generated by a Faster RCNN network [24] to evaluate the positioning accuracy of the algorithm we proposed for the sounding object. The parameters of the Faster R-CNN model are obtained by training the public datasets, Open Image Datasets. Since this algorithm studies the detection of sounding objects, the instrument category in the Open Image Datasets were selected as the training set when training the Faster R-CNN model. During the training process, 15 musical instrument categories were selected from the Open Image Datasets, and a total of more than 30,000 images participated in the training of the Faster R-CNN model. These instrument categories include accordion, banjo, cello, drum, guitar, harp, harmonica, oboe, piano, saxophone, trombone, trumpet, violin, flute and horn. These 15 musical instrument categories basically cover the main categories of common musical instruments, so the target detection model Faster R-CNN trained using them can accurately identify musical instruments in the video and label relevant regions. Although these 15 categories do not include all musical instruments, they meet the criteria for this algorithm test. When detecting potential sounding objects in an actual video scene, many detection results are obtained. In order to filter out the most valuable results, we only keep the detection results with a confidence level higher than 90%, and manually clean up the remaining detection boxes. After such a screening process, the obtained target detection coordinates basically meet the requirements of the sound source localization test.
C I o U = k = 1 K δ k I o U k k = 1 K δ k ,
where I o U k is computed by the predicted sounding object area and annotated bounding box. The indicator is δ k = 1 where the object is sounding, otherwise, it is 0.

3.2. Implementation Details

We divide each video equally into one-second clips with no intersection and randomly sample one image from the divided clips as the visual inputs, which is resized to 256 × 256. Then, we randomly crop them to a 224 × 224 size. As for audio, the inputs are first re-sampled into 16 KHz, then translated into a spectrogram via Short Time Fourier Transform (STFT). Log-Mel projection is performed over the spectrogram to better represent sound features. The audio and visual input from the same video clip are regarded as a positive sample. We use lightweight ResNet-18 [25] as audio and visual backbones. Our network is trained with the Adam optimizer with a learning rate of 10 4 . Similar to the baselines [23], we pretrain the backbone of visual on ImageNet. In addition, we use θ = 0.65, τ = 0.03 and k = 11 as our network’s hyper-parameters. Finally, all these experiments are run on the computer with a GPU of 3080 10 G, and it takes about three weeks to complete the training stage.

3.3. Quantitative Results

In this section, we compare our method (SPM) with the existing sound localization on the MUSIC dataset. First, in Figure 2, we show the reason why we separately improve the audio contribution. With the increase in training time, the accuracy of the audio classifier is much lower than that of the image classifier. According to the inference in [22], we can deduce the classifier performance of sound to show its contribution to the final result from the side. It was found that the accuracy of the image classifier was approximately 65%, while the accuracy of the audio classifier was only about 27%. This indicates that although both audio and images are used as inputs for network training, images contribute more to the final results, while audio has a lower level of involvement. Through in-depth analysis, it was discovered that the image feature extraction network, due to its pre-training, may become stuck in the vicinity of local optima, resulting in the phenomenon of “weak listening”. Thus, our method works in the freezing of the backbone of audio with a higher extraction rate to help it better learn the sound features.
Then, in Figure 3, it can be observed that at the beginning of training, the DSOL algorithm performs well in overall accuracy because it is only pre-trained on images, ignoring audio information. However, as training progresses, the accuracy of DSOL starts oscillating and converges at 94.5%. In contrast, the convergence curve of our method shows a steady improvement trend. Although the overall accuracy is lower than that of DSOL at the beginning of training, with an increase in the number of iterations, the proposed algorithm gradually surpasses the comparative algorithm and ultimately achieves an accuracy of 96.5%. We believe that the multimodal fusion in the comparative network is mainly dominated by visual information. When the visual network’s extractor reaches its limit, the contribution of the audio extractor is lower, resulting in network oscillation. In contrast, for our algorithm, the audio extractor is frozen, and only the image extractor and the final classifier are trained. Due to the larger operatable space of image information, the overall network can progress towards the optimal point, avoiding the oscillation phenomenon. This indicates that increasing the contribution of audio features can lead to higher accuracy, bringing the network closer to the global optimum.
In Table 1, it is shown that the method we propose outperforms the existing method on the test set. Due to the lack of data or the use of undisclosed optimization algorithms in the comparative algorithms during training, we cannot fully reproduce the experimental results from other papers. Therefore, we can only compare the accuracy provided in the other papers. Even so, the algorithm we propose still demonstrates better performance. When the CIoU is set to 0.3, in comparison to the best DSOL algorithm, our method improves by 2.8 percentage points and significantly surpasses the benchmark results of previous studies. We hypothesize that because the previous method did not have the operation like that of our model, the network could not escape from the local optimum when it converged. After the addition of our method, the network successfully converges to the better optimum, thereby improving the accuracy.

3.4. Qualitative Results

We visualize the prediction results in Figure 4, and we can find that our proposed method can produce a smaller and more precise heatmap output on the MUSIC dataset. Even in cases where the comparative algorithms fail to detect musical instruments, the algorithm we propose is still capable of detecting the position of the instruments, further confirming the effectiveness of our method. We hypothesize that, due to the frozen audio network and a large amount of redundant information that can be discarded in the image field, we keep the audio representation still and the network achieves better effect by adjusting image learning.

3.5. Alation Study

From the results shown in Figure 5, it can be observed that if only our algorithm is used for pre-training involving audio, the accuracy of the final network gradually approaches that of the DSOL algorithm, reaching only a 94.3% accuracy. A deep analysis of the curve reveals that after the first epoch of training, the accuracy of not freezing the audio extractor starts to decline compared to that of the proposed algorithm. This is because the performance of the audio extractor is affected, leading to an overall decrease in performance. The results eventually converging near the accuracy of DSOL also confirm the “weak listening” phenomenon mentioned above, where regardless of whether the audio extractor is involved in pre-training or not, the network tends to be biased towards visual dominance, suppressing the contribution of audio information. This validates the effectiveness of our algorithm once again.
The comparative results of the ablation experiments further demonstrate the superiority of the proposed algorithm and emphasize the importance of audio features. It is necessary to enhance their contribution by freezing the audio extractor in order to achieve better localization accuracy.

3.6. Additional Study

In the current situation, due to the large storage space occupied by large video datasets, there is no fixed packaging method available for download. Thus, the number of samples and the distribution of datasets that each person can obtain vary, which leads to biases in experimental results.
Figure 6 shows the number of samples in each category after removing five test samples from each category. It can be observed that the accordion, guitar, and cello have 50, 54, and 49 different samples, respectively, while the saxophone and xylophone only have 18 and 30 samples. This significant sample difference inevitably leads to dataset imbalance and can also affect the accuracy of the final results.
In response to the complexity of multimodal data augmentation, we attempted the following three experimental approaches to enhance the modal-specific features. By comparing these three different data augmentation methods, the optimal strategy can be determined, and further improvements can be made in the performance and accuracy of the multimodal sound source localization task.
Method A: Different frames are selected as image inputs within the same video while still using the original audio corresponding to the video for data augmentation.
Method B: The same frame of the video is selected as the image input, but the most similar audio to the original audio is found using Equation (6) as a replacement.
Method C: Different frames of the video are used as image inputs, and the most similar audio is found using Equation (6) as the corresponding audio.
Based on the results shown in Figure 7 and Table 2, the CIoU under the three data augmentation methods and the comparative results can be observed. In Figure 7, the CIoU values for different categories under the three methods are depicted. Table 2 provides detailed data comparisons for the categories with the highest and lowest values, as well as the overall results.
For Method A, overall performance is inferior to that of Methods B and Method C, indicating that augmenting the image with the original audio yields inferior results compared to replacing the audio. This aligns with the expectations for traditional data augmentation methods, which do not apply well to multimodal tasks. The results demonstrate that merely enhancing the image while keeping the audio unchanged does not achieve optimal performance.
In comparison, there is little difference between Method B and Method C, but Method C slightly outperforms Method B. This suggests that better performance can be achieved by leveraging the diversity of images and audio and maintaining their correspondence. Method C, which comprehensively considers the characteristics of both image and audio in multimodal data augmentation, achieves favorable results.

4. Discussion

We investigate and explore the task of sound source localization in audio-visual correspondence learning, and we propose a self-supervised sound promotion method and achieves remarkable results. However, the algorithm we propose still has some limitations. The following are the issues that are worth further research and discussion: a. consideration of visual information temporal coherence factor; b. limitations of feature extraction algorithms. It is hoped that more and more researchers can enter this field to explore and help make the tasks related to audio-visual correspondence learning more universal and robust.

5. Conclusions

In this paper, we mine the inherent audiovisual contribution to the result and present a self-supervised sound promotion method to tackle the problem of low sound contribution. We validate our method on benchmarks showing great improvement. Experimental results indicate that our approach does increase the proportion of sound in the result, which also improves the accuracy of sound localization, as shown in Figure 4. Furthermore, future work should put emphasis on the more specific balance between vision and audio, because we believe in the following: the more balanced, the more accurate.

Author Contributions

Conceptualization, Y.L.; methodology, Y.L.; software, Y.L.; validation, Y.L.; formal analysis, Y.L.; investigation, Y.L.; resources, Y.L.; data curation, Z.Z.; writing—original draft preparation, Y.L.; writing—review and editing, X.Z.; visualization, Y.L.; supervision, Y.L.; project administration, Y.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Scientific and technological innovation 2030—major project of new generation artificial intelligence grant number 2020AAA0109300.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

References

  1. Gabbay, A.; Ephrat, A.; Halperin, T.; Peleg, S. Seeing through noise: Visually driven speaker separation and enhancement. In Proceedings of the 2018 IEEE International Conference on Acoustics, Speech and Signal Processing, Calgary, AB, Canada, 15–20 April 2018; pp. 3051–3055. [Google Scholar]
  2. Ephrat, A.; Mosseri, I.; Lang, O.; Dekel, T.; Wilson, K.; Hassidim, A.; Freeman, W.T.; Rubinstein, M. Looking to listen at the cocktail party: A speaker-independent audiovisual model for speech separation. ACM Trans. Graph. 2018, 37, 112.1–112.11. [Google Scholar] [CrossRef]
  3. Lu, R.; Duan, Z.; Zhang, C. Listen and look: Audio–visual matching assisted speech source separation. IEEE Signal Process. Lett. 2018, 25, 1315–1319. [Google Scholar] [CrossRef]
  4. Morrone, G.; Bergamaschi, S.; Pasa, L.; Fadiga, L.; Tikhanoff, V.; Badino, L. Face landmark-based speaker-independent audio-visual speech enhancement in multi-talker environments. In Proceedings of the International Conference on Acoustics, Speech, and Signal Processing, Brighton, UK, 12–17 May 2019; pp. 6900–6904. [Google Scholar]
  5. Arandjelovic, R.; Zisserman, A. Objects that sound. In Proceedings of the European Conference on Computer Vision, Munich, Germany, 8–14 September 2018; pp. 451–466. [Google Scholar]
  6. Tian, Y.P.; Shi, J.; Li, B.C.; Duan, Z.Y.; Xu, C.L. Audio-visual event localization in unconstrained videos. In Proceedings of the European Conference on Computer Vision, Munich, Germany, 8–14 September 2018; pp. 252–268. [Google Scholar]
  7. Owens, A.; Efros, A.A. Audio-visual scene analysis with self-supervised multisensory features. In Proceedings of the European Conference on Computer Vision, Munich, Germany, 8–14 September 2018; pp. 639–658. [Google Scholar]
  8. Petridis, S.; Li, Z.; Pantic, M. End-to-end visual speech recognition with lstms. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing, New Orleans, LA, USA, 5–9 March 2017; pp. 2592–2596. [Google Scholar]
  9. Wand, M.; Koutn, J.; Schmidhuber, J. Lipreading with long short-term memory. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing, Shanghai, China, 20–25 March 2016; pp. 6115–6119. [Google Scholar]
  10. Chung, J.S.; Senior, A.W.; Vinyals, O.; Zisserman, A. Lip reading sentences in the wild. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 3444–3453. [Google Scholar]
  11. Trigeorgis, G.; Ringeval, F.; Brueckner, R.; Marchi, E.; Nicolaou, M.A.; Schuller, B.; Zafeiriou, S. Audio features? end-to-end speech emotion recognition using a deep convolutional recurrent network. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing, Shanghai, China, 20–25 March 2016; pp. 5200–5204. [Google Scholar]
  12. Suris, D.; Duarte, A.; Salvador, A.; Torres, J.; Nieto, X.G. Cross-modal embeddings for video and audio retrieval. In Proceedings of the European Conference on Computer Vision, Munich, Germany, 8–14 September 2018; pp. 711–716. [Google Scholar]
  13. Hong, S.; Im, W.; Yang, H.S. Deep learning for content-based, cross-modal retrieval of videos and music. arXiv 2017, arXiv:1704.06761. [Google Scholar]
  14. Arandjelovic, R.; Zisserman, A. Look, listen and learn. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 609–617. [Google Scholar]
  15. Senocak, A.; Oh, T.H.; Kim, J.; Yang, M.H.; Kweon, I.S. Learning to localize sound source in visual scenes. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 1605–1619. [Google Scholar]
  16. Zhao, H.; Gan, C.; Rouditchenko, A.; Vondrick, C.; McDermott, J.H.; Torralba, A. The Sound of Pixels. In Proceedings of the European Conference on Computer Vision, Munich, Germany, 8–14 September 2018; pp. 587–604. [Google Scholar]
  17. Hu, D.; Nie, F.P.; Li, X.L. Deep multimodal clustering for unsupervised audiovisual learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 9248–9257. [Google Scholar]
  18. Lin, Y.B.; Tseng, H.Y.; Lee, H.Y.; Lin, Y.Y.; Yang, M.H. Unsupervised Sound Localization via Iterative Contrastive Learning. CVIU 2023, 227, 103602. [Google Scholar] [CrossRef]
  19. Chen, H.L.; Xie, W.D.; Afouras, T.; Nagrani, A.; Vedaldi, A.; Zissermanin, A. Localizing Visual Sounds the Hard Way. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Virtual, 19–25 June 2021; pp. 16867–16876. [Google Scholar]
  20. Song, Z.J.; Song, Z.J.; Wang, Y.X.; Fan, J.S.; Tan, T.N.; Zhang, Z.X. Self-Supervised Predictive Learning: A Negative-Free Method for Sound Source Localization in Visual Scenes. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 3222–3231. [Google Scholar]
  21. Senocak, A.; Ryu, H.; Kim, J.; Kweon, I.S. Less Can Be More: Sound Source Localization with a Classification Model. In Proceedings of the 2022 IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 3–8 January 2022; pp. 577–586. [Google Scholar]
  22. Zhao, H.; Gan, C.; Ma, W.C.; Torralba, A. The sound of motions. In Proceedings of the IEEE International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 1735–1744. [Google Scholar]
  23. Hu, D.; Qian, R.; Jiang, M.Y.; Tan, X.; Wen, S.L.; Ding, E.; Lin, W.Y.; Dou, D.J. Discriminative Sounding Objects Localization via Self-supervised Audiovisual Matching. In Proceedings of the NeurIPS, Virtual, 6–12 December 2020; pp. 10077–10087. [Google Scholar]
  24. Gao, R.H.; Grauman, K. Co-separating sounds of visual objects. In Proceedings of the IEEE International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 3878–3887. [Google Scholar]
  25. He, K.M.; Zhang, X.Y.; Ren, S.Q.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 770–778. [Google Scholar]
Figure 1. Sound Promotion Method (SPM) is in the dashed frame: we cluster audio information features separately. Then, we use the pseudo-label of sound to train the backbone of audio.
Figure 1. Sound Promotion Method (SPM) is in the dashed frame: we cluster audio information features separately. Then, we use the pseudo-label of sound to train the backbone of audio.
Electronics 12 03558 g001
Figure 2. Category prediction accuracy. The accuracy of the audio classifier increases slowly, and the speed is far lower than that of the image classifier.
Figure 2. Category prediction accuracy. The accuracy of the audio classifier increases slowly, and the speed is far lower than that of the image classifier.
Electronics 12 03558 g002
Figure 3. Training precision. Our method is close to the global optimal point.
Figure 3. Training precision. Our method is close to the global optimal point.
Electronics 12 03558 g003
Figure 4. Qualitative Results on MUSIC dataset.
Figure 4. Qualitative Results on MUSIC dataset.
Electronics 12 03558 g004
Figure 5. Ablation study on MUSIC dataset.
Figure 5. Ablation study on MUSIC dataset.
Electronics 12 03558 g005
Figure 6. Quantity of each category.
Figure 6. Quantity of each category.
Electronics 12 03558 g006
Figure 7. Comparisons of all classes.
Figure 7. Comparisons of all classes.
Electronics 12 03558 g007
Table 1. Quantitative results on the MUSIC testset. Note that the CIoU reported in this table is [email protected], and our method outperforms the existing method on the test set.
Table 1. Quantitative results on the MUSIC testset. Note that the CIoU reported in this table is [email protected], and our method outperforms the existing method on the test set.
MethodsCIoU(0.3)AUC
Sound-of-Pixel [18]8.111.8
Object-Sound [5]3.710.2
DMC [17]7.016.3
DSOL [23]30.622.4
Ours33.424.7
Table 2. Comparison of three data enhancement methods.
Table 2. Comparison of three data enhancement methods.
MethodsAccordionAcoustic_GuitarSaxophoneXylophoneAll _Classes
A35.035.433.633.734.4
B35.736.133.834.335.6
C36.235.834.134.535.8
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, Y.; Zhao, X.; Zhang, Z. Self-Supervised Sound Promotion Method of Sound Localization from Video. Electronics 2023, 12, 3558. https://doi.org/10.3390/electronics12173558

AMA Style

Li Y, Zhao X, Zhang Z. Self-Supervised Sound Promotion Method of Sound Localization from Video. Electronics. 2023; 12(17):3558. https://doi.org/10.3390/electronics12173558

Chicago/Turabian Style

Li, Yang, Xiaoli Zhao, and Zhuoyao Zhang. 2023. "Self-Supervised Sound Promotion Method of Sound Localization from Video" Electronics 12, no. 17: 3558. https://doi.org/10.3390/electronics12173558

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop