Next Article in Journal
Evidence Supporting Open-Loop Control During Early Vergence
Previous Article in Journal
17th European Conference on Eye Movements 11–16 August 2013, Lund, Sweden
 
 
Journal of Eye Movement Research is published by MDPI from Volume 18 Issue 1 (2025). Previous articles were published by another publisher in Open Access under a CC-BY (or CC-BY-NC-ND) licence, and they are hosted by MDPI on mdpi.com as a courtesy and upon agreement with Bern Open Publishing (BOP).
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Different Types of Sounds Influence Gaze Differently in Videos

by
Guanghan Song
1,
Denis Pellerin
1 and
Lionel Granjon
2
1
GIPSA-Lab, CNRS UMR 5216, Grenoble University, France
2
LPP, CNRS UMR 8158, Paris Descartes University, France
J. Eye Mov. Res. 2013, 6(4), 1-13; https://doi.org/10.16910/jemr.6.4.1
Published: 12 October 2013

Abstract

:
This paper presents an analysis of the effect of different types of sounds on visual gaze when a person is looking freely at videos, which would be helpful to predict eye position. In order to test the effect of sound, an audio-visual experiment was designed with two groups of participants, with audio-visual (AV) and visual (V) conditions. By using statistical tools, we analyzed the difference between eye position of participants with AV and V conditions. We observed that the effect of sound is different depending on the kind of sound, and that the classes with human voice (i.e. speech, singer, human noise and singers) have the greatest effect. Furthermore, the results of the distance between sound source and eye position of the group with AV condition, suggested that only particular types of sound attract human eye position to the sound source. Finally, an analysis of the fixation duration between AV and V conditions showed that participants with AV condition move eyes more frequently than those with V condition.

Introduction

In daily life, humans receive a large quantity of information about the environment through sight and hearing. The fast processing of this information helps us to react rapidly and properly. Hence, there exists a mechanism in the brain to direct attention towards particular regions or events, called salient regions or events. This attentional bias is not only influenced by visual and auditory information separately, but is also influenced by audio-visual interaction.
From psychophysical studies, we know that humans react faster to overlapping bimodal audio-visual stimuli than to unimodal (audio or visual) stimuli (Corneil, van Wanrooij, Munoz, & Van Opstal, 2002 ; Sinnett, Faraco, & Spence, 2008). The studies on audio-visual interaction concentrate on two areas: the influence of visual input on auditory perception and the influence of acoustic input on visual perception.
Early evidence of the influence of visual input on auditory perception is the “McGurk Effect”. The “McGurk Effect” is a phenomenon that demonstrates a perceptual fusion between auditory and visual (lip-reading) information in speech perception. In this experiment a film of a young woman repeating utterances of the syllable [ba] was dubbed on to lip movements for [ga]: normal adults reported hearing [da] (McGurk& MacDonald, 1976). This “McGurk Effect” works with perceivers of all language backgrounds (Cohen & Massaro, 1994), and it also works on young infants (Rosenblum, Schmuckler, & Johnson, 1997). Another well-known audio-visual interaction is that visual “lip-reading” helps speech to be understood, when speech is in poor acoustical conditions or in a foreign language (Jeffers & Barley, 1971 ; Summerfield, 1987).
Speech is a special audio stimulus: numerous current studies are focused on audio-visual interaction of speech (Alho et al., 2012). A study from (Tuomainen, Andersen, Tiippana, & Sams, 2005) provided evidence of the existence of a specific mode of multi-sensory speech perception. More recently, some observations of the mechanisms of speech stimuli and visual interaction have demonstrated that lip-read information was more strongly paired with speech information than non-speech information (Vroomen & Stekelenburg, 2011). Other types of sound have been investigated less.
Auditory cues also influence visual perception. Previous studies showed that when auditory and visual signals come from the same location, the sound can guide attention toward a visual target (Perrott, Saberi, Brown, & Strybel, 1990; Spence & J.Driver, 1997). Moreover, other studies demonstrated that synchronous auditory and visual events can improve visual perception (Vroomen & De Gelder, 2000; Dalton & Spence, 2007). Another study considered the situation in which audio and visual information do not come from the same spatial place. The result showed that the synchronous sound “pip” makes the visual object pop out from its complex environment phenomenally (Van der Burg, Olivers, & Bronkhorst, 2008).
Inspired by these studies of the influence of audiovisual interaction on human behavior, computer scientists have tried to simulate this attentional mechanism to create a computational attention model, which helps to select important objects from a mass of information. This computational attention model provides another way to better understand the attentional mechanism. Furthermore, these computational attention models are useful for applications such as video coding (Lee, Simone, & Ebrahimi, 2011) and video summarizing (Wang & Ngo, 2012).
Studies in cognitive neurosciences show that eye movements are tightly linked to visual attention (Awh, Armstrong, & Moore, 2006). The study of eye movements enables a better understanding of the visual system and the mechanisms in our brain to select salient regions. Furthermore, eye movements also represent the influence of audio-visual interaction on human behavior. Quigley and her colleagues (Quigley, Onat, Harding, Cooke, & König, 2008) investigated how different locations of sound source (played by loudspeakers in different locations: left, right, up and down) influence eye movement in static images (Onat, Libertus, & König, 2007). The results showed that eye movements were spatially biased towards the regions of the scene corresponding to the location of the loudspeakers. Auditory influences on visual location also depend on the size of the visual target (Heron, Whitaker, & McGraw, 2004). In videos, during dynamic face viewing, sound influences gaze to different face regions (Vö, Smith, Mital, & Henderson, 2012).
Although the interaction of features within audio and visual modalities has been actively studied, the sound effect on human gaze when looking at videos with their original soundtrack has been explored less. Our previous research (Song, Pellerin, & Granjon, 2011) showed that sound affects human gaze differently depending on the type of sound, and the effect is greater for the on-screen speech class (the speakers appear on screen) rather than the non-speech class (any kind of audio signal other than speech) and the non-sound class (intensity below 40 dB). Recently, (Coutrot, Guyader, Ionescu, & Caplier, 2012) showed that original soundtrack of videos impacts on eye position, fixation duration and saccade amplitude, and (Vilar et al., 2012) using non-original soundtrack also concluded that sound affects human gaze.
In our previous research, we only considered three sound classes and no strict control of sound event over time. In this paper, we provide a deeper investigation of the question of which type of sound influences human gaze with a controllable sound. A preliminary analysis was published in (Song, Pellerin, & Granjon, 2012). We first describe an audio-visual experiment with two groups of participants: with original soundtrack called audiovisual (AV) condition; and without sound called visual (V) condition. Then, we observe the difference of eye position between two groups of participants for thirteen more refined sound classes. The fixation duration between groups with AV and V conditions is also studied.

Method

Participants

Thirty-six human participants (18 women and 18 men, aged from 20 to 34) took part in the experiment. 18 participants first viewed 5 clips with V condition, then viewed another 5 clips with AV condition. The other 18 participants first viewed 5 clips with AV condition, then viewed another 5 clips with V condition. Each clip appeared with AV and V condition in the same number of occurrences. All participants had normal or corrected-to-normal vision, and reported normal hearing. They were ignorant to the purpose of the experiment.

Materials

In this experiment, eighty video excerpts (called clip snippets) were chosen from heterogeneous sources of films (with original soundtrack). Each clip snippet lasted around 200 frames (8 seconds). The sum of all the clip snippets represents 16402 frames (around 11 minutes). In the visual domain, each clip snippet consists of just one shot. In the audio domain, the sound signal is divided into two parts. The first sound lasts to about the middle of the clip snippet, and is then followed by the second sound. In order to prevent the participants from understanding the language in the video, we chose foreign languages for each participant, such as Chinese, Indian, and Japanese etc. An example of a clip snippet is presented in Figure 1. The 80 clip snippets were then recombined into 10 clips (Carmi & Itti, 2006), each clip being the concatenation of 8 clip snippets from different film sources and different sound classes of the second sound. The sound amplitude of each clip snippet was “normalized” in order to reduce the amplitude transition between two successive clip snippets, so as to partially preserve the original sound amplitude information. Normalization is defined as follows:
Jemr 06 00012 i001
where, Ai is the mean amplitude of each clip snippet, and A ¯ is the mean of the mean amplitude Ai for an entire clip. All the clip snippets were converted in gray level to the same video format (25 fps, 842×474 pixels/frame). Two sets of stimuli were built from these clips, one with AV condition (frames + soundtrack), and the other one with V condition (frames only).
We only observe the behavior of human gaze after the onset of the second sound. The aim is to analyze the effect of an audio change unrelated to the visual changes that occur when a new clip snippet starts. Furthermore, the first sound lasts at least two seconds before the second sound occurs, which is enough to avoid center bias (Tseng, Carmi, Cameron, Munoz, & Itti, 2009). This so called “center bias” is the relocation of gaze near the center of the frame to collect information about the new visual scene. It influences eye movement when viewing dynamic scenes (Dorr, Martinetz, Gegenfurtner, & Barth, 2010).
We classify the second sound into thirteen classes (see Figure 2), based on other research (Niessen, Maanen, & Andringa, 2008). For each class, there are 5 to 11 clip snippets. The numbers of clip snippets and frames in each class are given in Table 1.
The difference between clusters of classes “on-screen with one sound source” and “on-screen with more than one sound source” is the number of sound sources on the screen. Here, we call one sound source a visual event in the scene associated with the soundtrack. In this instance the sound can be associated with a spatial location. The “off-screen sound source” cluster is different from the other two in that there is no sound source on the screen when the second sound appears.
Table 1. Number of clip snippets and frames in each class. 
Table 1. Number of clip snippets and frames in each class. 
Jemr 06 00012 i003
Before this experiment, we did a pre-experiment to validate the classification of the second sound. In this pre-experiment, the participant heard only the second sound from the headphone, then chose the sound class from the proposed list (Figure 2). If the participants chose the same class as we proposed, we considered it as a correct recognition. 5 participants took part in this experiment. For each class, the minimal correct classification rate is 80%, and the mean correct classification rate is 90%. Hence, we can conclude that the classification is suitable for an audio-visual experiment.
Figure 2. Hierarchical classification of the second sound, including 3 sound clusters, which consist of 13 sound classes. 
Figure 2. Hierarchical classification of the second sound, including 3 sound clusters, which consist of 13 sound classes. 
Jemr 06 00012 g002

Procedure

Human eye position was tracked by an Eye tracker Eyelink II (SR Research). The clips were shown by SoftEye (a software tool) synchronized with the eye tracker (Ionescu, Guyader, & Guérin-Dugué, 2009). During the experiment, the participants were sitting in front of a 19-inch color monitor (60 Hz refresh rate) with their chin supported. The viewing distance between the participant and the monitor was 57 cm. The usable field of vision was 35 ° × 20 ° . A headphone carried the stereo sound. The apparatus did not allow participants to locate the sound source on the small size of the video because of the stereo headphone. A 9-point calibration was carried out every five clips. Before each clip, we presented a drift correction, then a fixation in the center of the screen. Figure 3 illustrates the time course of this experimental trial. Participants were asked to look at the ten clips without any particular task. All ten clips were presented to each participant in random order. As said above, each participant watched half the clips with AV condition and half the clips with V condition.

Metrics

Kullback-Leibler divergence. In order to measure the difference of eye position between two groups (with AV and V conditions) for each frame, a metric named Kullback-Leibler divergence was calculated. The Kullback-Leibler divergence metric was already adopted to compare distributions of eye position between groups by other researchers, such as (Tatler, Baddeley, & Gilchrist, 2005). For a given frame, a 2-D Gaussian was added to each eye position to build the density map of a group of participants with AV condition ( M h m ), respectively with V condition ( M h v ). The standard deviation of the Gaussian was chosen to have a diameter at half the height of the Gaussian equal to 2°of visual angle. Here, we use symmetric Kullback-Leibler divergence ( K L ). For each frame, we calculated the following equation:
Jemr 06 00012 i002
where q represents the same size of video frame (842×474 pixels). High K L values represent large differences between two distributions of eye position.
Linear Correlation Coefficient. To confirm the measurement, two other metrics are adopted. One is the linear correlation coefficient, noted as c c . The c c describes the linear relationship between the respective probability densities of two data sets. It is defined as follows:
c c ( M h a v , M h v ) = c o v ( M h a v , M h v ) σ M h a v σ M h v
where, M h a v (respectively M h v ) represents the eye position density maps with the AV (respectively V) condition, c o v ( M h a v , M h v ) is the covariance value between M h a v and M h v .
For c c , a value of zero indicates no linear relationship between the two maps: there is no correspondence between the eye position of the two groups with AV and V conditions, and higher values of c c indicate higher correspondence between the eye positions of the two groups.
Median distance. The other metric we adopted is median distance m d . It is defined as:
m d = m e d i a n ( d i , j ) , i C , j C
where, C is the group with AV condition and C′ is the group with V condition. d i , j is the Euclidean distance between eye positions of participants i and j , who belong respectively to the group with AV condition and the group with V condition.
Distance to the sound source. From observation, we notice that participants with AV condition seemed to move their eye to the sound source after the beginning of the second sound. To verify this assumption, we located the approximate coordinates of the center of the sound source manually. Then, the Euclidean distance between the eye position of each participant with AV condition and the sound source was calculated. The mean of these Euclidean distances gives the D value, which represents the distance of eye position to the sound source. A high D value represents a large distance from the sound source to eye position of participants from one group.

Results

In order to investigate the effect of sound on visual gaze, we analyzed the difference of eye position between participants with AV condition and with V condition. Figure 4 (a) shows an example of the eye positions of two groups of participants. Figure 4 (b) shows an example of the density map of groups of participants with AV condition ( M h m ) and with V condition ( M h v ).

Comparison among three clusters of sound classes

We analyzed the Kullback-Leibler divergence (KL) between the eye position of the participants in two groups with AV and V conditions, among three clusters of classes (see Figure 2): “on-screen with one sound source”, “on-screen with more than one sound source” and “off-screen sound source”.
In this section, for each clip snippet, we investigated one second after the beginning of the second sound (from frame 6 to 30, to eliminate reaction time of about 5 frames). We used the ANOVA test to compare KL among different clusters of classes. This test requires the samples in each cluster to be independent samples. Because we consider continuous measurement over time, the eye position for most participants does not change much between two adjacent frames, they could not be considered as independent samples. To solve this problem, we took the mean of KL values over one second (from frame 6 to 30 after the beginning of the second sound) as one independent sample.
Figure 4. A sample frame in singer class of eye position of participants in groups with AV (red points) and with V condition (blue points), and corresponding density map of groups of participants with AV condition Mhav (red region) and with V condition Mhv (blue region). 
Figure 4. A sample frame in singer class of eye position of participants in groups with AV (red points) and with V condition (blue points), and corresponding density map of groups of participants with AV condition Mhav (red region) and with V condition Mhv (blue region). 
Jemr 06 00012 g004
In Figure 5, with the ANOVA test, “off-screen sound source” presents the lowest KL among the three clusters of classes. The difference is significant between “on-screen with one sound source” and “off-screen sound source” (F(1,63)=4.72, p=0.034), and also significant between “on-screen with more than one sound source” and “off-screen sound source” (F(1,25)=4.67, p=0.041). The difference between “on-screen with one sound source” and “on-screen with more than one sound source” is not significantly different (F(1,69)=0.03, p=0.859). These results indicate that one and more than one localizable sound sources lead to a greater distance between the groups with AV and V conditions compared to non-localizable sound source.
The results above were confirmed by two other metrics: cc and md.
To verify that the effect measured is really due to the second sound, we performed the same calculation for a period of one second (25 frames, from frame -24 to 0) before the transition from first sound to second sound for all the classes. This “pre-transition” cluster (in Figure 5) can be considered as a baseline, compared to the three other clusters. The difference is significant between “on-screen with one sound source” and “pre-transition” (F(1,133)=9.09, p=0.0031), and also significant between “on-screen with more than one sound source” and “pre-transition” (F(1,95)=4.65, p=0.034). The difference is not significant between “off-screen sound source” and “pre-transition” (F(1,89)=0.01, p=0.915). These results show that one and more than one localizable sound sources for the second sound lead to a greater distance between the groups with AV and V conditions compared to pre-transition (first sound).
Figure 5. Kullback-Leibler divergence (KL) between participants with AV and V conditions in three clusters of classes: “on-screen with one sound source”, “on-screen with more than one sound source” and “off-screen sound source”, and compared to the “pre-transition” cluster. Larger KL values represent greater difference between groups with AV and V conditions. 
Figure 5. Kullback-Leibler divergence (KL) between participants with AV and V conditions in three clusters of classes: “on-screen with one sound source”, “on-screen with more than one sound source” and “off-screen sound source”, and compared to the “pre-transition” cluster. Larger KL values represent greater difference between groups with AV and V conditions. 
Jemr 06 00012 g005
To complete the previous study, we analyzed entropy variation between before and after sound transition. More precisely, in AV condition (respectively V condition) for each clip snippet, we calculated the mean of entropy for a period of one second after the transition (from frame 6 to 30) and subtracted the mean of entropy for one second before the transition (from frame -24 to 0). Then, we compared the results of entropy variation between the two conditions (AV and V) by using paired t-test. For “on-screen with one sound source” and for “on-screen with more than one sound source” clusters, the mean of entropy variation is significantly larger in AV condition compared to V condition (respectively t(53)=2.95, p=0.004 and t(15)=2.52, p=0.023) (Figure 6). Participants with AV condition are not only attracted by salient regions from visual aspect, such as face, motion regions, but also attracted by sound sources from audio aspect. For these two clusters, the entropy variation is negligible in V condition. For “off-screen sound source”, the entropy variation is not significantly different between AV and V conditions (t(9)=0.84, p=0.42) (Figure 6). In this case, participants with AV condition modify their behavior slightly compared with V condition.

Analysis of thirteen sound classes

We analyzed the thirteen sound classes separately. We did not analyze sound effect directly through audio information, but through the eye position of participants which are also based on visual information. In order to reduce the influence of visual information, we created a baseline for statistical comparison by performing a randomization (Edgington & Onghena, 2007): We fused two groups of participants with AV and V conditions into one set of 36 participants. We extracted 18 participants from this set randomly to create a new group called G1. The rest of the participants formed another new group, called G2. Afterwards, we calculated the KL between G1 and G2 for each frame. We repeated this procedure 5000 times, obtaining for each frame a distribution of 5000 random KL values (KLi, i=1,2,…5000). Then, we took the mean of the 5000 KL values as the baseline ( K L R ). This K L R , which is influenced by image only, is an estimate of the KL that can be expected between two random groups of participants. Finally, we calculated the difference ( K L A V V K L R ) where K L A V V represents the difference between participants with AV and V conditions. Because K L A V V is caused by the effect of both image and sound, and K L R is caused by the effect of image only, the difference ( K L A V V K L R ) is mainly caused by the effect of sound.
Figure 7 shows the difference over time between K L A V V and K L R for two classes: “speech” (human) and “impact and explosion” (non-human). If ( K L A V V K L R ) is above 0, the difference between AV and V groups is greater than that between two random groups. The behavior over time is different for two presented sound classes.
Jemr 06 00012 i004
Table 2 shows the results for frames 6 to 30 after the beginning of the second sound. The high K L ¯ A V V values (therefore low p values) for the marked classes (with ■): speech, singer, human noise, and singers, show that human voice affects visual gaze significantly (p<0.05).
To verify that the effect measured above is really due to the second sound, we perform the same calculation for a period of one second (25 frames) before the beginning of the second sound. Results of probability estimations of K L ¯ i values higher than K L ¯ A V V of all the sound classes from frame -24 to 0 are higher than 0.1, suggesting that before the second sound, eye position of participants between groups with AV and V conditions are not significantly different for all the sound classes.
The results above were confirmed by other two metrics: cc and md.

Analysis of distance between sound source and eye position

In the previous section, we showed that the Kullback-Leibler divergence K L between eye position of participants with AV and V conditions is greater for speech, singer, human noise and singers classes than others. In this section, we want to verify the assumption that participants with AV condition moved their eye to the sound source after the beginning of the second sound. We only analyzed the “on-screen with one sound source” cluster of sound classes. We first located the approximate coordinates of the center of the sound source manually. Then, we calculated the Euclidean distance between the eye position of each participant with AV condition and the sound source. The mean of these Euclidean distances gives the D A V S value, which is affected by both image and sound information. Similarly, in order to reduce the influence of visual information, we created a baseline for statistical comparison by performing a randomization (Edgington & Onghena, 2007). We considered the mean Euclidean distance between eye position of participants of G1 (consists of 18 participants, which are randomly selected from the set of all the participants in groups with AV and V conditions) and sound source (Di, i=1,2…5000). We took the mean of 5000 distance values as the baseline ( D R ), which was affected only by image information. Afterwards, for each frame, we calculated D A V S D R for all the classes with one sound source. This difference reflects the influence of the sound information.
Figure 8 shows the difference over time between D A V S and D R for “speech” and “impact and explosion” classes. When the values are negative, the group with AV condition is closer to the sound source than the random group. Again, different sound classes behave differently.
To find out which classes give the higher difference between DAVS and DR and quantify the sound effect, we investigate the same duration of one second (25 frames) as in previous analysis, from frame 6 to 30 after the beginning of the second sound. We compared D ¯ A V S (the mean of DAVS over the 25 frames) to the distribution of D ¯ i (i=1,2,...5000), where D ¯ i is the mean of Di between G1 and sound source over the 25 frames for the random trial i. To estimate the probability of D ¯ i being smaller than D ¯ A V S , we calculate p=n/5000, where n is the number of D ¯ i which are smaller than D ¯ A V S .
Figure 8. Average difference (DAVS - DR) over time for “speech” and “impact and explosion” classes. Dark regions represent (DAVS - DR) below 0, suggesting that the group with AV condition is closer to the sound source than the random group. 
Figure 8. Average difference (DAVS - DR) over time for “speech” and “impact and explosion” classes. Dark regions represent (DAVS - DR) below 0, suggesting that the group with AV condition is closer to the sound source than the random group. 
Jemr 06 00012 g008
In Table 3, D ¯ i is smaller than D ¯ A V S (p<0.05), from frame 6 to 30 after the beginning of the second sound, for speech, singer, human noise classes (marked with ■) suggesting that participants tend to move their eyes to the sound source only when they hear human voice.

Analysis of musical instrument subclass

Compared to human voice classes, which have been well discussed in recent decades, music class has been explored less. To better understand the influence of audio-visual interaction, we propose a deeper investigation of eye movement behavior of music class. In our music class database, four snippets are humans playing musical instruments. They represent the musical instrument subclass. In this subclass, there is more than one face in the scene. However, only one person is playing an instrument (piano or guitar), when the corresponding music begins.
In musical instrument subclass, what is more attractive to the participants? There is evidence that faces in the scene are preferred by the visual system compared to other object categories (Rossion et al., 2000 ; Langton, Law, Burton, & Schweinberger, 2008), and can be processed at the earliest stage after stimulus presentation (Ro, Russell, & Lavie, 2001). From our observation, we assume that a particular face -- Face of the player attracts more attention than other faces. In previous calculations, we know that the sound source in the scene was attractive for participants with AV condition in human voice sound classes. In the musical instrument subclass, do participants have a preference for sound source, that is, the Musical instrument?
To measure which region (musical instrument or the face of the player) is more attractive to the participants, we calculate the Euclidean distance between the eye position of participants with AV condition and Musical instrument ( D A V M ). Respectively, we calculate the Euclidean distance between the eye position of participants with AV condition and the Face of the player ( D A V F ). Again, we introduce a baseline D R M , which is the mean Euclidean distance between random group G1 and Musical instrument for 5000 randomization times. Respectively, the baseline D R F is the mean Euclidean distance between random group G1 and the Face of the player.
Figure 9 illustrates the distances from group with AV condition to Musical instrument (a) and to Face of the player (b) over time. Here, the dark regions below zero represent smaller distances from the Face of the player or the Musical instrument. The Face of the player is reached more frequently after the beginning of the music sound until around frame 14. After that, both the Face of the player and the Musical instrument are reached somewhat equally.
To quantify the measurement, we further investigated a period of one second (25 frames), from frame 6 to 30 after the beginning of the second sound. The probability of D ¯ R M i ( D R M i is the mean Euclidean distance between G1 and Musical instrument) being smaller than D ¯ A V M is p=0.164. The probability of D ¯ F i ( D F i is the mean Euclidean distance between G1 and Face of the player) being smaller than D ¯ A V F is p=0.042. The results indicate that during this period of one second, participants move their eyes to the Face of the player rather than the Musical instrument.
Figure 9. Average distances (a) from Musical instrument (DAVM - DRM), (b) from Face of the player (DAVF - DRF) for 4 clip snippets of musical instrument subclass over time. 
Figure 9. Average distances (a) from Musical instrument (DAVM - DRM), (b) from Face of the player (DAVF - DRF) for 4 clip snippets of musical instrument subclass over time. 
Jemr 06 00012 g009

Fixation duration analysis

We also investigated the effect of sound on the distribution of fixation duration for the whole database. It is typical to study such parameters (Tatler, Hayhoe, Land, & Ballard, 2011). For each participant, we calculated the mean of fixation duration for each clip. A traditional method -- paired t-test was adopted. Per clip, AV condition has a shorter average duration of fixation (6.17 frames, 247 ms) than V condition (6.82 frames, 273 ms), and the difference is significant (t(9)=2.479, p=0.035). Per participant, AV condition still has a shorter average duration of fixation (6.19 frames, 248 ms) than V condition (6.75 frames, 270 ms), and the difference is also significant (t(35)=2.697, p=0.011). This means that the participants with AV condition tend to move their eyes more frequently compared to the participants with V condition. Additionally, this result is confirmed by a more recent method -- mixed effect model (Baayen, Davidson, & Bates, 2008).

Discussion

This study demonstrates that not only does human speech have a higher effect on human gaze when looking freely at videos, but also singer(s) and human noise.
The Kullback-Leibler divergence ( K L ) between the groups with AV and V conditions is lower for “off-screen sound source” cluster than for two “on-screen sound source” clusters. The result indicates that a change in auditory information affects human gaze, when the information is linked to a visual event in the video (Hidaka, Teramoto, Gyoba, & Suzuki, 2010) (Gordon & Hibberts, 2011). The reason is perhaps that synchronized audio-visual events capture attention rather than unpaired audio-visual stimuli (Van der Burg, Brederoo, Nieuwenstein, Theeuwes, & Olivers, 2010). The entropy variation between before and after sound transition in AV condition (compared to V condition) shows that eye positions of participants tend to be more dispersed after transition when the sound source(s) is on-screen.
By calculating the difference between K L ¯ A V V (the temporal mean of K L between two groups of participants) and randomization distribution K L ¯ t , we conclude that the difference between participants with AV and V conditions is greater for four human classes (speech, singer, human noise, and singers). To explain this difference, we assume that the participants with AV condition move their eyes to the sound source after the beginning of the second sound. The result of D ¯ A V S (mean of distance between participants with AV condition and sound source) is smaller than D ¯ t (randomization distribution), and implies that after the auditory stimuli, participants searched for the sound source, associated with auditory information in the scene. This kind of behavior is obvious when the auditory stimulus is a human voice. This kind of behavior has also been observed by other researchers, but only for speech class. Kim and colleagues (Kim, Davis, & Krins, 2004 ; Tuomainen et al., 2005) provided evidence that acoustic and visual speech is strongly integrated only when the perceiver interprets the acoustic stimuli as speech. More recently observations of the mechanisms of speech stimuli and visual interaction demonstrated that lip-read information was more strongly paired with speech information than non-speech information (Vroomen & Stekelenburg, 2011).
Temporally, reaction time of participants is also observed. In Figure 7 (a), the K L value between participants with AV and V conditions of “speech” class increases around frame 7. However, in Figure 8 (a), the eye position of participants with AV condition seems to reach the sound source after frame 14. It takes 7 frames on average (280 ms) for a participant to move their eyes to the sound source after hearing the second sound.
Face in the scene not only influences human voice sound classes, but also influences the musical instrument subclass. In this subclass, the distance between the eye position of participants with AV condition and the Face of the player is smaller than the distance between the eye position of participants with AV condition and the Musical instrument. The visual event linked to the acoustic stimuli is the instrument, not the face. The result shows that after the participants hear music, first they tend to move their eyes to the Face of the player. After a while, both the human face and musical instrument are reached. One possible explanation for this behavior is that participants responded faster to social stimuli (like faces) compared to non-social stimuli (like houses) (Escoffier, Sheng, & Schirmer, 2010). This special attractability of the Face of the player among other faces only appears when the music (from a musical instrument) can be heard simultaneously.
The comparison of fixation duration between the groups of participants with AV and V conditions was carried out for the whole database. We observed that the group with AV condition had a shorter fixation duration than the group with V condition. It may be caused by the fact that the responses of the participants to bimodal audio-visual stimuli were significantly faster than unimodal visual stimuli (Sinnett et al., 2008). Recent research from (Zou, Mller, & Shi, 2012) also confirms that synchronous audio-visual stimuli facilitate visual search performance, and have shorter reaction time than visual stimuli only.
In conclusion, our results provide evidence of sound influence on gaze when looking at videos. This sound effect is different depending on the type of sounds. Sound effect can be measured only when the sound is human voice. More precisely, human voice drives participants to move their eyes towards the sound source. In future work, by simulating this eye movement behavior influenced by sound, it would be interesting to add auditory influence to the traditional computational visual saliency model (such as (Itti, Koch, & Niebur, 1998)) to create an audio-visual saliency model. It could help to increase the prediction accuracy when the model is applied to videos with an original soundtrack.

Acknowledgments

This research is supported in part by the Rhône-Alpes region (France) through the LIMA project. The authors would like to thank their colleagues R. Drouilhet for the advices on statistical analysis, and A. Rahman and K. Wang for the helpful comments about this article.

References

  1. Alho, K.; Salonen, J.; Rinne, T.; Medvedev, S. V.; Hugdahl, K.; Hmlinen, H. Attention related modulation of auditory-cortex responses to speech. sounds during dichotic listening. Brain Research 2012, 1442(9), 47–54. [Google Scholar] [CrossRef]
  2. Awh, E.; Armstrong, K. M.; Moore, T. Visual and oculomotor selection: Links, causes and implications for spatial attention. Trends in Cognitive Sciences 2006, 10(3), 124–130. [Google Scholar]
  3. Baayen, R. H.; Davidson, D. J.; Bates, D. M. Mixed-effects modeling with crossed random effects for subjects and items. Journal of Memory and Language 2008, 59, 390–412. [Google Scholar] [CrossRef]
  4. Carmi, R.; Itti, L. Visual causes versus correlates of attentional selection in dynamic scenes. Vision Research 2006, 46, 4333–4345. [Google Scholar] [CrossRef] [PubMed]
  5. Cohen, M. M.; Massaro, D. W. Development and experimentation with synthetic visible speech. Behavior Research Methods, Instruments, and Computers 1994, 26, 260–265. [Google Scholar] [CrossRef]
  6. Corneil, B. D.; van Wanrooij, M.; Munoz, D. P.; Van Opstal, A. J. Auditory-visual interactions subserving goal-directed saccades in a complex scene. Journal of Neurophysiology 2002, 88(1), 438–454. [Google Scholar] [CrossRef]
  7. Coutrot, A.; Guyader, N.; Ionescu, G.; Caplier, A. Influence of soundtrack on eye movements during video exploration. Journal of Eye Movement Research 2012, 5(4), 1–10. [Google Scholar] [CrossRef]
  8. Dalton, P.; Spence, C. Attentional capture in serial audiovisual search tasks. Perception & Psychophysics 2007, 69, 422–438. [Google Scholar]
  9. Dorr, M.; Martinetz, T.; Gegenfurtner, K. R.; Barth, E. Variability of eye movements when viewing dynamic natural scenes. Journal of Vision 2010, 10(28), 1–17. [Google Scholar] [CrossRef]
  10. Edgington, E. S.; Onghena, P. Randomization tests; Chapman Hall, 2007. [Google Scholar]
  11. Escoffier, N.; Sheng, D. Y. J.; Schirmer, A. Unattended musical beats enhance visual processing. Acta Psychologica 2010, 135(1), 12–16. [Google Scholar] [CrossRef]
  12. Gordon, M. S.; Hibberts, M. Audiovisual speech from emotionally expressive and lateralized faces. The Quarterly Journal of Experimental Psychology 2011, 64(4), 730–750. [Google Scholar] [CrossRef]
  13. Heron, J.; Whitaker, D.; McGraw, P. V. Sensory uncertainty governs the extent of audio-visual interaction. Vision Research 2004, 44, 2875–2884. [Google Scholar] [CrossRef] [PubMed]
  14. Hidaka, S.; Teramoto, W.; Gyoba, J.; Suzuki, Y. Sound can prolong the visible persistence of moving visual objects. Vision Research 2010, 50, 2093–2099. [Google Scholar] [CrossRef] [PubMed]
  15. Ionescu, G.; Guyader, N.; Guérin-Dugué, A. Softeye software. IDDN.FR.001.200017.000.S.P.2010.003.31235. 2009. [Google Scholar]
  16. Itti, L.; Koch, C.; Niebur, E. A model of saliency-based visual attention for rapid scene analysis. Transaction on Pattern Analysis and Machine Intelligence 1998, 20(11), 1254–1259. [Google Scholar] [CrossRef]
  17. Jeffers, J.; Barley, M. Speech reading (lipreading); Springfield, 1971. [Google Scholar]
  18. Kim, J.; Davis, C.; Krins, P. Amodal processing of visual speech as revealed by priming. Cognition 2004, 93, B39–B47. [Google Scholar] [CrossRef]
  19. Langton, S. R. H.; Law, A. S.; Burton, A. M.; Schweinberger, S. R. Attention capture by faces. Cognition 2008, 107, 330–342. [Google Scholar] [CrossRef]
  20. Lee, J. S.; Simone, F. D.; Ebrahimi, T. Subjective quality evaluation of foveated video coding using audio-visual focus of attention. IEEE Journal of Selected Topics in Signal Processing 2011, 5(7), 1322–1331. [Google Scholar] [CrossRef]
  21. McGurk, H.; MacDonald, J. Hearing lips and seeing voices. Nature 1976, 264, 746–748. [Google Scholar] [CrossRef]
  22. Niessen, M. E.; van Maanen, L.; Andringa, T. C. Disambiguating sounds through context. IEEE International Conference on Semantic Computing; 2008; pp. 88–95. [Google Scholar]
  23. Onat, S.; Libertus, K.; König, P. Integrating audiovisual information for the control of overt attention. Journal of Vision 2007, 7(10), 1–16. [Google Scholar] [CrossRef] [PubMed]
  24. Perrott, D. R.; Saberi, K.; Brown, K.; Strybel, T. Z. Auditory psychomotor coordination and visual search performance. Perception & Psychophysics 1990, 48, 214–226. [Google Scholar]
  25. Quigley, C.; Onat, S.; Harding, S.; Cooke, M.; König, P. Audio-visual integration during overt visual attention. Journal of Eye Movement Research 1 2008, 1–17. [Google Scholar] [CrossRef]
  26. Ro, T.; Russell, C.; Lavie, N. Changing faces: a detection advantage in the flicker paradigm. Psychological Science 2001, 12(1), 94–99. [Google Scholar] [CrossRef] [PubMed]
  27. Rosenblum, L. D.; Schmuckler, M. A.; Johnson, J. Q. The mcgurk effect in infants. Perception & Psychophysics 1997, 59(3), 347–357. [Google Scholar]
  28. Rossion, B.; Gauthier, I.; Tarr, M. J.; Despland, P.; Bruyer, R.; Linotte, S.; et al. The n170 occipito temporal component is delayed and enhanced to inverted faces but not to inverted objects: an electrophysiological account of face-specific processes in the human brain. Neuroreport 2000, 11(1), 69–74. [Google Scholar] [CrossRef] [PubMed]
  29. Sinnett, S.; Faraco, S. S.; Spence, C. The cooccurrence of multisensory competition and facilitation. Acta Psychologica 2008, 128(1), 153–161. [Google Scholar] [CrossRef]
  30. Song, G.; Pellerin, D.; Granjon, L. Sound effect on visual gaze when looking at videos. 19th European Signal Processing Conference (EUSIPCO); 2011; pp. 2034–2038. [Google Scholar]
  31. Song, G.; Pellerin, D.; Granjon, L. How different kinds of sound in videos can influence gaze. International Workshop on Image Analysis for Multimedia Interactive Services (WIAMIS) 2012, 1–4. [Google Scholar]
  32. Spence, C.; Driver, J. Audiovisual links in exogenous covert spatial orienting. Perception & Psychophysics 1997, 59, 1–12. [Google Scholar]
  33. Summerfield, Q. Dodd, B., Campbell, R., Eds.; Some preliminaries to a comprehensive account of audio-visual speech perception. In Hearing by eye: The psychology of lip-reading; London: Erlbaum, 1987; pp. 3–52. [Google Scholar]
  34. Tatler, B. W.; Baddeley, R. J.; Gilchrist, I. D. Visual correlates of fixation selection: effects of scale and time. Vision Research 2005, 45, 643–659. [Google Scholar] [CrossRef]
  35. Tatler, B. W.; Hayhoe, M. M.; Land, M. F.; Ballard, D. H. Eye guidance in natural vision: Reinterpreting salience. Journal of Vision 2011, 11(5), 1–23. [Google Scholar]
  36. Tseng, P. H.; Carmi, R.; Cameron, I. G. M.; Munoz, D. P.; Itti, L. Quantifying center bias of observers in free viewing of dynamic natural scenes. Journal of Vision 2009, 9(7), 1–16. [Google Scholar] [CrossRef]
  37. Tuomainen, J.; Andersen, T. S.; Tiippana, K.; Sams, M. Audio-visual speech perception is special. Cognition 2005, 96(1), B13–B22. [Google Scholar] [CrossRef] [PubMed]
  38. Van der Burg, E.; Brederoo, G.; Nieuwenstein, M. R.; Theeuwes, J.; Olivers, C. N. Audiovisual semantic interference and attention: Evidence from the attentional blink paradigm. Acta Psychologica 2010, 134(2), 198–205. [Google Scholar] [CrossRef]
  39. Van der Burg, E.; Olivers, C. N. L.; Bronkhorst, A. W. Pip and pop: nonspatial auditory signals improve spatial visual search. Journal of Experimental Psychology: Human Perception and Performance 2008, 34(5), 1053–1065. [Google Scholar] [CrossRef] [PubMed]
  40. Vilar, A.; Duchowski, A. T.; Orero, P.; Grindinger, T.; Tetreault, S.; di Giovanni, E. How sound is the pear tree story? testing the effect of varying audio stimuli on visual attention distribution. Perspectives: Studies in Translatology 2012, 20(1), 55–65. [Google Scholar] [CrossRef]
  41. Võ, M. L. H.; Smith, T. J.; Mital, P. K.; Henderson, J. M. Do the eyes really have it? Dynamic allocation of attention when viewing moving faces. Journal of Vision 2012, 12(13), 1–14. [Google Scholar] [CrossRef]
  42. Vroomen, J.; De Gelder, B. Sound enhances visual perception: cross-modal effects of auditory organization on vision. Journal of Experimental Psychology: Human Perception and Performance 2000, 26, 1583–1590. [Google Scholar] [CrossRef]
  43. Vroomen, J.; Stekelenburg, J. J. Perception of intersensory synchrony in audiovisual speech: not that special. Cognition 2011, 118, 75–83. [Google Scholar] [CrossRef]
  44. Wang, F.; Ngo, C. W. Summarizing rushes videos by motion, object, and event understanding. IEEE Transaction on Multimedia 2012, 14(1), 76–87. [Google Scholar] [CrossRef]
  45. Zou, H.; Mller, H. J.; Shi, Z. Non-spatial sounds regulate eye movements and enhance visual search. Journal of Vision 2012, 12(5), 1–18. [Google Scholar] [CrossRef]
Figure 1. An example of several frames of a clip snippet (one shot) with the associated soundtrack (time-frequency representation). The soundtrack is a succession of two types of sound. In this example, the first sound is birds singing, and the second sound is the boy in the center talking. Frame 1 corresponds to the beginning of the second sound. 
Figure 1. An example of several frames of a clip snippet (one shot) with the associated soundtrack (time-frequency representation). The soundtrack is a succession of two types of sound. In this example, the first sound is birds singing, and the second sound is the boy in the center talking. Frame 1 corresponds to the beginning of the second sound. 
Jemr 06 00012 g001
Figure 3. Time course of two clips with AV condition. To control the gaze of participant, a fixation cross is presented at the center of the screen before each clip. This sequence is repeated for five clips with AV condition and five clips with V condition. 
Figure 3. Time course of two clips with AV condition. To control the gaze of participant, a fixation cross is presented at the center of the screen before each clip. This sequence is repeated for five clips with AV condition and five clips with V condition. 
Jemr 06 00012 g003
Figure 6. Entropy variation between AV and V conditions, for “on-screen with one sound source”, “on-screen with more than one sound source”, and “off-screen sound source” clusters. 
Figure 6. Entropy variation between AV and V conditions, for “on-screen with one sound source”, “on-screen with more than one sound source”, and “off-screen sound source” clusters. 
Jemr 06 00012 g006
Figure 7. Average difference (KLAVV - KLR) over time for “speech” (11 clip snippets) and “impact and explosion” (8 clip snippets) classes. Frame 1 corresponds to the beginning of the second sound. Dark regions represent (KLAVV - KLR) over 0, suggesting that the difference be- tween AV and V groups is greater than that between two random groups. 
Figure 7. Average difference (KLAVV - KLR) over time for “speech” (11 clip snippets) and “impact and explosion” (8 clip snippets) classes. Frame 1 corresponds to the beginning of the second sound. Dark regions represent (KLAVV - KLR) over 0, suggesting that the difference be- tween AV and V groups is greater than that between two random groups. 
Jemr 06 00012 g007
Table 2. Probability estimations of  K L ¯ i  values higher than  K L ¯ A V V  (respectively with metrics of cc and md) from frame 6 to 30 after the beginning of the second sound for all the sound classes. Smaller p value represents a higher possibility that KL between groups with AV and V conditions is larger than KL between random groups.
Table 2. Probability estimations of  K L ¯ i  values higher than  K L ¯ A V V  (respectively with metrics of cc and md) from frame 6 to 30 after the beginning of the second sound for all the sound classes. Smaller p value represents a higher possibility that KL between groups with AV and V conditions is larger than KL between random groups.
Jemr 06 00012 i005
Table 3. Probability estimation of  D ¯ i  being smaller than  D ¯ A V S  from frame 6 to 30 after the beginning of the second sound for “on-screen with one sound source” cluster.
Table 3. Probability estimation of  D ¯ i  being smaller than  D ¯ A V S  from frame 6 to 30 after the beginning of the second sound for “on-screen with one sound source” cluster.
Jemr 06 00012 i006

Share and Cite

MDPI and ACS Style

Song, G.; Pellerin, D.; Granjon, L. Different Types of Sounds Influence Gaze Differently in Videos. J. Eye Mov. Res. 2013, 6, 1-13. https://doi.org/10.16910/jemr.6.4.1

AMA Style

Song G, Pellerin D, Granjon L. Different Types of Sounds Influence Gaze Differently in Videos. Journal of Eye Movement Research. 2013; 6(4):1-13. https://doi.org/10.16910/jemr.6.4.1

Chicago/Turabian Style

Song, Guanghan, Denis Pellerin, and Lionel Granjon. 2013. "Different Types of Sounds Influence Gaze Differently in Videos" Journal of Eye Movement Research 6, no. 4: 1-13. https://doi.org/10.16910/jemr.6.4.1

APA Style

Song, G., Pellerin, D., & Granjon, L. (2013). Different Types of Sounds Influence Gaze Differently in Videos. Journal of Eye Movement Research, 6(4), 1-13. https://doi.org/10.16910/jemr.6.4.1

Article Metrics

Back to TopTop