Next Article in Journal
Incorporation of Prior Knowledge and Habits While Solving Anagrams
Previous Article in Journal
The Mechanism of Word Satiation in Tibetan Reading: Evidence from Eye Movements
 
 
Journal of Eye Movement Research is published by MDPI from Volume 18 Issue 1 (2025). Previous articles were published by another publisher in Open Access under a CC-BY (or CC-BY-NC-ND) licence, and they are hosted by MDPI on mdpi.com as a courtesy and upon agreement with Bern Open Publishing (BOP).
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Do Weak Readers in Rural India Automatically Read Same Language Subtitles on Bollywood Films? An Eye Gaze Analysis

by
Somnath Arjun
1,
Brij Kothari
2,
Nirav Kumar Shah
3 and
Pradipta Biswas
1
1
Indian Institute of Science, Bangalore, India
2
Indian Institute of Management, Ahmedabad, India
3
PlanetRead, Pondicherry, India
J. Eye Mov. Res. 2022, 15(5), 1-15; https://doi.org/10.16910/jemr.15.5.4
Submission received: 8 March 2022 / Published: 24 November 2022

Abstract

:
Same Language Subtitling (SLS) of audio-visual content on mainstream TV entertainment to improve mass reading literacy was first conceived and piloted in India. SLS is now being scaled up nationally to ensure that the reading skills of one billion TV viewers, including 600 million weak readers, remain on a lifelong pathway to practice, progress, and proficiency. Will weak readers ignore or try to read along with SLS? Our eye-tracking study investigates this question with 136 weak readers drawn from a remote village in Rajasthan state by showing them popular Hindi film clips of dialog and songs, with and without SLS. We developed an interactive web-based visual analytics tool for exploring eye-tracking data. Based on an analysis of fixations, saccades, and time spent in the subtitle and non-subtitle areas, our main finding is that 70 percent of weak readers engaged in unprompted reading while watching film clips with SLS. We observed that saccadic eye movement is a good indicator to quantify the amount of reading with SLS, and saccadic regression can further differentiate weak readers. Eye-tracking studies of weak readers watching subtitles are rare, and ours may be the first with subjects from rural India.

Introduction

Ninety-five percent of the world’s illiterate people live in developing countries (Verner, 2005). India’s literacy rate is 78 percent (NSO, 2020) but the quality of literacy is extremely low. Sixty percent “literates” cannot read simple texts, much less a newspaper (Kothari & Bandyopadhyay, 2010). Children fall behind in reading from the early grades and weak skills only erode later in life due to a lack of reading practice. The nation and practically every state confronts weak reading at scale and the overwhelming majority of non-literates and weak readers are rural and female (Chandra, 2019). The long-term goal of the Billion Readers (BIRD) initiative at the Indian Institute of Management, Ahmedabad (IIM-A) is to ensure daily and lifelong reading practice for a billion people in India. In order to achieve this goal, BIRD aims to scale up Same Language Subtitling (SLS) on all existing entertainment content on television and streaming platforms, expressly for reading literacy practice and skill improvement at population scale. This is a global first in some important respects.
Till date, no country’s TV and/or streaming policy has implemented text on screen explicitly for mass reading literacy. Several majority English-speaking countries have implemented captions for media access among the Deaf and Hard or Hearing (DHH) (NCI, 2020). Since Price, (1983) first researched the availability of captions for learning English as a Second Language (ESL), a large body of evidence has supported captions for second language learning (Burger Guenter (2022) maintains a comprehensive online bibliography).
The suggestion of leveraging captions for reading literacy among the DHH (Cooper, 1973) and the hearing (Koskinen, 1985) is as old as the idea itself. However, the use of captions to strengthen reading has been far less researched than their use for language learning and has certainly not driven national broadcast policy on mainstream TV or streaming, anywhere in the world, for mass reading literacy. The ‘SLS’ project was first conceived in India in 1996 with the intent to move national media policy and impact reading literacy at scale (Kothari, 1998). Existing terms like ‘bimodal’ and ‘intralingual’ subtitling, while meaning exactly the same, did not suit the project’s need for a self-evident term that spoke widely to policy makers and viewers alike.
SLS is the idea of subtitling audio-visual (AV) content in the ‘same’ language as the audio. What you hear, is what you read. SLS means Hindi subtitles on Hindi content, Tamil subtitles on Tamil content and likewise on all existing and popularly watched Indian language content like films, serials, cartoons, and songs. In India and globally, most entertainment content in English is available with SLS. That is not the case on content in most of the world’s other, and especially, non-Roman script languages. Streaming platforms in India offer translation subtitles in a number of languages, but not in the ‘same’ audio language.
SLS is a deceptively simple social innovation with the power to transform a large population of struggling readers into functional and even fluent readers. Globally, the enormous value of SLS for reading literacy at population scale remains mostly untapped. After field testing in villages and poor urban communities, researchers at IIM-A piloted SLS on TV in Gujarat state in 1999 and found it to be an effective intervention for improving reading skills (Kothari et al., 2004). Since then, other studies have confirmed that frequent matching text-sound exposure improves reading (e.g., Kothari & Bandyopadhyay, 2014).
The theory of change driving BIRD is that a population that engages in reading every day, all through life, cannot remain weak reading. Currently there is nothing that ensures daily reading practice for all Indians, across the average lifespan of 70 years. If all the entertainment content on TV and streaming platforms carried SLS, millions of viewers would automatically try and associate the matching text and sound. Reading skills would remain in a constant state of reinforcement.
To our knowledge eye-tracking studies of beginning or weak-readers’ viewing of subtitles are rare, if any, anywhere in the world. Most eye-tracking studies on dynamic texts assume functionally reading subjects and hardly any have been conducted in a developing country (Kruger et al., 2015). Negi and Mitra (2020) study in India is with good readers with a minimum 10th Grade education. Our present eye-tracking study in rural Rajasthan, with weak readers – those who can decode some letters or simple words but struggle to read the words in a 2nd Grade level text as single units, i.e., – they lack the ability to unitize (Ehri, 2005) – may be a first.
An eye-tracking study undertaken earlier (Kruger & Steyn, 2014) found that viewers who saw videos with SLS and read them had better comprehension than those who saw the same videos but did not read the SLS. The participants were literate college students in South Africa who spoke English as a second language. The study was conducted in the context of English subtitles on academic lectures delivered in English. It demonstrated the educational potential of SLS in reading instruction and language learning. In rural India, the primary source of entertainment is via native language films and television programs. Could SLS on entertainment content be an approach to improved reading skills?
There is some evidence in the literature that SLS significantly impacts reading skills (Bird & Williams, 2002; Danan, 2004; Linebarger et al., 2010). A controlled experiment with schoolchildren in India demonstrated that SLS leads to reading improvement (Kothari & Takeda, 2000). A national TV pilot of reading improvement further supported that there is a significant effect of SLS on reading skills (Kothari & Bandyopadhyay, 2014).
Most of the analyses for previous studies focused on the change in reading performance but did not eye-track the reading engagement with SLS. This paper analyzes the processing of SLS on popular Hindi songs and dialogs with eye gaze metrics. In addition to fixation and scan path measures, we investigated regressions, saccade amplitudes, and revisits. These metrics helped us to examine the reading behavior of participants and the proportion of engagement with SLS. We developed a visual analytic tool to investigate eye-tracking data. The tool provided information about eye gaze patterns that would have gone unnoticed otherwise. For example, the tool could differentiate between the reading pattern of weak readers who can read with difficulty (functional weak readers) and very weak readers who cannot read functionally (poor readers) with respect to subtitle reading in videos with SLS.
We discuss the related work on SLS in Section 2 followed by an overall design of our user study in Section 3. The methodology is discussed in Section 4. Section 5 presents our analysis and results followed by concluding remarks in Section 6.

Related Work

The most sustained push for SLS on TV for reading literacy has been in India, through research, pilot TV implementations in 10 Indian languages and evidence-based policy advocacy (Kothari & Bandyopadhyay, 2020). However, the positive impact of SLS use on reading literacy has also been affirmed in research studies from other countries (Koskinen et al., 1997; D. Linebarger et al., 2001, 2010; Parkhill & Davey, 2014). Taken together, these studies make a strong case that routine and regular exposure to SLS can have an impact at scale on a population’s reading development.
Previous studies have found that the presence of SLS on popular entertainment content causes automatic reading skill practice and improvement among viewers who can minimally recognize letters (Kothari et al., 2004; Kothari & Bandyopadhyay, 2014). Moreover, there is compelling evidence that SLS also promotes language learning and media access among the Deaf and Hard of Hearing (DHH) (Danan, 2004; Gernsbacher, 2015; Pavakanun, 1992; Vanderplank, 2016). While any of these benefits provides a strong rationale for SLS, the evidence for all three makes it a compelling proposition on TV and streaming platforms.
SLS on popular entertainment content is a critical aspect of BIRD for two factors that powerfully drive reading skill acquisition: first, an inordinate amount of practice and second, with text embedded in a context that the reader will be passionate about for life. Toste et al. (2020) review the powerful bi-directional relations between motivation and reading skill. Both drive each other. For proficient reading, grapheme-phoneme associations need to fire frequently and sufficiently over a long enough period of time to achieve and sustain automaticity. As Frey and Fisher (2010) state “When we experience something, neurons fire. Repeated firings lead to physical changes in the brain that, over time and with repetition, become more permanent.” They further point out that “The challenge, of course, with automaticity is to not allow repetition to turn into a rut.”
When reading acquisition occurs in a person’s first language, the auditory-to-language pathways are well-established in the brain. In a weak reader, the letter-to-soundto-meaning pathways are weak (van Atteveldt et al., 2009). Reading fluency and comprehension are achieved over a long period of sustained exposure to congruent lettersound correspondence and decoding practice. As decoding approaches fluency and automaticity, cognitive resources are freed to focus on the key task of meaning-making. SLS on popular songs, nursery rhymes and repeatedly watched cartoons has the added advantage of reading practice with predictable sound-to-text reinforcement, on content of high-interest (Hill-Clarke & Robinson, 2004; Iwasaki et al., 2013). The visual and auditory pathways involved in reading are strengthened incidentally, subconsciously, and inevitably, as a by-product of entertainment.
There are few longitudinal studies on the impact of captions on reading skills (Koskinen, 1985). Linebarger et al. (2010) study indicated that children who viewed video with captions improved their reading faster than their counterparts who viewed without captions, and the improvement was most pronounced among children at risk for poor reading outcomes. Similarly, in New Zealand, Parkhill and Johnson (2009) found that in their six-week ‘AVAILLL’ program for children aged 5-13 years, which uses popular, subtitled movies and accompanying novels to engage students in reading literacy, the greatest gains occurred for ‘low-progress’ readers. A positive impact was also observed for average and higher-level readers. Kothari and Bandyopadhyay (2014) evaluated the impact of SLS after sustaining it for 5 years on a weekly hour-long program of Hindi film songs telecast nationally in prime time. Among school children who could not read a single letter in Hindi at the baseline (2002), 70% in the high-SLS viewing group became functional readers by the end line (2007) as compared to 34% in the low-SLS group. The benefits of SLS or Closed-Captioning are not limited to reading literacy. The range of benefits attributable to SLS include reading, media access and language acquisition (Gernsbacher, 2015).
Although there is a substantial body of research on measuring the effect of SLS on reading using standardized tests, eye-tracking studies that focus on the development of reading skills are uncommon (Schroeder et al., 2015). Studies have examined gaze patterns while reading text or phrases (Beck & Konieczny, 2020; Geyer et al., 2020; Magyari et al., 2020). Smyrnakis et al. (2021) compared the silent and loud reading ability of typical and dyslexic readers using eye-tracking while Faber et al. (2020) investigated the differences in stable eye movement patterns during narrative reading. Eye-tracking studies were also used to understand cognitive load while reading (Babu et al., 2019; Kucharský et al., 2020).
A key finding of eye-tracking research on SLS and subtitling in general is that viewers will engage with the onscreen text automatically (d’Ydewalle & de Bruycker, 2007; Schroyens et al., 1999a, 1999b). Viewers just cannot ignore the subtitles in movies, although they may do so periodically. It is important to note, however, that all these studies were conducted with subjects who could read. Reading along with SLS is preferred because of the efficiency in following and understanding the movie. Evidence of such eye-tracking studies enables us to attribute possible learning outcomes to the subtitles. d’Ydewalle and de Bruycker (2007) reported that subtitle reading is inescapable and viewers have little difficulty in distributing visual attention.
Several studies have explored the reading behavior of participants and investigated to what extent they read subtitles (Kruger & Steyn, 2014; Pavakanun, 1992). These studies primarily limit the eye-tracking metrics to fixation count, fixation time, scan path, and average fixation duration in two areas of interest (AOI): the subtitle band and image. Negi and Mitra (2020) eye-tracking study of subtitled videos revealed that fixation duration can be a useful metric to trace the learning process. Other eye-tracking studies investigate the impact of audio on the reading of subtitles (Liao et al., 2022) and make recommendations for future cognitive research in the field of audiovisual translation (Kruger et al., 2015).
Reading along with SLS is inescapable, among good readers (d’Ydewalle et al., 1991). This pioneering study found that American subjects watching an English movie with SLS and Dutch subjects watching a Dutch movie with SLS, spent considerable time in the subtitle area. Reading SLS was inevitable and comparable for both groups, even though the Dutch subjects had much more experience with subtitles on TV. Reading SLS did not depend much on habit formation. The critical question for us is, would struggling readers, especially those from economically disadvantaged backgrounds, also try and engage automatically with SLS? PlanetRead (2018) completed an eyetracking study of government school children in Grades 25 in rural Rajasthan, India, by showing them animated stories with and without SLS. Almost all viewers beginning, struggling, or good readers automatically engaged with SLS and could not ignore it.

User Study

We undertook user studies with participants from remote villages in Rajasthan. We approached an NGO named Dusra Dashak that has been working in Abu Road, Rajasthan for over 15 years and has an established on-theground connect. The head of the NGO advised us on the villages where we could conduct our eye-tracking study. In the selected villages we first got the required permissions with Dusra Dashak’s help and implicit support from the village head to conduct our study. We went door-todoor to mainly identify individuals who had completed their schooling but were still weak readers. As a quick filter, we asked villagers to read a simple Grade 2 level paragraph. Those who struggled to read the text were considered to be weak readers, including poor readers, and selected as study participants. Individually, every participant was asked to see video clips of Hindi songs and dialogs, with and without SLS, on a computer monitor. We collected ocular data while participants watched both versions of the video. In particular, we conducted both statistical and visual data analysis, described in the next section, to address the following four questions.
  • What proportion of viewers’ attention is divided between two regions (video and SLS) while watching a video with SLS?
  • How can we quantify the amount of engagement with the subtitles?
  • How can we differentiate between functional weak readers and poor readers?
  • Is there any difference in engagement when the SLS are highlighted?
Table 1. Media descriptions of audio-visual clips.
Table 1. Media descriptions of audio-visual clips.
ContentMovie NameSong NameDuration
DialogueBaahubaliNot applicable64 secs
SongBesharamDil ka jo haal hai67 secs
SongAbhimaanTeri bindiya re63 secs
SongDhadkanTum dil ki dhadkan42 secs
DialogueSholayNot applicable100 secs
SongKarzKamaal hai94 secs

Procedure

The study procedure had the following steps:
(a)
We gathered background information on the participant such as age, education, gender, and occupation.
(b)
An Early Grade Reading Assessment (EGRA) tool was used to assess the participant’s reading level more accurately than the earlier filter test (USAID, 2019).
(c)
The participant was asked to sit in front of the monitor with screen-based eye-tracker attached to the monitor. They were then asked to undertake the task of watching video clips with and without SLS.
(d)
Each participant was asked to watch twelve short videos ranging from 42 to 100 seconds. The eye-tracking study was conducted with six unique videos with two versions of each video, one with SLS and the other without SLS. Videos are presented in a randomized order to reduce the order effects. The description of media considered for the study is given in Table 1.
(e)
We investigated the collected data using statistical and visual analysis (see results section).
(f)
We developed a web-based hardware agnostic software to analyze eye-tracking data visually. The tool is platform independent and does not require any datapreprocessing, making it easier for users.

Visualization Tool

The tool investigates important eye-tracking metrics: fixation, saccade, scan path and regressions. The tool explores temporal data without losing spatial information and retains the direction of saccades while analyzing the saccadic length. The eye gaze movements across AOIs are usually studied statistically, and occasionally AOI transitions are examined visually (Blascheck et al., 2013). We introduced an interactive method to study and analyze direct and indirect transitions across AOIs. Direct transitions are eye movements from source AOI to target AOI without other AOIs between them. However, if there are one or more AOIs in the eye movement path from source to target AOIs, it is called indirect transition. Unlike existing webbased visualization tools where AOIs are specified as rectangles (Bakardzhiev et al., 2021), the proposed tool allows us to define AOI manually in any convex-shaped polygon with four vertices. The software is developed entirely on JavaScript, and the D3.js library was used for rendering visualizations. It is hosted on a python server and tested with Edge, Chrome, and Mozilla Firefox browsers on a Windows operating system. The tool’s features are described below.
(1) AOI visualization: This technique keeps track of the saccadic eye movements between AOIs. Users can manually choose AOIs by defining points of a four-sided polygon from the control panel on the right side of the canvas. After specifying AOIs, we can visualize the number of saccadic movements between AOIs in a bar chart. The tool’s node-link graph shows the frequency of the saccadic movement. Nodes are used for depicting AOIs, and the saccadic movement between them is represented by an edge connecting the nodes. The higher the number of movements among two AOIs, the thicker is the edge connecting the nodes. We can also compare fixations and total durations of AOIs at a particular time interval with a horizontal stacked bar chart used in the tool.
(2) Saccade visualization: For investigating saccade amplitudes and direction visually, we developed a two canvas technique in the tool. Visualizing saccades can be a difficult task because of the varying saccadic length. The longest or shortest saccades might be of interest, and the direction of saccades gives an overview of the eye movement in that specific area of the stimulus. It becomes challenging to track both the saccadic length and direction as points get cluttered with gaze plots. We developed a unicolor-coded technique to address this issue. A set of rectangles representing saccades is plotted on a separate graph. Rectangles have identical colors, but their opacity changes according to the length of the corresponding saccade. The more the opacity of the rectangle, the longer is its length. Selecting the rectangle highlights the original saccade in the gaze plot by changing the color and width of the line. The selected saccade’s start and target position colors are also changed to red and green, respectively. This technique helps in identifying the direction of the saccade. Both gaze plots and a set of rectangles for saccade visualization are shown in Figure 1 and Figure 2 respectively. It can be noticed that the saccadic movement across the AOIs is one of the biggest saccade amplitude.
(3) Fixation and Regression visualization: The tool allows interactive fixation and regression visualization. Regression is a type of eye movement that orients backward with respect to normal reading behavior. The software separates regressions from normal saccades by highlighting them with different colors. Fixations are shown as scatter plots where each circle represents one gaze fixation. The size of the circle depends on the duration of the fixation, bigger circles depict higher durations. An interactive brush is used to select a range of fixations, the size of the brush can be adjusted vertically. The range of the brush corresponds to the timeline of the task. For example, if a user want to explore fixations of a particular time segment of the task, he or she can adjust the brush so that it corresponds to that time segment. The technique would help users to reduce the clutter and concentrate only on specific fixations.

Participants

We selected overall 136 participants who are weak readers and our goal was to track their eye movements while they are watching Hindi film-based songs and dialog clips, with and without SLS. The sex and reading ability profile is given in Table 2.
All the participants were above 14 years of age, except for one. The average age was 24 years. Parents accompanied children below 18 years to the eye-tracking location set up in the community and gave oral permission since most of them were non-literate themselves. The average grade completed was 7.9 with a standard deviation of 3.3. Over 32% in 7th and 27% in 8th grade are not able to read a 2nd grade text in rural India (ASER, 2018).

Materials

A Gazepoint 3 screen mounted eye tracker with an accuracy of 0.5º-1º of visual angle, sampling rate of 60 Hz, and an average calibration error of 0.5-1 degrees of visual angle was used to collect gaze-based data. A 19-inch computer monitor was used to display the videos for participants. An EGRA (USAID, 2019) was used to assess the participants’ reading ability. The test is in increasing order of difficulty starting with characters and finishing with a simple story. The subsection below explains the test in detail.
The reading test, pitched at 2nd Grade level, comprises five simple exercises with increasing difficulty. The participants were asked to read:
  • Hindi syllabary in random order comprising 52 syllables.
  • 50 meaningful words in Hindi (Figure 3).
  • 50 meaningless words in Hindi.
  • Four simple Hindi sentences in Hindi with 19 words in total.
  • A Hindi story of 65 words.

Data Collection and Analysis

All 136 participants were asked to sit in front of a computer monitor at 100 cm from the monitor and 60 cm from a screen-based eye-tracker. They were then asked to calibrate with the eye-tracker with a 9-point calibration routine built into the eye tracking software. It required participants to follow dots on the screen. This ensured accurate eyetracking across the entire screen for all participants. Participants were then asked to watch the videos and the order of the videos was randomized to reduce any learning impacts. The eye-tracking software collected data while participants were undertaking the task. Eye-tracking data were primarily analyzed with respect to AOIs corresponding to the videos shown to users. We considered two sets of AOIs for the analysis. In the first set, the screen is divided into two AOIs: the first AOI is the region with video (2/3rd of the screen) and the second AOI is the region with SLS, as shown in Figure 4. From the visualization tool, we found that fixations of poor readers tend to be more in the left region than the right region of the subtitle area. Based on this observation, we have split the SLS region into two equal parts (left and right regions) to delve deeper into the pattern, as shown in Figure 5.

Results

This section lists the indicators and parameters that can answer the questions addressed in section 1 on the resulting reading behavior from SLS. In the subsections below, we first list the parameters and then discuss the results of our statistical analysis to justify why the selected indicators can address our research questions. We also report a visual analysis of eye tracking data to support the results. For the analysis, we calculated the average values of the indicators or parameters for all 136 participants. We prepared a table of 12 columns corresponding to two versions (SLS and No SLS) of six unique videos and 136 columns corresponding to 136 participants.
We further analyzed the data statistically with two independent variables and each variable has two levels: (a) Video: with and without SLS, and (b) AOIs: AOI_SLS and AOI_Video. In particular we undertook Video(2) × AOI(2) repeated measure ANOVA for identifying statistically significant differences between the experimental conditions. We report the statistical results in the sections below along with other results.

What Proportion of Viewers’ Attention Is Divided Between Two Regions (Video and SLS) While Watching a Video with SLS?

Indicators considered for this question are explained below:
(1)
Change in the number of fixations and saccades in AOI_Video
Participants who try to read subtitles while watching a video with SLS are expected to devote more time to AOI_SLS. Participants’ fixations and saccades at AOI_Video would be less if they try to scan subtitles on AOI_SLS while watching a video with SLS. A consideration of both fixations and saccades ensures that participants were indeed trying to concentrate at the region rather than just moving their eyes. Figure 6 and Figure 7 show how fixation counts can change in both the AOIs for the same video, without SLS and with SLS. The change of participants’ fixations and saccades at AOI_Video was calculated by subtracting the fixation numbers in the video without SLS from those in the video with SLS. We calculated the change for each pair of videos (with and without SLS) across all the participants and the videos. Table 3 illustrates the average percentage of change in eye movements across all the participants for six pairs. In four out of the six pairs of videos the percentage change in the eye movements is more than 25%. We also undertook paired-samples t-test for each pair of videos to statistically investigate the change of eye movements in AOI_Video. The results listed in Table 4 indicates that the eye movements in AOI_Video are significantly different between each pair of all six videos.
(2)
Total time spent in AOI_Video
The time spent by participants in AOI_Video while watching a video without SLS is expected to be more than a video with SLS. Based on this assumption, the total time spent could be helpful in identifying readers’ attention in both the AOIs. We computed the differences in time spent in AOI_Video between the pair of videos (with and without subtitles) as this is the constant AOI in both the situations. The difference would express the divided attention of participant due to the inclusion of subtitles. Table 5 shows the average time differences in AOI_Video across all participants for each pair of video. After undertaking repeated measure ANOVA, we found significant difference for the interaction effects of video × AOI: F(1, 135) = 287.42, η2 = 0.68, p < 0.05. We also found that video AOI are significantly different for the time spent. F(1, 135) = 754.77, η2 = 0.84, p < 0.05. We also undertook a pairwise comparison test like least significant difference (LSD) and found that AOI_Video in the video without subtitle is significantly different from the video with subtitle (p < 0.05).
(3)
Saccades across AOIs or revisits at AOIs
Participants whose attention are divided between the two AOIs while watching videos with SLS would repeatedly visit the subtitle area. While we cannot expect much eye movements between the two AOIs in the videos without SLS as there were no subtitles. However, it is intuitive that the introduction of subtitles would increase the eye movements across the AOIs in videos with SLS. The comparisons of saccades across AOIs would allow us to investigate the quantity with which the participants were trying to visit the subtitle region while watch videos with SLS. We compared the average number of saccades across AOIs for all the participants on all SLS videos and its counterpart. The result conveys that the rate of increase in saccadic eye movement is approximately 60% for videos with SLS as compared to videos without SLS. We undertook 2 × 2 repeated measure ANOVA and found significant difference for the interaction effect of video × AOI: F(1, 135) = 286.53, η2 = 0.68, p < 0.05. We further found that video and subtitle AOIs are significantly different for saccades across AOIs F(1, 135) = 329.69, η2 = 0.7, p < 0.05. After undertaking the LSD test, we found that the video with SLS is significantly different from the video without SLS for AOI_Video (p < 0.05).

In What Ways Can We Quantify the Amount of Engagement with SLS?

The indicators used are explained below:
(1)
Fixation rate and number of fixations
Fixation rate and the number of fixations in AOI_Video could be good indicators for assessing the amount of engagement. We calculated the average number of fixations and average fixation rate at AOI_Video across all participants for all videos with and without SLS. The fixation rate is calculated as:
Jemr 15 00033 i001
Figure 8 compares the average fixation rate across all participants. We found from repeated measure ANOVA that the interaction effects of video × AOI are significantly different: F(1, 135) = 153.14, η2 = 0.535, p < 0.05. We found that videos with and without SLS are significantly different F(1, 135) = 156.02, η2 = 0.54, p < 0.05.
(2)
Total durations in AOI_Video
Viewers would be expected to spend more time in AOI_SLS when they try to read the subtitles. Average and total fixation duration could thus be an important variable for quantifying the amount of reader engagement. Higher duration might not always indicate that participants were reading the subtitles. However, it would at least indicate that they were trying to engage in that region while watching the video. We undertook repeated measure ANOVA for total durations in AOI_Video across all participants for all videos and found that there is a significant difference for the interaction effects of video × AOI: F(1, 135) = 287.42, η2 = 0.68, p < 0.05. We also found that video AOI are significantly different for total durations. F(1, 135) = 754.77, η2 = 0.84, p < 0.05.
(3)
Saccade amplitude at AOI_Video
Eye movements in the video region will be less if participants try to scan through the subtitle while watching the video. It will lead to smaller number of saccades and hence shorter total saccade lengths are expected in the video area and higher total saccade lengths are expected in the SLS area. We investigated the saccade amplitudes for both versions of video and found that saccade amplitude decreases significantly in AOI_Video. Table 6 compares the change of saccade length in AOI_Video for each pair of six videos. We found that the interaction effect of video × AOI are significantly different: F(1, 135) = 221.99, η2 = 0.62, p < 0.05. The videos with and without SLS are significantly different for saccade length F(1, 135) = 16.82, η2 = 0.11, p < 0.05. After undertaking LSD test, we found that both versions of video are significantly different with respect to AOI Video (p < 0.05).

How Can We Differentiate Between Functional Weak Readers and Poor Readers?

(1)
Number of regressions at AOI_SLS
Regression is a type of saccadic eye movement that orients backward for normal reading behavior on the screen while reading (Booth & Weger, 2013). It is also termed backward eye movement. The standard reading behavior of participants in the Hindi language is from left to right on the screen, and we expect their eye movement in a similar direction. Regression could be an indicator for differentiating between functional weak readers (FWR) and poor readers (PR). We expect poor readers to have higher number of regressions than the functional weak readers, as poor readers would read with greater difficulty than their counterparts. We compared the average number of regressions for FWR and PR only on videos with SLS and observed that number of regressions for both the groups are almost similar. The results of the comparison are listed in Table 7. As regressions in AOI_Video would not help us in identifying the reading behaviour of participants, we analysed regressions only at AOI_SLS. We also undertook paired sample t-test and did not find any significant difference between the two groups.
(2)
Fixations at left AOI_SLS and right AOI_SLS
We noticed from the visualization tool, as shown in Figure 9 and Figure 10, that poor readers have relatively more fixations at the left part of the subtitle area than functionally weak readers. We found that the average fixation rate at AOI_SLS_Left is relatively more for poor readers than functional weak readers in all videos. However, the difference is not significantly high as can be seen from Table 8. We also undertook paired samples t-test and did not find any significant difference between the two groups.

Is There Any Difference in Engagement When the SLS Are Highlighted?

Among the six videos with SLS, the subtitle words of two videos are not highlighted and the words of four videos are highlighted. We considered the number of fixations and durations for analysis.
(1)
Number of fixations and durations
As stated earlier, the number of fixations and durations could be a good indicator for identifying reading engagement. We used these same variables for comparing our two design versions of SLS. We did not find any considerable difference among the subtitle designs for both the parameters.

Discussion

Our study is unique in several respects. First, a number of eye-tracking studies have established automatic reading behavior while watching video content, but, in almost all these studies, the subjects were good readers. Our subjects were all weak and many of them poor readers and we established that a majority of them also try to read along with SLS on video.
Second, eye-tracking studies in India generally, and of video consumption in particular, are rare. With a billion TV viewers and 356 million mobile video viewers in 2021 (INMOBI, 2021), and growing fast, eye gaze research on how viewers consume video content, with or without subtitles, or interact with screens, is a fertile area for study.
Third, to our knowledge, there is no eye gaze study conducted with subjects from rural India. Perhaps the reason is that it may seem difficult to transport and operate eye-tracking equipment in rural areas or bring rural subjects to urban centers. Our study has demonstrated that the former is indeed possible.
Impact measurement studies have found that regular SLS exposure among weak readers leads to reading skill improvement, over time (Kothari et al., 2004; Kothari & Bandyopadhyay, 2014). Future studies could explore whether eye gaze measurement can also capture reading skill improvement resulting from SLS exposure over time. How would a subject’s reading of SLS on the same video be different at the baseline and end line, assuming that the subject’s reading skills have improved? The visualization tool proposed in the paper was useful to support the results of the statistical analysis.
This paper presents a detailed eye-tracking study on processing SLS while watching Hindi film songs and dialogs. We analyzed five gaze-based metrics to investigate eye movements of users in the subtitle region of video clips. We also observed the proportion of viewers who try to read along while watching videos. We noticed that regression between poor and functional weak readers are almost similar. We also found significant differences for the interaction effect between the experimental conditions. Weak readers tend to linger more in the left half of the subtitle area. From a policy perspective, a key finding is that both poor readers and functional weak readers tend to engage their eye movements in the subtitle regions with the introduction of SLS in the video clips. We plan to undertake a similar eye-tracking study of SLS film videos to understand changes in eye-gaze patterns when an individual becomes a better reader. This could even be conducted with a subset of our study participants who have measurably improved their reading skills since this study was conducted. Our results could potentially be used to design an adaptive user interface for learning while watching videos with SLS.

Conclusions

In 2019, SLS became a part of India’s Accessibility Standards (India’s Accessibility Standards, 2019), mandating that half the entertainment content on TV, in every language, state, and channel, is required to carry SLS by 2025. Our finding that weak readers do try to read along with SLS on films, supports the leveraging of the Accessibility Standards, for both, media access among the DHH and improving the reading skills of over half a billion weak readers in India, an overwhelming majority being female. The Accessibility Standards, if implemented with the intent to benefit all, will then be designed to benefit all. But if they are framed as good only for the hearing impaired, then the implementation risks not leveraging their full potential for the hearing. This perspective is important to keep in mind for policy makers in countries, like India, that are in the process of adopting SLS
Countries that do have captioning on TV might consider re-designing captions with universal visual appeal. By not doing so, they lost out on a massive opportunity to contribute, at scale, to their populations’ reading literacy and language learning. The hearing also want to turn TV captions on for different reasons, as they now do on streaming platforms. Networks might consider making captions available on content in all languages and turning them on by default, especially on children’s programming, like Turn On The Subtitles (TOTS) is campaigning for in the UK. Finally, parents the world over need to know that SLS or captions can make a massive contribution to their child’s reading literacy and language skills. All they have to do is turn them on whenever possible.

Conflicts of Interest

The author(s) declare(s) that the article’s contents are in agreement with the ethics described in http://biblio.unibe.ch/portale/elibrary/BOP/jemr/ethics.html and that there is no conflict of interest regarding the publication of this paper.

References

  1. ASER. 2018. Annual Status of Education Report (Rural). http://img.asercentre.org/docs/ASER%202018/Release%20Material/aserreport2018.pdf.
  2. Babu, M. D., D. v. JeevithaShree, G. Prabhakar, K. P. S. Saluja, A. Pashilkar, and P. Biswas. 2019. Estimating pilots’ cognitive load from ocular parameters through simulation and in-flight studies. Journal of Eye Movement Research 12, 3. [Google Scholar]
  3. Bakardzhiev, H., M. van der Burgt, E. Martins, B. van den Dool, C. Jansen, D. van Scheppingen, G. Wallner, and M. Burch. 2021. A Web-Based Eye Tracking Data Visualization Tool. International Conference on Pattern Recognition, 405–419. [Google Scholar]
  4. Beck, J., and L. Konieczny. 2020. Rhythmic subvocalization: An eye-tracking study on silent poetry reading. Journal of Eye Movement Research 13, 3. [Google Scholar] [CrossRef] [PubMed]
  5. Bird, S. A., and J. N. Williams. 2002. The effect of bimodal input on implicit and explicit memory: An investigation into the benefits of within-language subtitling. Applied Psycholinguistics 23, 4: 509–533. [Google Scholar]
  6. Blascheck, T., M. Raschke, and T. Ertl. 2013. Circular heat map transition diagram. Proceedings of the 2013 Conference on Eye Tracking South Africa, 58–61. [Google Scholar]
  7. Booth, R. W., and U. W. Weger. 2013. The function of regressions in reading: Backward eye movements allow rereading. Memory & Cognition 41, 1: 82–97. [Google Scholar]
  8. Burger Guenter. 2022. Research Bibliography. http://www.fremdsprache-und-spielfilm.de/Captions.htm.
  9. Caldwell Doris Cooper. 1973. Use of Graded Captions with Instructional Television for Deaf Learners. American Annals of the Deaf 118: 500–507. [Google Scholar] [PubMed]
  10. Chandra, T. 2019. Literacy in India: The gender and age dimension. Observ Res Foundation 322. [Google Scholar]
  11. Danan, M. 2004. Captioning and subtitling: Undervalued language learning strategies. Meta: Journal Des Traducteurs/Meta: Translators’ Journal 49, 1: 67–77. [Google Scholar] [CrossRef]
  12. d’Ydewalle, G., and W. de Bruycker. 2007. Eye movements of children and adults while reading television subtitles. European Psychologist 12, 3: 196–205. [Google Scholar]
  13. d’Ydewalle, G., C. Praet, K. Verfaillie, and J. v. Rensbergen. 1991. Watching subtitled television: Automatic reading behavior. Communication Research 18, 5: 650–666. [Google Scholar] [CrossRef]
  14. Ehri, L. C. 2005. Learning to read words: Theory, findings, and issues. Scientific Studies of Reading 9, 2: 167–188. [Google Scholar] [CrossRef]
  15. Faber, M., M. Mak, and R. M. Willems. 2020. Word skipping as an indicator of individual reading style during literary reading. Journal of Eye Movement Research 13, 3. [Google Scholar] [CrossRef] [PubMed]
  16. Frey, N., and D. Fisher. 2010. Reading and the brain: What early childhood educators need to know. Early Childhood Education Journal 38, 2: 103–110. [Google Scholar] [CrossRef]
  17. Gernsbacher, M. A. 2015. Video captions benefit everyone. Policy Insights from the Behavioral and Brain Sciences 2, 1: 195–202. [Google Scholar] [CrossRef] [PubMed]
  18. Geyer, T., F. Günther, H. J. Müller, J. Kacian, H. R. Liesefeld, and S. Pierides. 2020. Reading Englishlanguage haiku: An eye-movement study of the ‘cut effect’. Journal of Eye Movement Research 13, 2. [Google Scholar] [CrossRef] [PubMed]
  19. Hill-Clarke, K. Y., and N. R. Robinson. 2004. It’s as Easy as A-B-C and Do-Re-Mi: Music, Rhythm, and Rhyme Enhance Children’s Literacy Skills. YC Young Children 59, 5: 91–95. [Google Scholar]
  20. India’s Accessibility Standards. 2019. Accessibility Standars for Persons with Disabilities in TV Programmes; Ministry of Information & Broadcasting. https://mib.gov.in/sites/default/files/Accessibility%20Standards%20for%20Persons%20with%20Disabilities%20in%20TV%20Programmes%20.pdf.
  21. INMOBI. 2021. Programmatic Mobile Video Ads See 194% Growth in India. INMOBI. https://www.inmobi.com/company/press/programmatic-mobile-video-ads-see-194-percent-growth-in-india.
  22. Iwasaki, B., T. Rasinski, K. Yildirim, and B. S. Zimmerman. 2013. Let’s bring back the magic of song for teaching reading. The Reading Teacher 67, 2: 137–141. [Google Scholar] [CrossRef]
  23. Koskinen, P. S. 1985. Closed-Captioned Television: A New Tool for Reading Instruction. Literacy Research and Instruction 24, 4: 1–7. [Google Scholar]
  24. Koskinen, P. S., C. T. Bowen, L. B. Gambrell, C. J. Jensema, and K. W. Kane. 1997. Captioned television and literacy development: Effects of home viewing on learning disabled students. Meeting of the American Educational Research Association, Chicago, IL. [Google Scholar]
  25. Kothari, B. 1998. Film songs as continuing education: Same language subtitling for literacy. Economic and Political Weekly, 2507–2510. [Google Scholar]
  26. Kothari, B., and T. Bandyopadhyay. 2010. Can India’s “literate” read? International Review of Education 56, 5: 705–728. [Google Scholar]
  27. Kothari, B., and T. Bandyopadhyay. 2014. Same language subtitling of Bollywood film songs on TV: Effects on literacy. Information Technologies & International Development 10, 4: pp-31. [Google Scholar]
  28. Kothari, B., and T. Bandyopadhyay. 2020. Lifelong Reading for a Billion People. Stanford Social Innovation Review 18, 3: 34–41. [Google Scholar]
  29. Kothari, B., A. Pandey, and A. R. Chudgar. 2004. Reading out of the “idiot box”: Same-language subtitling on television in India. Information Technologies & International Development 2, 1: pp-23. [Google Scholar]
  30. Kothari, B., and J. Takeda. 2000. Same language subtitling for literacy: Small change for colossal gains. Information and Communication Technology in Development, 176–186. [Google Scholar]
  31. Kruger, J.-L., and F. Steyn. 2014. Subtitles and eye tracking: Reading and performance. Reading Research Quarterly 49, 1: 105–120. [Google Scholar]
  32. Kruger, J.-L., A. Szarkowska, and I. Krejtz. 2015. Subtitles on the moving image: An overview of eye tracking studies. Refractory: A Journal of Entertainment Media 25: 1–14. [Google Scholar]
  33. Kucharský, Š., I. Visser, G.-O. Truțescu, P. G. Laurence, M. Zaharieva, and M. E. J. Raijmakers. 2020. Cognitive strategies revealed by clustering eye movement transitions. Journal of Eye Movement Research 13, 1. [Google Scholar]
  34. Liao, S., L. Yu, J.-L. Kruger, and E. D. Reichle. 2022. The impact of audio on the reading of intralingual versus interlingual subtitles: Evidence from eye movements. Applied Psycholinguistics 43, 1: 237–269. [Google Scholar] [CrossRef]
  35. Linebarger, D. L. 2001. Learning to read from television: The effects of using captions and narration. Journal of Educational Psychology 93, 2: 288. [Google Scholar]
  36. Linebarger, D., J. T. Piotrowski, and C. R. Greenwood. 2010. On-screen print: The role of captions as a supplemental literacy tool. Journal of Research in Reading 33, 2: 148–167. [Google Scholar]
  37. Magyari, L., A. Mangen, A. Kuzmičová, A. M. Jacobs, and J. Lüdtke. 2020. Eye movements and mental imagery during reading of literary texts with different narrative styles. Journal of Eye Movement Research 13, 3. [Google Scholar]
  38. NCI. 2020. History of Closed Captioning. https://www.ncicap.org/history-of-cc.
  39. Negi, S., and R. Mitra. 2020. Fixation duration and the learning process: An eye tracking study with subtitled videos. Journal of Eye Movement Research 13, 6. [Google Scholar]
  40. NSO. 2020. Household Social Consumption on Education in India. http://14.139.60.153/han-dle/123456789/13670.
  41. Parkhill, F., and R. Davey. 2014. “I used to read one page in two minutes and now I am reading ten”: Using popular film subtitles to enhance literacy outcomes. Literacy Learning: The Middle Years 22, 2: 28–33. [Google Scholar]
  42. Parkhill, F., and J. Johnson. 2009. An unexpected breakthrough for rapid reading improvement: AVAILLL uses movies so students read it, see it end get it. Set: Research Information for Teachers (Wellington) 1: 28–35. [Google Scholar]
  43. Pavakanun, U. 1992. Incidental acquisition of foreign language through subtitled television programs as a function of similarity with native language and as a function of presentation mode. Unpublished Doctoral Thesis, Leuven, Belgium, University of Leuven. [Google Scholar]
  44. PlanetRead. 2018. Eye Tracking Study of AniBooks Draft Report; PlanetRead. https://www.planetread.org/pdf/Eye%20Tracking%20Study%20of%20AniBooks%20Draft%20Report%20(June%202018).pdf.
  45. Price, K. 1983. Closed-Captioned TV: An Untapped Resource. Newsletter. https://www.matsol.org/assets/documents/Currentsv12no2Fall1983.pdf.
  46. Schroeder, S., J. Hyönä, and S. P. Liversedge. 2015. Developmental eye-tracking research in reading: Introduction to the special issue. Journal of Cognitive Psychology 27, 5: 500–510. [Google Scholar]
  47. Schroyens, W., W. Schaeken, W. Fias, and G. d’Ydewalle. 1999a. Is there any need for eye-movement recordings during reasoning? In Current Oculomotor Research. Springer: pp. 279–285. [Google Scholar]
  48. Schroyens, W., F. Vitu, M. Brysbaert, and G. d’Ydewalle. 1999b. Eye movement control during reading: Foveal load and parafoveal processing. The Quarterly Journal of Experimental Psychology: Section A 52, 4: 1021–1046. [Google Scholar]
  49. Smyrnakis, I., V. Andreadakis, A. Rina, N. Boufachrentin, and I. M. Aslanides. 2021. Silent versus Reading Out Loud modes: An eye-tracking study. Journal of Eye Movement Research 14, 2. [Google Scholar] [CrossRef]
  50. Toste, J. R., L. Didion, P. Peng, M. J. Filderman, and A. M. McClelland. 2020. A meta-analytic review of the relations between motivation and reading achievement for K–12 students. Review of Educational Research 90, 3: 420–456. [Google Scholar]
  51. USAID. 2019. Analysis of Early Grade Reading Assessment in India, Imapct Assessment report. https://earlygradereadingbarometer.org/pdf/India_2018_Impact_Assessement.pdf.
  52. van Atteveldt, N., A. Roebroeck, and R. Goebel. 2009. Interaction of speech and script in human auditory cortex: Insights from neuro-imaging and effective connectivity. Hearing Research 258, 1–2: 152–164. [Google Scholar] [CrossRef] [PubMed]
  53. Vanderplank, R. 2016. Captioned media in foreign language learning and teaching: Subtitles for the deaf and hard-of-hearing as tools for language learning. Springer. [Google Scholar]
  54. Verner, D. 2005. What Factors Influence World Literacy?: Is Africa Different? World Bank Publications: Vol. 3496. [Google Scholar]
Figure 1. Selection of longest saccade in the main canvas.
Figure 1. Selection of longest saccade in the main canvas.
Jemr 15 00033 g001
Figure 2. Longest saccade selection with the rectangles.
Figure 2. Longest saccade selection with the rectangles.
Jemr 15 00033 g002
Figure 3. Sample Hindi reading test of 50 words.
Figure 3. Sample Hindi reading test of 50 words.
Jemr 15 00033 g003
Figure 4. Video and SLS region considered as first AOI set.
Figure 4. Video and SLS region considered as first AOI set.
Jemr 15 00033 g004
Figure 5. Two parts of the SLS area are second AOI set.
Figure 5. Two parts of the SLS area are second AOI set.
Jemr 15 00033 g005
Figure 6. Percentage of fixations in video without SLS.
Figure 6. Percentage of fixations in video without SLS.
Jemr 15 00033 g006
Figure 7. Percentage of fixations in video with SLS.
Figure 7. Percentage of fixations in video with SLS.
Jemr 15 00033 g007
Figure 8. Interaction effect of fixation rate across AOIs. The error bar indicates the standard deviation.
Figure 8. Interaction effect of fixation rate across AOIs. The error bar indicates the standard deviation.
Jemr 15 00033 g008
Figure 9. Fixations at AOI_SLS (left/right) of a poor reader.
Figure 9. Fixations at AOI_SLS (left/right) of a poor reader.
Jemr 15 00033 g009
Figure 10. Fixations at AOI_SLS (left/right) of a functional reader.
Figure 10. Fixations at AOI_SLS (left/right) of a functional reader.
Jemr 15 00033 g010
Table 2. Media descriptions of audio-visual clips.
Table 2. Media descriptions of audio-visual clips.
FemaleMaleTotal
Weak readers (Total)8155136
Poor readers381957
Functional weak readers433679
Table 3. Changes in eye movements in AOI_Video.
Table 3. Changes in eye movements in AOI_Video.
Pair No.Movie NamePercentage Change
1Baahubali27.87%
2Besharam19.48%
3Abhimaan34.35%
4Dhadkan22.84%
5Sholay27.55%
6Karz25%
Table 4. Paired-samples t-test.
Table 4. Paired-samples t-test.
Pair No.Movie NameResult
1Baahubalit(132) = -11.36, p < 0.05
2Besharamt(132) = 7.85, p < 0.05
3Abhimaant(132) = 14.3, p < 0.05
4Dhadkant(132) = 8.78, p < 0.05
5Sholayt(132) = -12.52, p < 0.05
6Karzt(132) = 9.86, p < 0.05
Table 5. Differences in time spent in AOI_Video (without SLS minus with SLS).
Table 5. Differences in time spent in AOI_Video (without SLS minus with SLS).
Movie NameTotal Time (secs)Time Differences (secs)
Baahubali6415.24
Besharam6710.91
Abhimaan6317.55
Dhadkan429.07
Sholay10023.36
Karz9421
Table 6. Changes in saccade amplitude (pixels) in AOI_Video.
Table 6. Changes in saccade amplitude (pixels) in AOI_Video.
Pair No.Movie NamePercentage of Change
1Baahubali26.82%
2Besharam19.22%
3Abhimaan31.48%
4Dhadkan19.95%
5Sholay25.89%
6Karz23.8%
Table 7. Average number of regressions.
Table 7. Average number of regressions.
Movie NamePRFWR
Baahubali13.9616.67
Besharam11.9812.36
Abhimaan17.1417.92
Dhadkan7.898.24
Sholay23.5926.97
Karz19.5619.15
Table 8. Comparisons of fixations at left AOI_SLS_Left.
Table 8. Comparisons of fixations at left AOI_SLS_Left.
Movie NamePRFWR
Baahubali61%52.8%
Besharam60%53%
Abhimaan50.2%40.5%
Dhadkan56%50%
Sholay57%54%
Karz57%54%

Share and Cite

MDPI and ACS Style

Arjun, S.; Kothari, B.; Shah, N.K.; Biswas, P. Do Weak Readers in Rural India Automatically Read Same Language Subtitles on Bollywood Films? An Eye Gaze Analysis. J. Eye Mov. Res. 2022, 15, 1-15. https://doi.org/10.16910/jemr.15.5.4

AMA Style

Arjun S, Kothari B, Shah NK, Biswas P. Do Weak Readers in Rural India Automatically Read Same Language Subtitles on Bollywood Films? An Eye Gaze Analysis. Journal of Eye Movement Research. 2022; 15(5):1-15. https://doi.org/10.16910/jemr.15.5.4

Chicago/Turabian Style

Arjun, Somnath, Brij Kothari, Nirav Kumar Shah, and Pradipta Biswas. 2022. "Do Weak Readers in Rural India Automatically Read Same Language Subtitles on Bollywood Films? An Eye Gaze Analysis" Journal of Eye Movement Research 15, no. 5: 1-15. https://doi.org/10.16910/jemr.15.5.4

APA Style

Arjun, S., Kothari, B., Shah, N. K., & Biswas, P. (2022). Do Weak Readers in Rural India Automatically Read Same Language Subtitles on Bollywood Films? An Eye Gaze Analysis. Journal of Eye Movement Research, 15(5), 1-15. https://doi.org/10.16910/jemr.15.5.4

Article Metrics

Back to TopTop