You are currently viewing a new version of our website. To view the old version click .
Information
  • Article
  • Open Access

12 January 2020

Comparing Web Accessibility Evaluation Tools and Evaluating the Accessibility of Webpages: Proposed Frameworks

Department of Computer Science, College of Computer Science and Engineering, Taibah University, Medina 42353, Saudi Arabia
This article belongs to the Section Information and Communications Technology

Abstract

With the growth of e-services in the past two decades, the concept of web accessibility has been given attention to ensure that every individual can benefit from these services without any barriers. Web accessibility is considered one of the main factors that should be taken into consideration while developing webpages. Web Content Accessibility Guidelines 2.0 (WCAG 2.0) have been developed to guide web developers to ensure that web contents are accessible for all users, especially disabled users. Many automatic tools have been developed to check the compliance of websites with accessibility guidelines such as WCAG 2.0 and to help web developers and content creators with designing webpages without barriers for disabled people. Despite the popularity of accessibility evaluation tools in practice, there is no systematic way to compare the performance of web accessibility evaluators. This paper first presents two novel frameworks. The first one is proposed to compare the performance of web accessibility evaluation tools in detecting web accessibility issues based on WCAG 2.0. The second framework is utilized to evaluate webpages in meeting these guidelines. Six homepages of Saudi universities were chosen as case studies to substantiate the concept of the proposed frameworks. Furthermore, two popular web accessibility evaluators, Wave and SiteImprove, are selected to compare their performance. The outcomes of studies conducted using the first proposed framework showed that SiteImprove outperformed WAVE. According to the outcomes of the studies conducted, we can conclude that web administrators would benefit from the first framework in selecting an appropriate tool based on its performance to evaluate their websites based on accessibility criteria and guidelines. Moreover, the findings of the studies conducted using the second proposed framework showed that the homepage of Taibah University is more accessible than the homepages of other Saudi universities. Based on the findings of this study, the second framework can be used by web administrators and developers to measure the accessibility of their websites. This paper also discusses the most common accessibility issues reported by WAVE and SiteImprove.

1. Introduction

Web Content Accessibility Guidelines 2.0 (WCAG 2.0) [] were developed to provide recommendations and guidance for creating accessible web content to meet the needs of different disabled users. Some countries have adapted these guidelines and used them as a law like section 508 in the United States []. The Saudi government established a program called “Yesser” to focus on digital transformation and the provision of e-services. One of the aspects that Yesser covers is creating accessible web content according to W3C guidelines [].
The evaluation of a website in terms of web accessibility is a hard task []. In the literature, there are many approaches to evaluating the accessibility of webpages. One of the most common approaches is to examine the web accessibility of a webpage using automatic evaluation tools. These tools can also be called web accessibility checkers, and the terms are used interchangeably in this paper. Furthermore, with the development of WCAG 2.0, various automated web accessibility evaluation tools have been widely used to determine to what extent a specific webpage meets accessibility guidelines, especially WCAG 2.0. The widespread utilization of automated web accessibility checkers is due to the fact that WCAG 2.0 was designed to be more testable and measurable than WCAG 1.0. Many organizations rely on automated web accessibility evaluation tools as the main indicator of their accessibility level in the absence of expert evaluators [].
Web accessibility checkers have been used to evaluate the web accessibility in different domains such as e-learning [], e-Commerce [] and banking []. Alshamari [] evaluated the accessibility of three well-known Arab e-commerce websites using AChecker, TAW, Eval Access, MAUVE and FAE tools. Yakup and Kemal [] evaluated the web accessibility of 25 official government websites in Turkey using AChecker, eXaminator, TAW, Total Validator, WAVE, Web Accessibility Assessment Tool, Eval Access, Cynthia Says, MAGENTA, HERA, Amp and Sort Site. Solomon and Ibrahim [] used TAW and site analyser to report accessibility issues of Nigerian e-government websites. Basel and Faouzi [] evaluated the homepages of 21 e-government websites using the TAW tool.
It is vital to evaluate websites using web accessibility evaluation tools, as they play an important role in assisting web developers with designing and developing more accessible websites. Despite the popularity of automated web accessibility evaluation tools in practice, few studies focus on comparing the performance and quality of web accessibility evaluation tools systematically. It is important to establish a systematic method of tool performance comparisons to help webmasters to select an appropriate evaluation tool for checking the compliance of their websites with guidelines. Similar to Brajnik [], one of the main motivations of this paper is to help web developers compare web accessibility evaluation tools and select an appropriate tool. In order to establish a systematic method for comparing the performance of accessibility evaluation tools, this study proposed a framework to compare the performance of web-accessibility checkers. The comparison of tools’ performances will be based on the idea of measuring how effective these tools are in detecting web accessibility issues compared to other tools, taking WCAG 2.0 guidelines into the consideration. Hence, a specific metric called a coverage error ratio (CER) metric is proposed.
Besides that, measuring the accessibility of given webpages compared to others based on WCAG 2.0 guidelines is necessary, as web developers require to measure whether the new version of webpages is more accessible than the old ones. For achieving this, another framework is proposed to evaluate how accessible webpages that meet these guidelines are. Therefore, a specific metric called the web accessibility accuracy (WAA) metric is proposed.
The organization of the paper is as follows. In Section 2, we present the background of WCAG 2.0. Section 3 describes the related works, including web accessibility evaluation approaches, the previous works that compare the performance of web accessibility tools, and relevant studies that evaluate the accessibility of websites in Saudi Arabia. Section 4 presents the proposed frameworks for comparing the performance of automatic web accessibility evaluation tools and measuring the accessibility of webpages. Section 5 describes the methodology that we followed in carrying out the study. Section 6 presents the results and discussion. Section 7 discusses the findings of the study and future works.

2. Background

WCAG 2.0 was developed to cover recommendations that make web content more accessible [], taking into account various web technologies []. It comprises 12 guidelines related to four main web accessibility principles: perceivable, operable, understandable, and robust. Each guideline comprises various success criteria (SCs).
The guidelines are considered a framework that guides developers and webmasters aiming to make content easily accessible for disabled users. It is important to comply with these guidelines to allow elderly and disabled users to access contents without any barriers. However, it is difficult to measure and test whether the contents comply with the guidelines or not. Therefore, SC were proposed to be testable manually or automatically using web accessibility checkers. WCAG 2.0 [] organized SC into levels of conformance: the minimum level of conformance (denoted by A) covering 25 SC, the intermediate level of conformance (denoted by AA) covering 13 SC as well as all the criteria in level A, and the highest level of conformance (AAA) covering all the criteria in level AA and 23 additional SC. In other words, each of the SC belongs to a level of conformance. A webpage is said to meet a specific level of conformance if it meets all the SC in that level and the preceding level.
The next two subsections describes two of well-known accessibility tools that will be used as case study to proof the concept of the proposed framework.

2.1. WAVE

WAVE is an automatic tool developed by WebAIM that allow users to enter the web address of a current site. It aims to help web developers check the accessibility of a given webpage to make it more accessible []. It adds icons to a webpage that allow users and experts to check potential accessibility issues. Red icons refer to accessibility errors, yellow icons indicate alerts, green icons indicate accessibility features, and all the light blue icons indicate structural, semantic, or navigational elements.

2.2. SiteImprove

SiteImprove is an online service that allows webmasters to check the web accessibility of a webpage with respect to WCAG 2.0. The browser extension of SiteImprove can be activated to automatically analyze the webpages for accessibility violations regarding the A, AA, or AAA level of the WCAG standards. It allow users to choose the conformance level, either A, AA, or AAA, and distributes the reported errors into different responsibilities. Each reported accessibility error is associated with a direct link to the corresponding WCAG manual to obtain a more detailed explanation of the reasons for the existence of errors. Furthermore, the reports produced by SiteImprove comes with suggestions to fix accessibility errors to gain compliance with WCAG 2.0. Similar to WAVE, SiteImprove has the ability to highlight the location of errors on the site itself and to point out the snippet source code in the browser’s developer tools.

4. The Proposed Frameworks for Comparing Tool and Web Accessibility

The next sections present two frameworks. The first is proposed to compare the performance of automatic web accessibility tools and the second is introduced to evaluate various webpages in terms of their accessibility.

4.1. Study 1: Framework for Comparing the Performance of Web Accessibility Tools

In this section, a framework for comparing the performance of web accessibility tools is proposed. It relies on collecting accessibility errors using a number of tools. In this study, a web accessibility error denotes a contradiction that may violate one or more WCAG 2.0 criteria. Let us assume that there are a number of tools to measure web accessibility for a given webpage. Therefore, with respect to various accessibility tools, it can be said that a specific tool is considered the one that performs the best if it has the ability to detect more web accessibility errors accurately. To achieve this, a coverage error ratio (CER) metric is proposed to be computed for each tool and a given webpage. The performance of web accessibility tools can be measured by computing a CER score for each tool. By comparing the attained CER scores for tools, the tool with the highest CER score can be considered the one with the best performance.
C E R = N u m b e r   o f   E r r o r s   d e t e c t e d   b y   a   g i v e n   t o o l   ( t ) T h e   t o t a l   n u m b e r   o f   E r r o r s   d e t e c t e d   b y   a l l   t o o l s
Figure 1 illustrates the general framework for comparing the performance of web accessibility tools with respect to WCAG 2.0 criteria for a given webpage. In general, this framework relies on a number of webpages and different web accessibility checkers. Therefore, it begins by providing numbers of webpages and various accessibility tools. Subsequently, for the given webpage, the web accessibility is evaluated using tools such as Wave, Achecker, SiteImprove, and so on. Once the webpages have been evaluated using the tools, accessibility errors are collected for each web accessibility checker for each webpage separately.
Figure 1. Evaluation framework for tools comparison.
The union set of errors is gathered manually by analyzing the reported errors obtained from various tools. The aim of this stage is to collect distinctive errors without redundancies. This phase requires non-expert users to map errors generated by different tools. For instance, SiteImprove showed this error message “Image link has no alternative text”, which corresponds to the following message “Linked image missing alternative text” generated by WAVE. This mapping process is required to collect the union errors generated by different tools. The union set of errors will be used as a reference to measure the effectiveness of each tool in finding these distinctive errors. Subsequently, the CER scores are computed, as described in Equation (1). This ratio represents the proportion of errors detected by the given tool divided by the union set of errors detected by all the tools.
Equation (7) can be rewritten using the following equation, where d e ( t x , w ) denotes the number of accessibility errors detected using the tool t x for a webpage w and t T d e ( t , w ) denotes the number of errors in the union set of errors detected by all the web accessibility evaluation tools for the given website w.
C E R ( t x , w ) = d e ( t x , w ) t T d e ( t , w )
The procedures of the comparison process to assess the performance of various tools is described in Algorithm 1. The comparison starts by selecting a number of web accessibility tools and webpages, as shown in lines (1) and (2). Then, an iteration over the webpages set is performed to collect accessibility errors using various tools. Following that, the Errors Map is initialized to record the detected errors via tools as illustrated in line (4). The UnionErrors set is defined to store the union set of errors detected by all tools for a given webpage. The next step is to iterate over tools to collect accessibility errors. The CollectErrors (t,w) function is responsible for collecting accessibility errors given the tool (t) and webpage (w). The tool and the detected errors will be recorded as a pair (t, detectedErrors) with the Errors as shown in line 9. Subsequently, the union errors are updated for the current webpage until there is no tool to select.
Algorithm 1: Computing CER score for each tool.
Information 11 00040 i001
Once the union errors have been collected, another iteration over the tool set is performed to compute the CER score for each tool and the current webpage under analysis, as illustrated in lines (12)–(14). Moreover, Errors (t) is a function that returns a list of errors detected by a tool t. The computeCER (Errors(t), UnionErrors) function computes the CER score based on errors detected by a given tool and the UnionErrors set. Then, a pair of (t, CERScore) is mapped to the current webpage (w) as illustrated in line (14). The comparison process is carried out for the next webpage until no further webpages can be selected to compute the CER scores.
Our proposed CER metric in this paper is focused on false negatives and benefits from all issues reported by the tools to select the reference issues stored in the union set. The set of reference issues contains all error reported by all tools. Furthermore, all reported issues are assumed to be true and the effectiveness of the tool will be measured based on its ability to detect more issues from the reference list.

4.2. Study 2: Framework for Evaluating Webpages in Terms of Web Accessibility

In this section, a framework for evaluating various webpages in terms of web accessibility is proposed. The aim of this framework is to establish a systematic approach to measure the accessibility level of webpages. Moreover, it can be considered as a performance indicator to compare different webpages in terms of accessibility. For instance, the Ministry of Education can utilize this approach to determine which universities’ webpages meet WCAG 2.0 criteria. It can be used to compare two versions of webpages for the same sites. The key benefit of this approach is that is helps the web developers decide whether the new version of a webpage is more accessible than the old version.
To determine the most accessible webpage among a number of webpages, web designers can rely on a number of evaluation tools to measure the web accessibility of webpages. Let us assume that there are a number of webpages of which the accessibility needs to be measured with respect to WCAG 2.0 and there are various accessibility checkers to detect all violations, we can say that a webpage h is the most accessible webpage in considering WCAG 2.0 compared to other webpages if it violated fewer of those guidelines. In other words, a webpage with fewer accessibility violations than other webpages is the most accessible webpage in terms of web accessibility.
To achieve this, the web accessibility accuracy (WAA) metric is proposed to compute it for each tool and a given webpage, as shown in Equation (9). The accuracy of web accessibility can be measured by computing a WAA score for each webpage. By comparing the attained WAA scores, the webpage with the highest WAA score can be considered the most accessible one. In other words, a webpage with the highest WAA score among a list of webpages is the most accessible one according to WCAG 2.0.
W A A = 1 N u m b e r   o f   E r r o r s   i n   t h e   U n i o n E r r o r s   s e t   f o r   t h e   g i v e n   w e b p a g e N u m b e r   o f   E r r o r s   i n   t h e   U n i o n E r r o r s   s e t   f o r   a l l   w e b p a g e s
Figure 2 shows the framework for evaluating the web accessibility for any webpage with respect to WCAG 2.0. In general, the evaluation relies on gathering web accessibility errors that violated WCAG 2.0 criteria based on multiple known web accessibility checkers. Then, a WAA metric is proposed to evaluate the web accessibility of a given webpage. Similar to the previous comparison framework described in Section 4.1, the proposed framework in this section relies on webpages and different web accessibility checkers. Therefore, it begins by providing various webpages and various accessibility tools. It aims to measure and compare the web accessibility of webpages. First, the web accessibility for the given webpage is measured using tools such as Wave, SiteImprove, and so on. Second, for each accessibility checker tool, violations with respect to WCAG 2.0 are gathered. Third, the union of accessibility errors detected by various tools is collected separately for each webpage. The aim of this step is to find distinctive errors (a reference set of accessibility issues) collected without any redundancies using different tools for the webpage under assessment. The union set of errors will be used to assess the accessibility of the webpage with respect to WCAG 2.0. subsequently, the WAA is computed, as described in Equation (9). This ratio represents the proportions of union errors detected by all tools for the current webpage divided by the total number of errors in all union sets for all webpages involved in the comparison process.
Figure 2. Evaluation framework for webpage accessibility comparison.
Equation (9) can be rewritten using the following equation, where t T d e ( t , w y ) denotes the number of errors in the union set of errors detected by all web accessibility evaluation tools for the given website W y and Σ w W t T d e ( t , w ) denotes the number of errors in all union sets of errors detected by all web accessibility evaluation tools for all webpages W.
W A A ( t , w y ) = 1 t T d e ( t , w y ) Σ w W t T d e ( t , w )
The computation of web accessibility accuracies of various webpages is presented in Algorithm 2. The evaluation framework is provided with distinctive webpages and different web accessibility checkers, as shown in lines (1) and (2). The aim of entering both pices of information is to measure the web accessibility of the webpages. Then, for each webpage, the union of web accessibility errors is collected for all tools, as illustrated in lines (6)–(10). Subsequently, the AllUnionErrors Map is updated by recording the collection of union errors for each webpage.
Algorithm 2: Computing WAA score for each webpage.
Information 11 00040 i002
The second iteration shown in lines (13)–(16) aims to compute the WAA score, as described in Equation (10). The AllUnionErrors(w) function returns the set of union errors detected by all tools for the webpage of interest. Furthermore, the GetAllUnionErrors(AllUnionErrors) function is responsible for returning all the union errors for all webpages. The computeWAA function computes the WAA score by dividing the number of errors in the union set for the homepage under assessment by the total number of errors in all the union sets for all the webpages. Then, a pair (w, WAA) is added to the WAAScores Map to store the WAA score for the current homepage under assessment as illustrated in line (15).

5. Experimental Methodology

This section presents the methodology of our experiments. The conducted experiments are the proof of concept for the proposed frameworks. Therefore, six homepages of Saudi public universities were selected in our experiments as case studies. We followed the methodology used by Al-Khalifa [,], Alshamari [] and Rana et al. [] to evaluate only the homepages. The reason behind selecting the homepages only is that they are indicators for other webpages and the starting points for visitors. Moreover, two distinctive tools were selected to measure their performance using Algorithm 1 described in Section 4.2. We also evaluate the web accessibility of the six homepages using Algorithm 2.
The maturity level indicator of the electronic transformation of government core services is computed by Yesser for all Saudi public institutions. According to the service maturity indicators report generated by Yesser, the education and research sector in KSA has been classified into three categories according to their performance in providing e-services. The green (excellent) category includes universities and research centers that the performance ratio ranged from 85% to 100%. The yellow (average) category consists of educational institutions of which the ratio varied between 60% and 84%. The red (poor) category includes institutions of which the performance ratio ranged from 0% to 59%. Table 1 summarizes the maturity level indicator for various university websites. We selected these six universities in our experiments as they are distinctive samples from each category defined by Yesser. In other words, we have relied on the maturity level indicator to select these six homepages. This indicator works as key performance indicator and has been computed and publicly published by Yesser to measure the maturity level of electronic transformation of services provided by Saudi universities. In this way, we selected two homepages from each categories (excellence–average–poor).
Table 1. Maturity level indicator for e-services provided by various universities.
Moreover, two web accessibility checkers were selected, namely SiteImprove and Wave, in the conducted experiments. The reason for selecting these checkers is that they are free, open source, and descriptive tools where accessibility issues are described alongside with relevant violated SCs. Both tools allow evaluators to navigate accessibility issues on the webpages and the source codes as well.

6. Experimental Results and Discussion

6.1. Study 1: Result of Comparing Web-Accessibility Tools

The bar plots of the CER scores computed for all the university homepages are shown in Figure 3. It is apparent from Figure 3 that SiteImprove outperformed Wave in five homepages. However, Wave detected more accessibility errors than SiteImprove in the Taibah University homepage. This is due to the fact that the Wave checker detected ten empty links, and these were not discovered by SiteImprove.
Figure 3. Coverage error ratio computed for six universities’ homepages.

6.2. Study 2: Result of Evaluating Webpages in Terms of Web Accessibility

The bar plots illustrated in Figure 4 summarize the WAA scores attained for each homepage. From this figure, it is obvious that the homepage of Taibah University is more accessible than other homepages as its WAA score is 96.16%, which is the highest score. This means that the homepage of Taibah University violates fewer guidelines in WCAG 2.0 than other homepages. This is an interesting outcome as Taibah University was classified in Yesser in the red (poor) category in providing e-services. This indicates that the IT center is aware of WCAG 2.0. Conversely, the homepages of Prince Sattam University and King Saud University attained lower WAA scores than Taibah University even though they belong to the green (excellent) category in Yesser.
Figure 4. Web accessibility accuracy computed for six universities homepages.

6.3. Accessibility Issues in Saudi Universities Homepages

Table 2 summarizes the accessibility errors detected by WAVE and SiteImprove for the Taibah University homepage. The corresponding SC for the reported accessibility issues are shown in Table 2. One of the main advantages of WAVE and SiteImprove is that they describe accessibility errors with the related WCAG 2.0 criteria and provide guidelines to fix errors. It is clear that ten links do not contain descriptive texts, causing difficulties in describing different links for disabled users who use a screen reader, Braille, or text. SiteImprove detected two non-distinguishable links, which means the same link text is used for multiple links navigating to different destinations on the current webpage.
Table 2. Accessibility errors detected for Taibah University.
Table 3 shows the summary accessibility errors obtained for the Prince Sattam University homepage. It is obvious that SiteImprove detected more errors than WAVE. The number of elements that are not highlighted on focus is 87. Errors of this type cause a difficulty for keyboard users to highlight focused elements in a webpage, which aim to tell users where they are on the page. It is vital to mention that WAVE failed to detect errors of this type. Similar to the homepage of Taibah University, the homepage of Prince Sattam University has 21 empty links. WAVE describes this type of error as “ A link contains no text”. One of the main findings in the accessibility issues on the Prince Sattam University homepage is that multiple links should be combined. These errors occur for adjacent links pointing to the same destination in the case where one has a textual hyperlink and the other is associated with an iconic representation of the same link.
Table 3. Accessibility errors detected for Prince Sattam University.
Table 4 reports the accessibility errors detected by WAVE and SiteImprove for the King Khaled University homepage. Fifteen empty heading errors were detected by both tools, which means there are 15 heading tags, but the text is empty. However, both tools expressed these errors using different warning messages. These errors violate three criteria, 1.3.1, 2.4.1, and 2.4.6, according to the SiteImprove checker. The number of elements that are not highlighted on focus in the homepage of King Khaled University is 96. Furthermore, 11 images do not have an alt attribute. It is noted that each accessibility checker tool describes the errors in their own way. Moreover, there is a difference between the two web accessibility checker tools in the way of describing the violated criteria related to each error.
Table 4. Accessibility errors detected for King Khaled University.
Table 5 summarizes the accessibility errors detected by WAVE and SiteImprove for the King Saud University homepage. It is apparent that 13 links do not contain text and are considered by Wave as accessibility errors. The conducted experiments show that this type of error is common in all the homepages except the homepage of King Fahad University. As Table 5 shows, the number of elements that are not highlighted on focus is 57, and these are considered as accessibility errors that are detected by the SiteImprove checker.
Table 5. Accessibility errors detected for King Saud University.
The accessibility errors determined by WAVE and SiteImprove for the homepage of King Fahad University are illustrated in Table 6. It is clear that 15 images do not have the correct alternative text. Similar to other universities in the conducted study, the number of elements that are not highlighted on focus is 81. The number of images with no alt attribute detected by SiteImprove is 24, whereas WAVE detected only four. There are ten accessibility issues related to use of presentational attributes, in which attributes such as ’border’ and ’align’, are used in the HTML tags and these attributes should be used CSS for these attributes.
Table 6. Accessibility errors detected for King Fahad University.
Table 7 summarizes the accessibility errors detected by WAVE and SiteImprove for the Najran University homepage. It is clear that ten links do not contain text, as reported only by WAVE. As can be seen in Table 7, a number of text hyperlinks are not distinguishable (used the same link text) as they are pointing to different destinations. There are nine instances of select box without a descriptive title, and this should be fixed to allow users utilizing assistive technologies to know what the select box menu is for.
Table 7. Accessibility errors detected for Najran University.

6.4. General Finding

Among the six university homepages evaluated, all failed to add alternative text for image links. The homepage of King Fahad University has 22 image links without alternative texts, reaching a higher number for this type of failure than other homepages that were included in the study. In 2010, Al-Khalifa [] stated that one of the three most common accessibility errors encountered in governmental homepages was that they did not add a text alternative for non-text elements. This may be attributed to the web developers’ lack of knowledge of the importance of alternative texts for images [].
Furthermore, all the university homepages had accessibility issues in that they did not add text for links describing the functionality or the target of links. The homepage of Prince Sattam University has 21 empty links, more than any other homepage. The importance of adding descriptive text for links is to aid people using a screen reader, Braille, or a text browser to distinguish different links []. Moreover, all the university homepages failed to distinguish between links in the same webpage, as the same link texts are used.
Four university homepages failed to meet SC 2.4.7 that focuses on highlighting the components while the user uses keyboard navigation. It is important to mention that WAVE is not able to detect this type of accessibility issue. The result of accessibility evaluation showed that the homepages of Prince Sattam University, King Khalid University, King Saud University, and King Fahad University had 87, 96, 57, and 81 issues, respectively related to elements that are not highlighted on focus (no keyboard accessibility). Various studies [,,] showed that this type of accessibility issue is considered as one of the most common accessibility violations found in Saudi e-government websites. We recommend that web developers should ensure that elements receiving keyboard focus are highlighted on focus.
It is vital to find the common accessibility issues with respect to the four main principles. With regard to the operable principle, Al-Faries et al. [] stated that the most common violated guideline is 2.4. Furthermore, the conducted experiment showed that guideline 2.4 is the most common violated guideline, especially the 2.4.4 criterion. With respect to the understandable principle, Al-Faries et al. [] showed that the 3.2.5 criterion is the most common violated criterion. However, in the conducted experiment, this criterion is violated less than the 3.3.2 criterion. Regarding the robust accessibility principle, guideline 4.1 is intended to support compatibility with assistive technologies such as screen readers. The finding of the conducted experiment showed that the 4.1.2 criterion tend to be violated less than other criteria. According to Al-Khalifa [,], the major violations include the following: no text alternatives, no keyboard access, and no language identification. These findings are similar to those of our study. Moreover, the conducted study showed that other violations such as empty link and empty heading are very common in all the homepages.

7. Conclusions and Future Works

This study set out to propose novel frameworks in terms of tool comparison and webpage accessibility evaluation. WAVE and SiteImprove were selected as they are well-known tools and utilized to substantiate the concepts of the proposed frameworks. CER and WAA metrics were proposed as measurements for both frameworks. The CER metric was proposed to measure the capability of tools in detecting accessibility issues. The CER scores demonstrate the capability of SiteImprove compared to WAVE in detecting web accessibility issues. One of the main advantages of the tool comparison framework is the ability to compare more than two tools by implementing the same steps and utilizing the CER equation to compare tools’ performance. We recommend the webmasters and developers use multiple efficient web-accessibility tools in order to detect a variety of accessibility barriers for disabled users. In this context, selecting the most efficient tools can be performed relying upon CER scores that can be computed for multiple accessibility tools.
The WAA metric is proposed as an indicator of the accessibility level for a given webpage. In this study, we use the WAA metric to compare six homepages of Saudi universities to determine which homepage is most accessible with respect to WCAG 2.0. One can employ this metric to compare two versions of webpages for the same site. The key benefit of this approach is that is helps the web developers decide whether the new version of the webpage is more accessible than the old version. Section 6.4 summarized a general finding based on analysing accessibility issues reported by both WAVE and SiteImprove tools. Based on that, the majority of homepages have accessbility issues related to empty links, empty heading, image links without alternative texts, and missing alt attribute for images.
One future direction is to compare the proposed WAA metric with other relevant metrics reported in Section 3.3. An interesting future extension of the study would be creating online databases that contain all possible web accessibility violations and the corresponding error messages generated by all tools for each potential violation. This could be updated by the developers of web accessibility checkers. Once such a database exists, we could train various classifiers such as support vector machines to use all possible errors and their target categories such as perceivable, operable, understandable, and robust. Thus, the classifier will be able to categorize all accessibility issues without any human intervention. Furthermore, this classifier could be integrated into the proposed frameworks presented in this study to make them fully automated in computing the CER and WAA metrics.
Moreover, one could apply artificial intelligence and machine learning methods to help the IT manager assign each web accessibility violation to the person responsible (webmaster, editor, developer) for it to be fixed. This may include crawling distinct web accessibility tools to provide the optimal solution for any web accessibility error.

Funding

This research received no external funding.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Caldwell, B.; Cooper, M.; Reid, L.G.V.; Erheiden, G. Web Content Accessibility Guidelines (WCAG) 2.0. 2008. WWW Consortium (W3C). Available online: https://www.w3.org/TR/WCAG20/ (accessed on 20 August 2019).
  2. United States Lawsoverview of the Rehabilitation Act of 1973 (Section 504 And 508). Available online: https://webaim.org/articles/laws/usa/rehab (accessed on 15 September 2019).
  3. National Government Website Standard. Available online: https://www.yesser.gov.sa/EN/Methodologies/NATIONAL_ICT_DIGITALSTANDARDS_PRINCIPLES/Pages/National-Government-Website-Standard.aspx (accessed on 13 September 2019).
  4. Brewer, J. Web accessibility highlights and trends. In Proceedings of the 2004 International Cross-disciplinary Workshop on Web Accessibility (W4A ’04), New York, NY, USA, 17–22 May 2004; pp. 51–55. [Google Scholar] [CrossRef]
  5. Vigo, M.; Brown, J.; Conway, V. Benchmarking web accessibility evaluation tools: measuring the harm of sole reliance on automated tests. In Proceedings of the 10th International Cross-Disciplinary Conference on Web Accessibility, Rio de Janeiro, Brazil, 13–15 May 2013; ACM: New York, NY, USA, 2013; p. 1. [Google Scholar]
  6. Lundqvist, S.; Ström, J. Web Accessibility in E-Learning: Identifying And Solving Accessibility Issues for Wcag 2.0 Conformance in an E-Learning Application. Student Thesis. 2018. Available online: http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-148913 (accessed on 25 September 2018).
  7. Alshamari, M. Accessibility evaluation of arabic e-commerce web sites using automated tools. J. Softw. Eng. Appl. 2016, 9, 439–451. [Google Scholar] [CrossRef]
  8. Kaur, A.; Dani, D. Banking websites in india: An accessibility evaluation. CSI Trans. ICT 2014, 2, 23–34. [Google Scholar] [CrossRef]
  9. Akgül, Y.; Vatansever, K. Web accessibility evaluation of government websites for people with disabilities in turkey. J. Adv. Manag. Sci. 2016, 4, 201–210. [Google Scholar] [CrossRef]
  10. Adepoju, S.A.; Shehu, I.S.; Bake, P. Accessibility evaluation and performance analysis of e-government websites in nigeria. J. Adv. Inf. Technol. 2016, 7. [Google Scholar] [CrossRef]
  11. AlMourad, B.; Kamoun, F. Accessibility evaluation of dubai e-government websites: Findings and implications. J. -Gov. Stud. Best Pract. 2013, 2013, 978647. [Google Scholar]
  12. Brajnik, G. Comparing accessibility evaluation tools: A method for tool effectiveness. Univers. Access Inf. Soc. 2004, 3, 252–263. [Google Scholar] [CrossRef]
  13. Martín, A.; Cechich, A.; Rossi, G. Comparing Approaches to Web Accessibility Assessment. In Handbook of Research on Web Information Systems Quality; IGI Global: Hershey, PA, USA, 2008; pp. 181–205. [Google Scholar]
  14. Shadi Abou-Zahra, N.S.; Keen, L. Selecting Web Accessibility Evaluation Tools. Available online: https://www.w3.org/WAI/test-evaluate/tools/selecting/ (accessed on 5 May 2018).
  15. Eggert, E.; Abou-Zahra, S. Web Accessibility Evaluation Tools List. Available online: https://www.w3.org/WAI/ER/tools/index (accessed on 5 July 2018).
  16. Al-Khalifa, H.S. Wcag 2.0 semi-automatic accessibility evaluation system: Design and implementation. Comput. Inf. Sci. 2012, 5, 73. [Google Scholar] [CrossRef][Green Version]
  17. Abuaddous, H.Y.; Jali, M.Z.; Basir, N. Web accessibility challenges. Int. J. Adv. Comput. Sci. Appl. 2016, 7, 172–181. [Google Scholar] [CrossRef]
  18. Abanumy, A.; Al-Badi, A.; Mayhew, P. e-government website accessibility: In-depth evaluation of Saudi Arabia and oman. Electron. J. Gov. 2005, 3, 99–106. [Google Scholar]
  19. MasoodRana, M.; Fakrudeen, M.; Rana, U. Evaluating web accessibility of university web sites in the kingdom of Saudi Arabia. Int. J. Technol. Knowl. Soc. 2011, 7, 1–15. [Google Scholar]
  20. Al-Khalifa, H.S.; Al-Kanhal, M.; Al-Nafisah, H.; Al-soukaih, N.; Al-hussain, E.; Al-onzi, M. A pilot study for evaluating arabic websites using automated wcag 2.0 evaluation tools. In Proceedings of the 2011 International Conference on Innovations in Information Technology (IIT 2011), Abu Dhabi, United Arab, 25–27 April 2011; pp. 293–296. [Google Scholar]
  21. Alahmadi, T.; Drew, S. An evaluation of the accessibility of top-ranking university websites: Accessibility rates from 2005 to 2015. In Proceedings of the DEANZ Biennial Conference, Hamilton, New Zealand, 17–20 April 2016; pp. 224–233. [Google Scholar]
  22. Brajnik, G.; Yesilada, Y.; Harper, S. The expertise effect on web accessibility evaluation methods. Hum.–Comput. Interact. 2011, 26, 246–283. [Google Scholar]
  23. Hong, S.; Katerattanakul, P.; Lee, D. Evaluating government website accessibility: Software tool vs. human experts. Manag. Res. News 2007, 31, 27–40. [Google Scholar] [CrossRef]
  24. Elkabani, I.; Hamandi, L.; Zantout, R.; Mansi, S. Toward better web accessibility. In Proceedings of the 2015 5th International Conference on Information & Communication Technology and Accessibility (ICTA), Marrakech, Morocco, 21–23 December 2015; pp. 1–6. [Google Scholar] [CrossRef]
  25. Grantham, J.; Grantham, E.; Powers, D. Website accessibility: An australian view. In Proceedings of the Thirteenth Australasian User Interface Conference (AUIC ’12), Melbourne, Australia, 31 January–3 February 2012; Volume 126, pp. 21–28. Available online: http://dl.acm.org/citation.cfm?id=2512125.2512129 (accessed on 9 March 2019).
  26. Alayed, A.S. A Framework and Checklist for Localised Web Content Accessibility Guidelines for Arabic University Websites in Saudi Arabia. Ph.D. Thesis, University of Southampton, Southampton, UK, 2018. [Google Scholar]
  27. Petrie, H.; Kheir, O. The relationship between accessibility and usability of websites. In Proceedings of the SIGCHI Conference on Human Factors In Computing Systems, San Jose, CA, USA, 28 April–3 May 2007; ACM: New York, NY, USA, 2007; pp. 397–406. [Google Scholar]
  28. Petrie, H.; Bevan, N. The evaluation of accessibility, usability, and user experience. Univers. Access Handb. 2009, 1, 1–16. [Google Scholar]
  29. Petrie, H.; Hamilton, F.; King, N.; Pavan, P. Remote usability evaluations with disabled people. In Proceedings of the SIGCHI Conference on Human Factors In Computing Systems, Montreal, QC, Canada, 22–27 April 2006; ACM: New York, NY, USA, 2006; pp. 1133–1141. [Google Scholar]
  30. Kumar, K.L.; Owston, R. Evaluating e-learning accessibility by automated and student-centered methods. Educ. Technol. Res. Dev. 2016, 64, 263–283. [Google Scholar] [CrossRef]
  31. Latif, M.H.A.; Masrek, M.N. Accessibility evaluation on malaysian e-government websites. J. E-Gov. Stud. Best Pract. 2010, 2010, 935272. [Google Scholar]
  32. Al-Khalifa, H.S. Exploring the accessibility of Saudi Arabia e-government websites: A preliminary results. In Proceedings of the 4th International Conference on Theory and Practice of Electronic Governance (ICEGOV ’10), Cairo, Egypt, 1–4 December 2008; ACM: New York, NY, USA, 2010; pp. 274–278. [Google Scholar] [CrossRef]
  33. Al-Khalifa, H.S. The accessibility of Saudi Arabia government web sites: An exploratory study. Univers. Access Inf. Soc. 2012, 11, 201–210. [Google Scholar] [CrossRef]
  34. Al-Khalifa, H.S.; Baazeem, I.; Alamer, R. Revisiting the accessibility of Saudi Arabia government websites. Univers. Access Inf. Soc. 2017, 16, 1027–1039. [Google Scholar] [CrossRef]
  35. Khan, M.A.; Buragga, K.A. Effectiveness of accessibility and usability of government websites in Saudi Arabia. Can. J. Pure Appl. Sci. 2010, 4, 1127–1131. [Google Scholar]
  36. Alotaibi, S.J. Evaluating the blackboard system based on web accessibility and usability guidelines. Int. J. Digit. Soc. 2015, 6, 1094–1101. [Google Scholar] [CrossRef]
  37. Sullivan, T.; Matson, R. Barriers to use: Usability and content accessibility on the web’s most popular sites. In Proceedings of the 2000 Conference on Universal Usability (CUU ’00)y, Arlington, VA, USA, 16–17 November 2000; ACM: New York, NY, USA, 2000; pp. 139–144. [Google Scholar] [CrossRef]
  38. Parmanto, B.; Zeng, X. Metric for web accessibility evaluation. J. Am. Soc. Inf. Sci. Technol. 2005, 56, 1394–1404. [Google Scholar] [CrossRef]
  39. Cluster, W. Unified Web Evaluation Methodology (UWEM 10), 2006. In Proceedings of the 11th International Conference (ICCHP 2008), Linz, Austria, 9–11 July 2008; pp. 394–401. [Google Scholar]
  40. Bühler, C.; Heck, H.; Perlick, O.; Nietzio, A.; Ulltveit-Moe, N. Interpreting results from large scale automatic evaluation of web accessibility. In Computers Helping People with Special Needs; Miesenberger, K., Klaus, J., Zagler, W.L., Karshmer, A.I., Eds.; Springer: Berlin/Heidelberg, Germany, 2006; pp. 184–191. [Google Scholar]
  41. Song, S.; Bu, J.; Shen, C.; Artmeier, A.; Yu, Z.; Zhou, Q. Reliability aware web accessibility experience metric. In Proceedings of the Internet of Accessible Things (W4A ’18), Lyon, France, 23–25 April 2018; ACM: New York, NY, USA, 2018; pp. 24:1–24:4. [Google Scholar] [CrossRef]
  42. Al-Faries, A.; Al-Khalifa, H.S.; Al-Razgan, M.S.; Al-Duwais, M. Evaluating the accessibility and usability of top saudi e-government services. In Proceedings of the 7th International Conference on Theory and Practice of Electronic Governance (ICEGOV ’13), Seoul, Korea, 22–25 October 2013; ACM: New York, NY, USA, 2013; pp. 60–63. [Google Scholar] [CrossRef]
  43. Alturki, U.T.; Aldraiweesh, A. Evaluating the usability and accessibility of lms “blackboard” at king saud university. Contemp. Issues Educ. Res. 2016, 9, 33–44. [Google Scholar] [CrossRef]

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.