Next Article in Journal
Language Identification-Based Evaluation of Single Channel Speech Separation of Overlapped Speeches
Previous Article in Journal
Professional and Academic Digital Identity Workshop for Higher Education Students
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

The Effects of Social Desirability on Students’ Self-Reports in Two Social Contexts: Lectures vs. Lectures and Lab Classes

by
Konstantinos Lavidas
1,
Stamatios Papadakis
2,*,
Dionysios Manesis
3,
Anastasia Sofia Grigoriadou
4 and
Vasilis Gialamas
3
1
Department of Educational Sciences and Early Childhood Education, University of Patras, 26504 Patras, Greece
2
Department of Preschool Education, Faculty of Education, University of Crete, 74100 Crete, Greece
3
Department of Early Childhood Education, National and Kapodistrian University of Athens, 11527 Athens, Greece
4
Department of Economics, University Campus, University of Ioannina, 45110 Ioannina, Greece
*
Author to whom correspondence should be addressed.
Information 2022, 13(10), 491; https://doi.org/10.3390/info13100491
Submission received: 24 August 2022 / Revised: 20 September 2022 / Accepted: 9 October 2022 / Published: 11 October 2022

Abstract

:
Attempts to detect socially desirable responding bias have mainly focused on studies that explore sensitive topics. However, researchers concur that the sensitive character of the survey could be affected by the social context within which the research is conducted. Little research has been reported worldwide investigating the potential effects of social desirability on students’ self-reports, considering the social context within which the survey is conducted. In this paper, we investigate the potential effects of social desirability on students’ self-reports in two social contexts within which the survey was conducted. More specifically, with a sample of 111 Greek students, we explored the effects of social desirability on students’ attitudes towards statistics in two cases: when the questionnaire was administered to participating students after attending (a) lectures and (b) both lectures and laboratory classes. Only in the second case were the items’ attitudes toward statistics associated with a score of socially desirable responding; moreover, social desirability accounted for the relationship between attitudes toward statistics and perceived competence in mathematics. Implications and limitations are also discussed.

1. Introduction

Socially desirable responding (SDR) can be interpreted in terms of measurement bias [1]. The SDR is defined as the predisposition of survey respondents to present themselves in the best possible light, giving answers that are usually perceived as socially desired [2,3,4]. This tendency to answer in a way that is viewed sympathetically may lead to inaccurate self-reports and, ultimately, incorrect conclusions because the data are potentially biased [1,5].
Research on SDR has been ongoing for several decades [1,6,7,8,9]. Researchers from various backgrounds have provided insights into the debate about the nature of the phenomenon (whether a property of the respondent or the instrument used) [4,5], the motives and psychological mechanisms that underlie it, the survey variables that seem to influence its manifestation, the improvement measures that can be taken in the survey design [1,7,10], and the methods that may be employed to mitigate its impact [1,10]. Irrespective of the debate as to whether SDR constitutes a form of response bias or a variable of interest in itself, it seems to affect more or less all types of self-reports across all social sciences [1,11]. As a result, SDR continues to be a source of concern and attracts research attention and efforts, particularly in those fields that rely heavily on self-reports as a source of data, and even more so when some of the results are mixed or controversial [12].
Attempts to detect SDR bias have mainly focused on surveys that contain sensitive topics—that is, those that potentially pose a substantial threat for those involved [13,14], such as health, sexual, and drug-related behaviours [1,4,10]. Sensitive questions tend to lead to higher measurement error responses than questions on other topics [15]. The data collection mode is considered to be essential in explaining misreporting in sensitive studies. The main distinction is whether the questions are interviewer- or self-reported surveys. Self-report surveys have been found to reduce participants’ feelings of jeopardy and produce more honest answers to sensitive questions [10]. However, researchers argue that the sensitive character of research “seemingly inheres less in the topic itself and more in the relationship between that topic and the social context within which the research is conducted” [13] (p. 512). In this way, surveys about attitudes towards the teaching subject could transform the issue under research into a “sensitive topic”.
Self-report surveys of university students are among the most frequently used data sources because these data play an essential role in evaluating universities’ programs and affect the application of suitable policies [16]. Therefore, having valid and reliable data is vital for stakeholders such as faculty members and policymakers. However, although self-report surveys are frequently used in higher education, the presence of social desirability is rarely considered [16,17]. Recent studies on a broad spectrum of topics in higher education have found relationships between SDR and variables of interest [7,17,18]. Still, other studies in higher education have found no significant relationships with SDR [9], whereas others have found mixed results [12]. These controversial findings indicate the need further to explore the validity of surveys with university students, taking into account the social context within which the survey is conducted [13,16]. Moreover, to the best of the authors’ knowledge, no previous attempt has been reported worldwide to investigate the potential effects of social desirability on students’ self-reports considering the social context within which the research is conducted—the social contexts being (a) after lectures only or (b) after lectures combined with computer laboratory. This study explores whether students’ survey responses regarding a common but non-sensitive topic—such as students’ attitudes towards statistics—vary in terms of their susceptibility to SDR depending on the teaching setting employed (lecture only versus lecture and lab).
Subsequently, the theoretical background is presented, followed by the research objectives. Then, the method and the research procedure are presented. Lastly, this study’s results, limitations, practical implications, and future potential are discussed.

2. Theoretical Background

The most widely used teaching methods are lectures to small and large groups. Lectures have been “accused” of being ineffective in engaging students and addressing the varying needs of a vast and diverse audience [19]. In contrast, studies suggest that students strongly prefer a lab component when offered [19], and that the “hands-on” lab time is very effective in learning—for example, with statistical software [20,21]. Regarding teaching statistics, one study suggested that the most significant impact on final grades was found when computer-assisted instruction was incorporated into several class meetings [21]. It has been suggested that computer instruction reduces students’ anxiety and improves their general attitudes towards statistics [22].
Students’ attitudes toward statistics are important because there is evidence that they are related to the learning process and influence the learning outcome [23,24]. Although attitudes towards statistics are not considered to be a sensitive topic, the attitudes towards various subjects are often studied in education. In addition, it has been found that the teacher’s teaching approach often influences students’ attitudes towards the subjects they are taught [25,26]. Especially in smaller classes, such as a computer laboratory, where students have more opportunities for interaction with the teacher and one another [20], students may develop a social context that could improve their attitudes towards the subjects taught [27]. A possible explanation is that, through their positive responses, students may want to influence the assessment score that the teacher will assign to them [28]. On the other hand, in a lecture class, the emphasis is placed on knowledge transmission, so interaction is poor compared to the computer lab. Hence, it is crucial to investigate whether the social context developed in these different classes explains SDR’s effect on students’ attitudes toward statistics.

2.1. Partialling Out Social Desirability

SDR’s measurement error is a severe problem because it could provide an alternative explanation for the observed relationships between measures of different constructs [10]. Partialling out of measures of SDR is used to investigate their measurement error [1,14]. A high score on an SDR measure may indicate one of the following: (a) the respondents are engaged in socially desirable behaviours that influence their responses to what is being measured, (b) the respondents deliberately report a socially desirable behaviour to create an impression, or (c) the respondents are engaged in socially desirable behaviours, but their responses are sincere in terms of what is being measured [4,29]. Only the first two cases may influence the construct measured.
However, the total score of SDR by itself does not indicate the extent of the contamination [6,9]. A way to test the distortion of the scale of interest is by correlating the results with those from the scale used to measure SDR [1]. If no significant correlation is found, then the variables measured are probably not influenced by SDR [6], or the effect of SDR is insignificant if the correlation is negligible, i.e., less than |0.2| [29]. If the correlation is non-negligible, this is an indication of possible contamination [6,30], in which case it is essential to check whether SDR influences the fundamental correlation between the independent and dependent variables of interest [6,9,29]. Ganster et al. [6] introduced three models to describe the effects of SDR on the results of organisational behaviour research: (a) SDR may produce spurious correlations between study variables, (b) it may hide authentic relationships (suppression), and (c) it may moderate relationships between two other variables.

2.2. Research Objectives

The aim of our work was twofold:
  • To investigate the possible effects of socially desirable responses on university students’ attitudes towards statistics after attending an introductory course in statistics.
  • To highlight how the type of class that the students attend moderates the abovementioned effect.
We compared two different survey conditions: in the first condition, the survey was conducted with students who attended only the lectures, whereas in the second the survey was conducted with students who attended lectures combined with a computer laboratory.
In the following sections, we present the research procedure and analysis of our survey on whether university students’ self-reports on attitudes towards statistics are susceptible to socially desirable responses, considering the administration of the questionnaire in the two previous conditions.

3. Method

3.1. Research Procedure

This study employed a cross-sectional quantitative research method [31] (conducted in December 2019). Notably, a questionnaire was administered to students after nine weeks of teaching so that the students had some exposure to statistics. To increase the students’ participation in the research, we followed the guidelines suggested by Lavidas et al. [32]. The research protocol was approved by the Research Ethics Board designated by the University of Patras and, particularly, by the Department of Educational Science and Early Childhood Education. The participants gave their informed consent prior to the research. According to Tan et al. [33], web surveys are more likely to minimise SDB than offline surveys. Therefore, the data collection was conducted online using SurveyMonkey; the survey link was sent to each student via the learning management system that supported the lesson [34]. In a brief introduction, before the survey questions began, the students were informed that sincere answers were vital for the validity of the survey results. Moreover, ethical considerations concerning the privacy of individuals were carefully considered throughout the research process. Past studies have shown that the greater the degree of privacy and anonymity, the lower the level of SDR [10,35].

3.2. Participants and Teaching Settings

All students at the Department of Early Childhood Education at the University of Patras were enrolled in the statistics course and were asked to participate in the survey voluntarily. From them, 111 students were accepted to participate in the research and were randomly split into two groups. The first group (Lecture) comprised 61 students who attended only the class lectures. The second group (Lecture and Lab) consisted of 50 students who attended the lectures plus a two-hour computer laboratory per week. The participants ranged from 20 to 22 years old; they were at least in their third year of studies (juniors: 85%), and the vast majority were female students (96.4%).
In both cases, the course was a weekly compulsory introductory statistics course. The course content was identical for both groups and was taught by the same instructor (i.e., by the first author). The Lecture and Lab group, during the computer laboratory sessions, was divided into subgroups of 25 members. The students used spreadsheet software and worked independently (one student per computer). The lecturer gave them a list of tasks and demonstrated the solution on an interactive whiteboard. On the other hand, in the lecture group, the exercises were projected on a blackboard using presentation software or were solved manually by the instructor.

3.3. Research Instruments

A 41-item questionnaire was drawn up comprising closed-ended questions: a 28-item version of the Survey of Attitudes Toward Statistics (SATS-28) [36], 2 items for students’ perceived competence in mathematics [24,37], and an 11-item adapted version [8] of the Marlowe–Crowne Social Desirability (MCSD) Scale [2,3]. In terms of demographic factors, the respondents were asked to indicate their age and gender.
SATS-28 is one of the most widely used instruments to measure attitudes toward statistics [23,24,38]. It takes into account a multidimensional model that conceptualises attitudes towards statistics as a combination of four major components: (a) affect (6 items), i.e., the way the students feel about statistics (e.g., “I like statistics”); (b) cognitive competence, i.e., how competent they feel about their knowledge and skills in statistics (e.g., “I can learn statistics”); (c) value (9 items), i.e., how useful, relevant, and valuable statistics is in their everyday and professional life (e.g., “I use statistics in my everyday life”); and (d) difficulty (7 items), i.e., how difficult they believe it is to learn statistics (e.g., “Most people have to learn a new way of thinking to do statistics”). Students answer each question on a 7-point Likert-type scale ranging from 1 (I strongly disagree) to 7 (I strongly agree), with 4 being the neutral value. According to the creators of SATS-28 [38], Cronbach’s alpha coefficient appears to have high scores (>0.861) in the subscales “Affect”, “Cognitive Competence”, and “Value”. The corresponding internal consistency factor for the “Difficulty” subscale is 0.671, but this level is considered marginally sufficient. The validity and reliability of this version have also been tested on a similar sample of Greek students by Lavidas et al. [24].
Apart from SATS-28, the questionnaire included two items on perceived competence in mathematics. This variable was deemed necessary because many studies report that it is the main factor behind attitudes towards statistics [24,37]. Two questions were used to measure perceived competence in mathematics: (a) Were you good at high school maths? (b) How good are you at maths? Students responded to a 7-point Likert-type scale ranging from 1 (not good at all) to 7 (excellent) [24,37].
The use of separate SD scales prevails among the various methods proposed to detect, measure, or reduce SDR. The MCSD Scale (Crowne and Marlowe, 1960) is the most frequently employed SDR instrument [29]. To measure social desirability bias, a short version of the 11-item MCSD Scale [3,8] was distributed, and the respondents were asked to report whether a statement was true or false. “True” responses to items 5, 7, 9, 10, and 13 were added to “False” responses to items 1, 2, 6, 8, 11, and 12 to calculate the total score. This scale version was administered to students in a similar context by Lavidas and Gialamas [8], and its validity and reliability were established.

3.4. Data Analysis Strategy

We followed a two-level procedure to individually detect a possible SDR bias in the questionnaire. First, we calculated the bivariate product–moment correlation coefficients of the SDR scale with each of the 28 items of SATS-28, the 2 items measuring perceived competence in mathematics, and the 5 variables derived from there: 4 variables from the SATS scale (affect, cognitive competence, value, and difficulty) and 1 from the perceived competence in mathematics (PCM). Then, to assess the SDR effect on the correlation between the reported PCM (the independent variable) and the four dimensions of students’ attitudes towards statistics (the dependent variables), we used the techniques and the three models proposed by [6]. In more detail, we computed regressions for each of the four dimensions on both PCM and SDR. If the interaction between independent variable (X) and SDR (i.e., “X*SDR”) was non-zero while the zero-order correlations between SDR and the other variables were zero (unrelated), then the moderator model applied. If the beta coefficient for the independent variable (X) was zero or just less than the zero-order correlation between the independent (X) and dependent (Y) variables, then the spuriousness model applied. Finally, if the beta for the independent variable (X) was greater than the zero-order correlation between the independent (X) and dependent (Y) variables, then the suppression model was correct [6].

4. Results

Initially, to identify the multivariate outliers of the 28 statements of SATS-28, we relied on the Mahalanobis distance [39]. Three cases were identified as multivariate outliers (p < 0.001): one case from the Lecture group and two cases from the Lecture and Lab group. They were eliminated, narrowing down the sample size to 108 students (60 students from the Lecture group and 48 from the Lecture and Lab group). Before the data analysis, 19 negatively worded items were reversed so that, for all items, a higher score corresponded to a more positive attitude toward statistics [36]. Regarding the reliability of the SATS-28 scale, the internal consistency coefficient (Cronbach’s alpha) was satisfactory in the subscales affect (Total: 0.832. Lecture: 0.872, Lecture and Lab: 0.771), cognitive competence (Total: 0.804. Lecture: 0.876, Lecture and Lab: 0.699), and value (Total: 0.820. Lecture: 0.856, Lecture and Lab: 0.748), and marginally sufficient in the difficulty subscale (Total: 0.691. Lecture: 0.707, Lecture and Lab: 0.674). Regarding the reliability of the perceived competence in mathematics scale, the internal consistency coefficient (Cronbach’s alpha) was satisfactory (Total: 0.919. Lecture: 0.950, Lecture and Lab: 0.882). The internal consistency reliability of the SDR scale was estimated using the “Kuder-Richardson Formula 20” and was found to be marginally sufficient (Total: 0.648. Lecture: 0.648, Lecture and Lab: 0.655).
Finally, a new component variable was calculated for each dimension measured by SATS-28. We constructed four new variables based on the items’ scale (seven points). Similarly, a new variable for perceived competence in mathematics was derived. Lastly, an SDR variable was calculated. Initially, we present the correlations of the variables of interest with SDR; subsequently, the effects of SDR on the relationships between variables of interest are explored. Table 1 presents the descriptive statistics of the variables of interest in the two groups. In most cases, the mean scores of the variables in the two groups seemed similar, with a slight increase in the “Lecture” group.
Initially, we present the correlations of the variables of interest with SDR; subsequently, the effects of SDR on the relationships between variables of interest are explored.

4.1. Correlations between SDR and Variables of Interest

Following the analysis of the results from the Lecture and Lab group, SDR seemed to be significantly correlated with 11 statements on the SATS-28 scale and 2 statements on the perceived competence in mathematics scale (Table 2), and all of these cases were considered non-negligible (|r|>0.2). Non-negligible correlations were also observed with the component variables of the SATS-28 and the perceived competence in mathematics. The analysis of the Lecture group’s results did not indicate statistically significant correlations between SDR bias and the variables of interest, but did indicate some (with three statements) non-negligible correlations.

4.2. The Effect of SDR on the Correlation between the Independent and Dependent Variables

The moderator model was not observed in any of the four dimensions of the SATS-28 scale under either testing condition (i.e., the “Lecture” and “Lecture and Lab” groups). No statistically significant correlation was found in any of the tests for interaction between perceived competence in mathematics and SDR for each of the four dimensions of the attitudes scale. The spuriousness model was observed only in three dimensions (affect, cognitive competence, and difficulty), and only for the Lecture and Lab group (Table 3). More specifically, in these cases, the simple beta was slightly reduced (close to 0.08) compared to the simple correlation between the variables of interest, but remained statistically significant.

5. Discussion and Conclusions

The first aim of this study was to investigate the presence of SDR bias in the responses given by undergraduate students to the 28-item Survey of Attitudes Toward Statistics (SATS-28). The second aim was to compare the results yielded from the two survey conditions: Lecture versus Lecture and Lab. In the latter condition, there was a statistically significant, non-negligible correlation between SDR and most statements on the scale of interest. Regarding this correlation, it seems possible that students state a “false” image to the teacher through the questionnaire [28] in statements that are considered non-sensitive [1,4], such as their attitudes toward statistics and their perceived competencies in mathematics. This may indicate that students “reward” the teacher’s didactic intervention by expecting perhaps more leniency in the evaluation [28]. This “reward” might be because students develop an emotional climate through intense interaction with the teacher in the lab [27]. In addition, from the effects of SDR on the correlation between the four dimensions of the SATS-28 scale and the variable “perceived competence in mathematics”, we observed a slight variation of the effect (D (r-Beta): 7% to 8%) only in the case of three dimensions (affect, cognitive competence, and difficulty), which fits the spuriousness model. This means that the observed correlations between the independent and dependent variables are part of the shared variance of SDR. The effects of SDR on the relationships of variables of interest in higher education were also presented by Nauta [18], who found significant correlations of SDR with the variables that she examined (i.e., satisfaction and career decision self-efficacy). However, she did not observe a significant effect of SDR on the relationship between satisfaction and career decision self-efficacy [18].
In the case of the distribution of the questionnaire to students who attended only the lecture class, there was a non-negligible but statistically insignificant correlation of SDR bias with only three statements. Furthermore, the regression analysis did not indicate any alteration of the relationships between the variables in which we were interested that could be explained through SDR. Ferrari et al. [12] found similar results in higher education, finding no association between SDR and assessment of students’ perceptions of their institution’s values and mission.
Therefore, our findings suggest that SDR bias threatens the validity of student self-report responses to non-sensitive issues such as their attitudes toward statistics when the social context within which the survey is conducted could make it sensitive [13]. The effect of SDR on the correlation of the variables of interest was small but significant when the questionnaire was distributed to the Lecture and Lab group. The smaller group size and the more personal interaction with the teacher may have been more susceptible to socially desirable responses [28].
This study provides evidence that social desirability can influence students’ self-report responses to a non-sensitive issue. In our research, this non-sensitive issue was the students’ attitudes toward statistics, only for the students that attended the Lecture and Lab group. Considering that surveys worldwide are used to evaluate university programs’ effectiveness (Porter, 2011), SDR in such contexts must be examined [17]. However, to decrease respondents’ concerns about how their answers will be judged, questions should be presented in a neutral, inoffensive, and non-embarrassing way [1,10], and the privacy and the anonymity of the participants during the survey process should be ensured [35].
Future studies should evaluate the effects of social desirability on various survey instruments, considering the influence of social context, and propose methods to mitigate this form of bias. Thus, considering that SDR bias may be measured through well-documented instruments, we recommend administering a separate SDR scale along with the primary self-report scale in higher education. However, when SDR scales are applied, the potential contamination should not be assessed based only on the correlation between the SDR scale and the scale of interest. Rather, this should be coupled with analyses to detect the possible effects on the correlation between the variables of interest [6]. If SDR seems to affect the relationship between the dependent and independent variables under consideration, then SDR should be included as a covariate in further research [1].
Further empirical evidence is needed to relieve researchers’ concerns about self-report measures and to assist them in defining whether incipient contamination occurs. Moreover, more research is needed on how existing instruments may be improved to overcome concerns about SDR bias. It should be noted that a variety of SDR measurement scales and techniques are progressing over time [1,8]. Therefore, future research using at least one different SDR scale could confirm the previous findings.
Regarding our study, the main limitation was that we did not investigate whether shared variance between the independent and dependent variables of interest and the SDR variable included in the study was shared with other variables. Other variables might explain the difference between zero-order and partial correlations [1]. Although the course (i.e., duration, content) and the sample features were very similar, we cannot ignore the fact that differences in the two survey conditions may be considered to be confounding factors. Moreover, the small number of male participants prevented us from investigating whether gender might be a predictive factor of perceived social desirability. Finally, the SDR scale and the subscale “difficulty” had marginally acceptable Cronbach’s alphas. Therefore, a representative sample with students from different disciplines and universities including a more significant number of male students is necessary for future research.
Finally, despite the limitations of our study, we believe that our findings contribute to a better understanding of social desirability bias in responses to survey research and might be helpful for educators and researchers who investigate similar topics of interest in education.

Author Contributions

Conceptualisation, K.L. and S.P.; methodology, D.M.; software, A.S.G.; validation, K.L., V.G. and D.M.; formal analysis, K.L.; investigation, K.L.; resources, S.P.; data curation, A.S.G.; writing—original draft preparation, K.L.; writing—review and editing, S.P.; visualisation, D.M.; supervision, V.G.; project administration, K.L.; funding acquisition, S.P. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki and approved by the Institutional Review Board (or Ethics Committee) of Department of Educational Science and Early Childhood Education, University of Patras (protocol code 38225, 20/12/2021).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The data are available from the corresponding author upon reasonable request.

Acknowledgments

We would like to thank Emeritus Leonidas Sotiropoulos for our discussions in the initial conception of the idea. In addition, we thank our students of the Department of Educational Sciences and Early Childhood Education of the University of Patras for their participation in the study.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Podsakoff, P.M.; MacKenzie, S.B.; Lee, J.Y.; Podsakoff, N.P. Common method biases in behavioral research: A critical review of the literature and recommended remedies. J. Appl. Psychol. 2003, 88, 879–903. [Google Scholar] [CrossRef] [PubMed]
  2. Crowne, D.P.; Marlowe, D. A new scale of social desirability independent of psychopathology. J. Consult. Psychol. 1960, 24, 349–354. [Google Scholar] [CrossRef] [Green Version]
  3. Reynolds, W.M. Development of reliable and valid short forms of the Marlowe-Crowne Social Desirability Scale. J. Clin. Psychol. 1982, 38, 119–125. [Google Scholar] [CrossRef]
  4. Tourangeau, R.; Yan, T. Sensitive questions in surveys. Psychol. Bull. 2007, 133, 859–883. [Google Scholar] [CrossRef] [PubMed]
  5. McCrae, R.R.; Costa, P.T. Social desirability scales: More substance than style. J. Consult. Clin. Psychol. 1983, 51, 882–888. [Google Scholar] [CrossRef]
  6. Ganster, D.C.; Hennessey. H., W.; Luthans., F. Social Desirability Response Effects: Three Alternative Models. Acad. Manag. J. 1983, 26, 321–331. [Google Scholar] [CrossRef] [Green Version]
  7. Kelly, N.; Harpel, T.; Fontes, A.; Walters, C.; Murphy, J. An examination of social desirability bias in measures of college students’ financial behaviour. Coll. Stud. J. 2017, 51, 115–128. [Google Scholar]
  8. Lavidas, K.; Gialamas, V. Adaption and psychometric properties of the short forms Marlowe-Crowne social desirability scale with a sample of Greek university students. Eur. J. Educ. Stud. 2019, 6, 230–239. [Google Scholar] [CrossRef]
  9. Miller, A.L. Investigating social desirability bias in student self-report surveys. Educ. Res. Q. 2012, 36, 30–47. [Google Scholar]
  10. Krumpal, I. Determinants of Social Desirability Bias in Sensitive Surveys: A Literature Review. Qual. Quant. 2013, 47, 2025–2047. [Google Scholar] [CrossRef]
  11. Fischer, R. Social Desirability Bias and the Validity of Indirect Questioning. J. Consum. Res. 1993, 20, 303–315. [Google Scholar] [CrossRef]
  12. Ferrari, J.R.; Bristow, M.; Cowman, S.E. Looking good or being good? The role of social desirability tendencies in student perceptions of institutional mission and values. Coll. Stud. J. 2005, 39, 7–14. Available online: https://link.gale.com/apps/doc/A131318240/AONE?u=anon~cabc5eb2&sid=googleScholar&xid=34a3bb24 (accessed on 30 December 2019).
  13. Lee, R.M.; Renzetti, C.M. The problems of researching sensitive topics: An overview and introduction. Am. Behav. Sci. 1990, 33, 510–528. [Google Scholar] [CrossRef]
  14. Tan, H.C.; Ho, J.A.; Teoh, G.C.; Ng, S.I. Is social desirability bias important for effective ethics research? A review of literature. Asian J. Bus. Ethics 2021, 10, 205–243. [Google Scholar] [CrossRef]
  15. Koivula, A.; Räsänen, P.; Sarpila, O. Examining Social Desirability Bias in Online and Offline Surveys. In Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2019; pp. 145–158. [Google Scholar] [CrossRef]
  16. Porter, S.R. Do College Student Surveys Have Any Validity? Rev. High. Educ. 2011, 35, 45–76. [Google Scholar] [CrossRef]
  17. Bowman, N.A.; Hill, P.L. Measuring how college affects students: Social desirability and other potential biases in college student self-reported gains. New Dir. Inst. Res. 2011, 2011, 73–85. [Google Scholar] [CrossRef]
  18. Nauta, M.M. Assessing College Students’ Satisfaction With Their Academic Majors. J. Career Assess. 2007, 15, 446–462. [Google Scholar] [CrossRef]
  19. Fry, H.; Ketteridge, S.; Marshall, S. (Eds.) A Handbook for Teaching and Learning in Higher Education: Enhancing Academic Practice; Routledge: London, UK, 2009. [Google Scholar]
  20. Guardiola, J.; Duran-Hutchings, N.; Elsalloukh, H. Are Statistics Labs Worth the Effort?—Comparison of Introductory Statistics Courses Using Different Teaching Methods. Numeracy 2010, 3, 5. [Google Scholar] [CrossRef]
  21. Larwin, K.; Larwin, D. A Meta-Analysis Examining the Impact of Computer-Assisted Instruction on Postsecondary Statistics Education: 40 Years of Research. J. Res. Technol. Educ. 2011, 43, 253–278. [Google Scholar] [CrossRef]
  22. Ciftci, S.K.; Karadag, E.; Akdal, P. Instruction of Statistics via Computer-Based Tools: Effects on Statistics’ Anxiety, Attitude, and Achievement. J. Educ. Comput. Res. 2014, 50, 119–133. [Google Scholar] [CrossRef]
  23. Emmioğlu, E.S.M.A.; Capa-Aydin, Y. Attitudes and achievement in statistics: A meta-analysis study. Stat. Educ. Res. J. 2012, 11, 95–102. Available online: https://iase-web.org/documents/SERJ/SERJ11(2)_Emmioglu.pdf (accessed on 30 December 2019). [CrossRef]
  24. Lavidas, K.; Barkatsas, T.; Manesis, D.; Gialamas, V. A structural equation model investigating the impact of tertiary students’ attitudes toward statistics, perceived competence at mathematics, and engagement on statistics performance. Stat. Educ. Res. J. 2020, 19, 27–41. Available online: https://iase-web.org/documents/SERJ/SERJ19(2)_Lavidas.pdf?1593210287 (accessed on 30 December 2019). [CrossRef]
  25. Akinsola, M.K.; Olowojaiye, F.B. Teacher instructional methods and student attitudes towards mathematics. Int. Electron. J. Math. Educ. 2008, 3, 60–73. [Google Scholar] [CrossRef]
  26. Vogel, T.; Wanke, M. Attitudes and Attitude Change, 2nd ed.; Routledge: London, UK, 2016. [Google Scholar] [CrossRef]
  27. Reyes, M.R.; Brackett, M.A.; Rivers, S.E.; White, M.; Salovey, P. Classroom emotional climate, student engagement, and academic achievement. J. Educ. Psychol. 2012, 104, 700–712. [Google Scholar] [CrossRef] [Green Version]
  28. Basow, S.A.; Martin, J.L. Bias in student evaluations. In Effective Evaluation of Teaching: A Guide for Faculty and Administrators; Kite, M.E., Ed.; Society for the Teaching of Psychology—American Psychological Association: Washington, DC, USA, 2012; Available online: http://teachpsych.org/ebooks/evals2012/index.php (accessed on 30 December 2019).
  29. Steenkamp, J.B.E.; De Jong, M.G.; Baumgartner, H. Socially desirable response tendencies in survey research. J. Mark. Res. 2010, 47, 199–214. [Google Scholar] [CrossRef] [Green Version]
  30. Arnold, H.J.; Feldman., D.C. Social desirability response bias in self-report choice situations. Acad. Manag. J. 1981, 24, 377–385. [Google Scholar] [CrossRef]
  31. Bryman, A. Social Research Methods; Oxford University Press: New York, NY, USA, 2016. [Google Scholar]
  32. Lavidas, K.; Petropoulou, A.; Papadakis, S.; Apostolou, Z.; Komis, V.; Jimoyiannis, A.; Gialamas, V. Factors Affecting Response Rates of the Web Survey with Teachers. Computers 2022, 11, 127. [Google Scholar] [CrossRef]
  33. Tan, H.C.; Ho, J.A.; Kumarusamy, R.; Sambasivan, M. Measuring social desirability bias: Do the full and short versions of the Marlowe-Crowne Social Desirability scale matter? J. Empir. Res. Hum. Res. Ethics 2022, 17, 382–400. [Google Scholar] [CrossRef]
  34. Lavidas, K.; Komis, V.; Achriani, A. Explaining faculty members’ behavioral intention to use learning management. J. Comput. Educ. 2022. [Google Scholar] [CrossRef]
  35. Dodou, D.; de Winter, J. Social desirability is the same in offline, online, and paper surveys: A meta-analysis. Comput. Hum. Behav. 2014, 36, 487–495. [Google Scholar] [CrossRef]
  36. Schau, C.; Stevens, J.; Dauphinee, T.L.; Vecchio, A.D. The development and validation of the survey of attitudes toward statistics. Educ. Psychol. Meas. 1995, 55, 868–875. [Google Scholar] [CrossRef]
  37. Lavidas, K.; Manesis, D.; Gialamas, V. Investigation of the Statistical Anxiety Rating Scale psychometric properties with a sample of Greek students. Int. J. Educ. Psychol. 2021, 10, 116–142. [Google Scholar] [CrossRef]
  38. Schau, C. Students’ attitudes: The “other” important outcome in statistics education. In Proceedings of the Joint Statistical Meetings, San Francisco, CA, USA, 3–7 August 2003; pp. 3673–3681. [Google Scholar]
  39. Mahalanobis, P.C. On the generalized distance in statistics. Proc. Natl. Inst. Sci. India 1936, 2, 49–55. [Google Scholar]
Table 1. Descriptive statistics of the variables in the two groups.
Table 1. Descriptive statistics of the variables in the two groups.
Lecture and Lab (N = 48)Lecture (N = 60)
MinMaxMdnMSDMinMaxMdnMSD
SDR1.0011.007.007.042.351.0011.007.007.572.28
PCM2.0014.008.508.772.992.0014.007.007.172.72
Affect1.507.004.003.791.221.176.334.504.121.31
Cognitive competence1.837.004.424.411.011.336.504.674.571.19
Difficulty1.575.433.363.350.871.865.863.713.620.77
Value3.356.564.684.790.842.897.005.225.031.01
Notes: Perceived competence in mathematics (PCM), socially desirable responding (SDR), median (Mdn), mean (M), standard deviation (SD).
Table 2. Product–moment correlation coefficients; variables of interest with SDR in the two groups.
Table 2. Product–moment correlation coefficients; variables of interest with SDR in the two groups.
LectureLecture and Lab
Affect items1. I like statistics0.0990.382 **
2. I feel insecure when I have to do statistics problems0.2310.104
11. I get frustrated going over statistics tests in class0.0980.247
14. I am under stress during statistics classes−0.0250.183
15. I enjoy taking statistics courses0.2390.345 *
21. I am scared by statistics0.1810.320 *
Total Affect0.1710.383 **
Cognitive competence items3. I have trouble understanding statistics because of how I think0.0200.238
9. I have no idea of what is going on in statistics0.2530.305 *
20. I make many math errors in statistics0.0800.051
23. I can learn statistics0.0690.389 **
24. I understand statistics equations0.0760.286 *
27. I find it challenging to understand statistics concepts0.0270.403 **
Total Cognitive Competence0.1120.446 **
Difficulty items4. Statistics formulas are easy to understand.0.1040.484 **
6. Statistics is a complicated subject−0.107−0.037
17. Statistics is a subject quickly learned by most people−0.1190.187
18. Learning statistics requires a great deal of discipline−0.0020.170
22. Statistics involves massive computations−0.008−0.020
26. Statistics is highly technical−0.1640.122
28. Most people have to learn a new way of thinking to do statistics−0.1780.168
Total Difficulty −0.1030.253
Value items5. Statistics are worthless0.1860.226
7. Statistics should be a required part of my professional training0.0680.293 *
8. Statistical skills will make me more employable0.0160.322 *
10. Statistics is not helpful to the typical professional−0.0100.220
12. Statistical thinking is not applicable in my life outside my job0.0260.260
13. I use statistics in my everyday life−0.2470.120
16. Statistics conclusions are rarely presented in everyday life0.0080.088
19. I will have no application for statistics in my profession0.1330.182
25. Statistics is irrelevant in my life0.0590.355 *
Total Value 0.0390.406 **
PCMHow good were you in high school maths? 0.0500.409 **
How good are you at maths?−0.0290.496 **
Total PCM0.0120.476 **
Notes: perceived competence in mathematics (PCM), socially desirable responding (SDR), ** p < 0.01 and * p < 0.05.
Table 3. Beta and product–moment correlation coefficients among variables in the two groups.
Table 3. Beta and product–moment correlation coefficients among variables in the two groups.
tBetaSimple Correlation rΔ (r-Beta)
Dependent VariablesIndependent VariablesLectureLecture and LabLectureLecture and LabLectureLecture and LabLectureLecture and Lab
AffectPCM6.604 **3.231 **0.6490.4560.651 **0.535 **0.0020.079
SDR1.6671.1780.1640.166………………
Cognitive
competence
PCM6.984 **5.024 **0.6750.6100.676 **0.684 **0.0010.074
SDR1.0781.2780.1040.155………………
DifficultyPCM2.730 *1.3560.3380.2180.337 **0.289 *0.0010.071
SDR−0.8620.931−0.1070.150………………
ValuePCM1.216−1.2210.159−0.1860.1590.049….….
SDR0.2813.243 *0.0370.494………………
Notes: perceived competence in mathematics (PCM), socially desirable responding (SDR), ** p < 0.01 and * p < 0.05.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Lavidas, K.; Papadakis, S.; Manesis, D.; Grigoriadou, A.S.; Gialamas, V. The Effects of Social Desirability on Students’ Self-Reports in Two Social Contexts: Lectures vs. Lectures and Lab Classes. Information 2022, 13, 491. https://doi.org/10.3390/info13100491

AMA Style

Lavidas K, Papadakis S, Manesis D, Grigoriadou AS, Gialamas V. The Effects of Social Desirability on Students’ Self-Reports in Two Social Contexts: Lectures vs. Lectures and Lab Classes. Information. 2022; 13(10):491. https://doi.org/10.3390/info13100491

Chicago/Turabian Style

Lavidas, Konstantinos, Stamatios Papadakis, Dionysios Manesis, Anastasia Sofia Grigoriadou, and Vasilis Gialamas. 2022. "The Effects of Social Desirability on Students’ Self-Reports in Two Social Contexts: Lectures vs. Lectures and Lab Classes" Information 13, no. 10: 491. https://doi.org/10.3390/info13100491

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop