Next Article in Journal
Fuzzy Logic as a Tool for Assessing Students’ Knowledge and Skills
Previous Article in Journal
Improving Science Assessments by Situating Them in a Virtual Environment
Educ. Sci. 2013, 3(2), 193-207; doi:10.3390/educsci3020193
Article

Applying Models to National Surveys of Undergraduate Science Students: What Affects Ratings of Satisfaction?

* ,
 and
Received: 25 February 2013; in revised form: 21 May 2013 / Accepted: 22 May 2013 / Published: 30 May 2013
View Full-Text   |   Download PDF [466 KB, uploaded 30 May 2013]   |   Browse Figure
Abstract: Many countries use national-level surveys to capture student opinions about their university experiences. It is necessary to interpret survey results in an appropriate context to inform decision-making at many levels. To provide context to national survey outcomes, we describe patterns in the ratings of science and engineering subjects from the UK’s National Student Survey (NSS). New, robust statistical models describe relationships between the Overall Satisfaction’ rating and the preceding 21 core survey questions. Subjects exhibited consistent differences and ratings of “Teaching”, “Organisation” and “Support” were thematic predictors of “Overall Satisfaction” and the best single predictor was “The course was well designed and running smoothly”. General levels of satisfaction with feedback were low, but questions about feedback were ultimately the weakest predictors of “Overall Satisfaction”. The UK’s universities affiliated groupings revealed that more traditional “1994” and “Russell” groups over-performed in a model using the core 21 survey questions to predict “Overall Satisfaction”, in contrast to the under-performing newer universities in the Million+ and Alliance groups. Findings contribute to the debate about “level playing fields” for the interpretation of survey outcomes worldwide in terms of differences between subjects, institutional types and the questionnaire items.
Keywords: random forest analysis; data mining; student satisfaction; student surveys random forest analysis; data mining; student satisfaction; student surveys
This is an open access article distributed under the Creative Commons Attribution License which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Export to BibTeX |
EndNote


MDPI and ACS Style

Langan, A.M.; Dunleavy, P.; Fielding, A. Applying Models to National Surveys of Undergraduate Science Students: What Affects Ratings of Satisfaction? Educ. Sci. 2013, 3, 193-207.

AMA Style

Langan AM, Dunleavy P, Fielding A. Applying Models to National Surveys of Undergraduate Science Students: What Affects Ratings of Satisfaction? Education Sciences. 2013; 3(2):193-207.

Chicago/Turabian Style

Langan, Anthony M.; Dunleavy, Peter; Fielding, Alan. 2013. "Applying Models to National Surveys of Undergraduate Science Students: What Affects Ratings of Satisfaction?" Educ. Sci. 3, no. 2: 193-207.


Educ. Sci. EISSN 2227-7102 Published by MDPI AG, Basel, Switzerland RSS E-Mail Table of Contents Alert