Next Article in Journal
Cognitive Mechanisms Underlying the Influence of Facial Information Processing on Estimation Performance
Next Article in Special Issue
A Machine-Learning-Based Approach to Informing Student Admission Decisions
Previous Article in Journal
Pandemic-Induced PR Dilemmas Faced by Airlines: A Thematic Analysis of Spirit Airlines’ Incident Response from USA
Previous Article in Special Issue
Development of a Forced-Choice Personality Inventory via Thurstonian Item Response Theory (TIRT)
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Estimating the Minimum Sample Size for Neural Network Model Fitting—A Monte Carlo Simulation Study

by
Yongtian Cheng
1,*,
Konstantinos Vassilis Petrides
1 and
Johnson Li
2
1
Division of Psychology and Language Sciences, University College London (UCL), 26 Bedford Way, London WC1H 0AP, UK
2
Department of Psychology, University of Manitoba, Winnipeg, MB R3T 2N2, Canada
*
Author to whom correspondence should be addressed.
Behav. Sci. 2025, 15(2), 211; https://doi.org/10.3390/bs15020211
Submission received: 28 November 2024 / Revised: 7 February 2025 / Accepted: 11 February 2025 / Published: 14 February 2025

Abstract

:
In the era of machine learning, many psychological studies use machine learning methods. Specifically, neural networks, a set of machine learning methods that exhibit exceptional performance in various tasks, have been used on psychometric datasets for supervised model fitting. From the computer scientist’s perspective, psychometric independent variables are typically ordinal and low-dimensional—characteristics that can significantly impact model performance. To our knowledge, there is no guidance about the sample planning suggestion for this task. Therefore, we conducted a simulation study to test the performance of an NN with different sample sizes and the simulation of both linear and nonlinear relationships. We proposed the minimum sample size for the neural network model fitting with two criteria: the performance of 95% of the models is close to the theoretical maximum, and 80% of the models can outperform the linear model. The findings of this simulation study show that the performance of neural networks can be unstable with ordinal variables as independent variables, and we suggested that neural networks should not be used on ordinal independent variables with at least common nonlinear relationships in psychology. Further suggestions and research directions are also provided.

1. Introduction

Neural networks (NNs), which are a collection of machine learning algorithms, draw inspiration from the structure and functions of the human brain.
Neural networks with adequate width and depth—enabled by their hidden layers—can approximate any complex relationship between independent variables (IVs) and the dependent variable (DV) (Cybenko, 1989). This universal approximation property underpins their exceptional performance in tasks such as natural language processing (Zalake & Naik, 2019) and image classification (Rawat & Wang, 2017), in which there are complex features between the IVs and DV. In addition, NNs have been applied in diverse fields such as cancer prediction (Daoud & Mayo, 2019), transportation (Xiong & Schneider, 1992), engineering (Matel et al., 2022), and psychology (Choi et al., 2020).
In the field of psychology, supervised NNs are often utilized to discern patterns in psychological datasets (Witten et al., 2005). These NNs have not only pioneered a new direction in fully leveraging high-dimensional data but have also enhanced the performance of prediction tasks involving low-dimensional data, such as ordinal variables. Various studies by computer scientists have provided empirical evidence about the sample size requirements for high-dimensional data (Cho et al., 2015; Haykin, 2009; Kavzoglu & Mather, 2003). However, there is a notable gap in the literature regarding empirical evidence on sample size planning for fitting NN models with low-dimensional psychometric data, particularly from a prediction stability perspective in psychology, despite its importance in study design (Maxwell et al., 2008). Therefore, this study employs a Monte Carlo simulation to provide empirical evidence on the predictive performance of supervised NNs across various sample sizes and datasets, serving as a preliminary step in sample size planning. In the following paragraphs, the term ’NNs’ will refer exclusively to supervised neural networks, with a primary focus on prediction.
This study is structured into four sections. The first section offers a general introduction to the application of NNs in psychology, emphasizing existing sample size planning suggestions from previous studies. Following this, we introduce the design of our simulation study, providing justifications and detailed descriptions for each step. The subsequent section presents the simulation’s results, along with interpretations. Finally, this paper concludes with a general discussion, highlighting this study’s contributions and limitations and suggesting directions for further research.

1.1. Neural Network Application in Psychological Studies

NNs in psychology are typically applied to two categories of data for prediction purposes: high-dimensional data and low-dimensional data.
First, the ability of NNs to analyze high-dimensional data, such as natural language and video, has opened new possibilities for psychological research. For instance, Youyou et al. (2015) employed a natural generic digital footprint (i.e., Facebook likes), while L. Liu et al. (2016) used similar methods to analyze social media profile pictures. In another study, Dufour et al. (2020) applied an NN to assess vocal stereotypes in individuals with autism. The images in L. Liu et al. (2016) and vocal patterns in Dufour et al. (2020) were transformed into high-dimensional datasets using natural language processing (Liddy, 2001) or convolutional methods (Romanyuk, 2016). These advances in NNs and automated coding techniques are gradually replacing subjective human coding, enabling the creation of valuable high-dimensional datasets that facilitate prediction tasks in psychology.
In addition, many psychological studies utilize ordinal variables from psychometric scales (e.g., results from a 5-point Likert scale with values of 1, 2, 3, 4, or 5) as independent variables in NN models. These ordinal variables are ordered categorically and typically originate from psychometric scales with inherent measurement errors and self-correlation (Bland & Altman, 1997). Compared to datasets in various disciplines used by computer scientists, psychometric data generally exhibit lower dimensionality and higher measurement error (Jacobucci & Grimm, 2020). NNs have been leveraged in these contexts to outperform traditional regression methods (Yarkoni & Westfall, 2017). Suggested by Zeinalizadeh et al. (2015), NN models hold promise to learn and capture the behavior of highly nonlinear systems with proper accuracy and low computational efforts. These advantages cannot be achieved by common linearly structured models due to system nonlinearities and complexities.
For example, Marshall and English (2000) used various six-point Likert variables (from 0, indicating no risk, to 5, indicating high risk) to assess risk in child protective services, with 12,978 lines of data. They found that the NN model achieved 81% accuracy, outperforming the logistic regression model, which had 66% accuracy. Florio et al. (2009) employed a three-point (i.e., 0, 1, and 2) developmental behavior checklist (Einfeld & Tonge, 1995) for predicting Autism Spectrum Disorder and found that the NN performed better (ROC = 0.93) than logistic regression (ROC = 0.88) with a balanced sample size of 638. Zeinalizadeh et al. (2015) predicted bank customer satisfaction using a 51-item, five-point scale from 436 randomly selected customers. They found that the NN had a lower mean square error (MSE) than linear models (0.44 vs. 0.6).
However, the superior performance of neural network models relies heavily on having an adequate sample size. Rajput et al. (2023) and Pecher et al. (2024) demonstrated that insufficient sample sizes introduce randomness, which can undermine model stability. Similarly, Rajput et al. (2023), Kavzoglu and Mather (2003) and Haykin (2009) have shown that inadequate sample sizes during the model fitting process lead to unstable performance. Consequently, ensuring a sufficient sample size is a critical consideration for researchers employing neural networks in their studies. This issue will be elaborated on in the next section.

1.2. Previous Studies in Neural Network Sample Size Planning

While numerous psychological studies already employ ordinal variables as IVs in NN model fitting, there is limited empirical evidence available to help psychologists in designing a study that aims to use NNs for model fitting. Consider a scenario in which a group of psychologists wants to create a model that uses results from psychometric scales to make predictions. They believe there are some nonlinear relationships between the IVs and DV. Yet, they do not know what exactly those relationships are. Therefore, they want to use an NN to fit this model. Then, one of the most crucial questions they need to address is as follows: how many participants do they need to recruit?
Sample size planning is a crucial aspect of psychological research design. It ensures sufficient power, controls the budget, and addresses other concerns in psychological research (Maxwell et al., 2008). Various studies in the psychology discipline have provided recommendations for sample size planning for psychometric data (Kühberger et al., 2014), focusing on the validity of different statistics, such as correlation (Schönbrodt & Perugini, 2013), mediation (Fritz & MacKinnon, 2007), and CFA indexes (Marsh et al., 1988). The empirical evidence offered by these studies has supplied important guidelines for applied researchers. Additionally, given that measurement error is common in psychometric scales (Schmidt & Hunter, 1996), researchers have also developed numerous sample size planning techniques that take measurement error into account (Freedman et al., 1990; Levin & Subkoviak, 1977). For example, Bonett and Wright (2015) provided suggestions on the minimum sample size requirement for null hypothesis significance testing with the criterion of a desirable level of power based on Cronbach’s alpha level. In the following paragraph, we will review the sample size planning recommendations made by computer scientists for NNs.
Numerous studies conducted by computer scientists have addressed the sample size needed to fit various types of NN models (Alwosheel et al., 2018). Some rule-of-thumb guidelines have been developed for high-dimensional data, but these suggestions are inconsistent and stem from different perspectives. For instance, Cho et al. (2015) proposed that the sample size needs to be at least 50 to 1000 times the number of DVs based on a criterion of 99.5% multinomial accuracy. However, Kavzoglu and Mather (2003) suggested that the sample size should be at least 10 to 100 times the number of IVs based on the same criterion. In the meantime, these suggestions are based on findings from X-ray images or high-resolution visible images provided by NASA, which makes it doubtful whether they can applied to psychometric IVs.
Although a larger sample size can provide researchers with more confidence about the NN performance estimation reported in the study (Yarkoni & Westfall, 2017), there is no uniform sample size recommendation for fitting an NN (Rosenbusch et al., 2021).
These suggestions stem from different dimensions, which can result in an order of magnitude difference for a single study design. For instance, consider a scenario with three categorical IVs, each with five categories, and the researcher wants to fit a DV using an NN with three fully connected layers, each with 10 neurons. According to Cho et al. (2015), the sample size should be between 750 and 15,000; according to Kavzoglu and Mather (2003), it should be between 30 and 300; and according to Haykin (2009), a sample size of 600 is needed. These suggestions are based on findings from X-ray images or high-resolution visible images provided by NASA. While these recommendations may serve as references for psychological studies analyzing high-dimensional data like natural languages (Zalake & Naik, 2019) or images (Rawat & Wang, 2017), studies with such high-dimensional IVs can easily achieve large sample sizes. For example, Youyou et al. (2015) used participants’ Facebook likes to predict their personalities. With Facebook’s permission, they collected Facebook information from 86,220 participants.
Computer scientists have also conducted studies on how measurement error (i.e., noise) in the independent variables influences the sample size requirement. However, these studies mostly focus on voice and image independent variables (B. Liu et al., 2017; Tripathi, 2021). Some computer scientists also believe that NNs are robust to measurement error in independent variables (Zhang et al., 2018). The only study examining the influence of measurement error on NN models in psychological application research is work conducted by Jacobucci and Grimm (2020) from a performance perspective. In their study, they were concerned about whether the level of measurement noise (i.e., reliability of 0.3, 0.6, and 0.9) can influence the performance of the supervised machine learning model. A Monte Carlo simulation study by Jacobucci and Grimm (2020) found that the performance (i.e., R 2 ) of the boosting method (Freund, 1998) cannot always outperform the linear model even if there are interaction relationships between the IVs and DV. Meanwhile, they also found that the nonlinear feature is difficult to learn for boosting models in cases with a high measurement error level in the population.
In summary, although there are some studies on the sample size requirement for NN prediction tasks, there is no applicable guideline for the minimum sample size requirement when utilizing psychometric ordinal variables as IVs to make predictions about DVs of interest to studies. Additionally, there is a growing trend for psychologists to adopt a prediction perspective when reinterpreting psychological phenomena (Dwyer et al., 2018). As a result, it is likely that more studies will be conducted using a supervised NN model that fits ordinal IVs. Consequently, this study will use a Monte Carlo simulation to test the sample size requirement for NN model fitting with Likert IVs. Two criteria will be used to determine the sample size level: ensuring the performance of the NN model is stable in replication and ensuring that the NN can outperform the linear model. Several factors are included in this study: the number of IVs, the relationship between IVs and DVs, and the coefficients in the model. The performance of the model will be evaluated based on these two criteria separately, and two minimum sample size requirements (MSSRs) that meet these criteria will be reported in the Results Section.

2. Design

In this section, we will discuss the simulation design of the IVs and DV at first. Then, we will discuss the procedure of grid search, which is used to select the combination of hyperparamerters of the NN. Finally, we will provide two criteria for a sufficient sample size from different perspectives, with a pilot simulation to examine whether the criteria we propose are applicable.

2.1. Dataset Simulation Design

At first, we will discuss the simulation of the IVs and DV. In this study, we only include Likert IVs and a continuous DV.
We use a similar design to Maxwell (2000). IVs are simulated from an N(3,1) normal distribution and denoted as x 1 , x 2 , in the following paragraph. After that, x 1 , x 2 , are fixed by rounding up and the minimum is set at 1 and the maximum at 5; they will be denoted as X 1 , X 2 , in the following paragraph. The lowercase of x 1 , x 2 , are the true values behind the Likert scores, and the uppercase X 1 , X 2 , are the Likert scores observed by the researchers. The error caused by the Likertized procedure (i.e., rounding up in this case) is the only error included in the IVs in this study.
Three different numbers of IV conditions are included in this study, 3, 5, and 10, to calculate the simulated DV. In all conditions, the IVs and DV have both linear and nonlinear relationships in the simulated dataset. Suppose there is only a linear relationship between the IVs and DV. The NN cannot outperform linear regression in this case. Based on the Occam’s razor principle, linear regression, as a simpler model widely used in psychology, should be chosen unless the NN model demonstrates better performance. Therefore, there is no need to apply NNs when there is only a linear relationship in the population (Yarkoni & Westfall, 2017). As a result, this study only includes the IV and DV condition that there is both a linear and nonlinear relationship between them.
Nonlinear relationships are common in psychological studies (Richardson et al., 2017). This study includes several kinds of nonlinear relationships: two-way interaction effects ( x a x b ) (Mathieu et al., 2012), three-way interaction effects ( x a x b x c ) (Dawson & Richter, 2006; Wei et al., 2007), and quadratic effects ( x 2 ) (Guastello, 2001).
These nonlinear relationships are chosen as representative of the nonlinear relationships included in this simulation study. With these three nonlinear relationships, we aim at simulating datasets that have a complex relationship between the IVs and DV. Given their robust ability to capture these nonlinear patterns, NNs are ideally suited for this task (Almeida, 2002). We expect that the NN can detect these nonlinear relationships in the model fitting easily. Specifically, we include three levels of complexity of the nonlinear relationship: simple, medium, and high.
In the simple complexity level of the nonlinear relationship, only two-way interaction effects and linear relationships are included. In the medium complexity level of the nonlinear relationship, interaction effects, quadratic effects ( x 2 ), and linear relationships are included. In the high complexity level of the nonlinear relationship, interaction effects ( x a x b ), quadratic effects ( x 2 ), three-way interaction relationships ( x a x b x c ), and linear relationships are included.
For three complexities with three IV numbers, there are nine conditions between the IV and DV. The simulation method uses the study conducted by Li (2018) as a reference in the design.
The simulation formula is as follows.
Simple complexity with three IVs:
Y = c o f 1 × x 1 + c o f 2 × x 2 + c o f 3 × x 3 + c o n f 1 × x 1 × x 2 + c o n f 2 × x 2 × x 3 + c o f e r r × N ( 0 , 1 )
Medium complexity with three IVs:
Y = c o f 1 × x 1 + c o f 2 × x 2 + c o f 3 × x 3 + c o n f 1 × x 1 × x 2 + c o n f 2 × x 2 × x 3 + c o n f 3 × x 1 2 + c o n f 4 × x 2 2 + c o f e r r × N ( 0 , 1 )
High complexity with three IVs:
Y = c o f 1 × x 1 + c o f 2 × x 2 + c o f 3 × x 3 + c o n f 1 × x 1 × x 2 + c o n f 2 × x 2 × x 3 + c o n f 3 × x 1 2 + c o n f 4 × x 2 2 + c o n f 5 × e x p ( x 1 ) + c o f e r r × N ( 0 , 1 )
Simple complexity with five IVs:
Y = c o f 1 × x 1 + c o f 2 × x 2 + c o f 3 × x 3 + c o n f 1 × x 1 × x 2 + c o n f 2 × x 4 × x 5 + c o f e r r × N ( 0 , 1 )
Medium complexity with five IVs:
Y = c o f 1 × x 1 + c o f 2 × x 2 + c o f 3 × x 3 + c o n f 1 × x 1 × x 2 + c o n f 2 × x 4 × x 5 + c o n f 3 × x 1 2 + c o n f 4 × x 5 2 + c o f e r r × N ( 0 , 1 )
High complexity with five IVs:
Y = c o f 1 × x 1 + c o f 2 × x 2 + c o f 3 × x 3 + c o n f 1 × x 1 × x 2 + c o n f 1 × x 4 × x 5 + c o n f 2 × x 1 2 + c o n f 3 × x 5 2 + c o n f 4 × e x p ( x 4 ) + c o f e r r × N ( 0 , 1 )
Simple complexity with ten IVs:
Y = c o f 1 × x 1 + c o f 2 × x 2 + c o f 3 × x 3 + c o f 4 × x 4 + c o f 5 × x 5 +         c o f 6 × x 6 + c o n f 1 × x 7 × x 8 + c o n f 2 × x 9 × x 1 0 + c o f e r r × N ( 0 , 1 )
Medium complexity with ten IVs:
Y = c o f 1 × x 1 + c o f 2 × x 2 + c o f 3 × x 3 + c o f 4 × x 4 + c o f 5 × x 5 +         c o f 6 × x 6 + c o n f 1 × x 7 × x 8 + c o n f 2 × x 9 × x 1 0 + c o n f 3 × x 5 2 + c o n f 4 × x 1 0 2 + c o f e r r × N ( 0 , 1 )
High complexity with ten IVs:
Y = c o f 1 × x 1 + c o f 2 × x 2 + c o f 3 × x 3 + c o f 4 × x 4 + c o f 5 × x 5 + c o f 6 × x 6 +         c o n f 1 × x 7 × x 8 + c o n f 2 × x 9 × x 10 + c o n f 3 × x 5 2 + c o n f 4 × x 10 2 + c o n f 5 × x 9 + c o f e r r × N ( 0 , 1 )
Here, x 1 x 10 are the IVs, Ys are DVs, c o f x is the coefficient for linear relationships, c o n f x is the coefficient for nonlinear relationships, in which x = 1 , 2 , 3 , c o f e r r is the coefficient of the error, and N(0,1) is a random number simulated from a standard normal distribution.
Due to the time limitation in simulation, the orthogonal experimental design for item coefficients can not be applied because the time consumption associated with this design far exceeds the acceptable limits for the simulation, as this is a high-compute-intensive task, which will be explained in detail later.
Therefore, we include two conditions for these two coefficients separately: small or large. A coefficient with a small condition is simulated from a uniform distribution of [0.1, 0.3], and a coefficient with a large condition is simulated from a uniform distribution of [0.5, 1]. A gap between these two distributions is designed to explore the relationship between the MSSR and categorical coefficient conditions. All these coefficients are simulated from a selected random seed and are saved in the supplementary document. There are 2 × 2 = 4 conditions of coefficients in total, in which each coefficient is simulated independently by the conditions. This means a condition with small linear relationship coefficients and large nonlinear relationship coefficients can be calculated by
Y = 0.29 × x 1 + 0.28 × x 2 + 0.13 × x 3 + 0 . 6 1 × x 1 × x 2 + 0.7 × x 2 × x 3 + 0 . 8 1 × x 1 2 + 0 . 75 2 × x 2 2 + c o f e r r × N ( 0 , 1 )
Here, x 1 , x 2 , x 3 are the continuous IVs without Likertization and Y Is the simulated DV.
For all formulas in the Appendix A, c o f e r r in the formula is the noise coefficient and is a multiple with a number simulated by a normal distribution of N(0,1). There are three levels of noise coefficients: 1 , 4 , and 10. This design aims to add different levels of variance that can not be explained (i.e., noise).
To sum up, there are nine different kinds of relationships between the IVs and DV, four different kinds of coefficients, and three different error levels in this study. As a result, there are 144 conditions in this simulation. After Y is calculated with the true value of IVs like x 1 , x 2 , , the DV Y will be combined with the corresponding IVs like X 1 , X 2 , to create the dataset for NN model fitting. All these coefficients are simulated from a selected random seed and are saved in a supplementary document.
Based on the design proposed above, we can calculate several statistics for the simulation datasets. These statistics will be used in the next section to interpret the results.
The theoretical maximum explainable variance R t 2 is calculated by a linear regression to Y with all the items. For the example in (10), because X i is the best estimation of x i available in the dataset, R t 2 is calculated by the regression models with factors X 1 , X 2 , X 3 , X 1 × X 2 , X 2 × X 3 , and X 2 2 in this sample. As all factors used to calculate Y are included in the model, R t 2 should be the theoretical maximum explainable variance.
In addition, we also calculate the prediction performance calculated by linear regression with all items included in the model, and we denote the variance that can be explained as R l 2 , in which l stands for linear. In the sample, the R l 2 is calculated with a linear regression model that includes factors X 1 , X 2 , and X 3 , which are all the items included in this study. The performance of an ideal NN model should be expected to outperform R l 2 .

2.2. Neural Network Design

Researchers have a high degree of freedom in designing NN model fitting (Donda et al., 2022), particularly when it comes to selecting a combination of hyperparameters for optimal performance. Hyperparameters are parameters that significantly influence the model fitting process. However, unlike regular parameters, hyperparameters cannot be determined during model fitting; they must be decided beforehand (Dwyer et al., 2018).
In practice, researchers can choose hyperparameters through grid search (Erdogan et al., 2021) combined with cross-validation (Donda et al., 2022). Grid search is a technique used in machine learning for hyperparameter tuning, where all possible combinations of predefined hyperparameter values are evaluated to find the optimal settings for a model (Erdogan et al., 2021). For each condition, the best combination of hyperparameters is applied to the NN using this grid search. Below, we will provide a general introduction to the hyperparameters included in the grid search, with a focus on how they can influence the performance of the NN.

2.2.1. Neural Network Shape

The shape of the NN delineates its architecture in terms of the number of layers and the number of neurons (or nodes) in each layer. It is a fundamental hyperparameter that dictates the complexity and capacity of the model. An appropriate network shape is crucial: too shallow or with few neurons might lead to underfitting, while too deep or with many neurons can lead to overfitting and increased computational demands (Yu & Zhu, 2020). In the meantime, the selection of shape influences the MSSR based on the research of (Alwosheel et al., 2018). A larger sample size is needed if more neurons and layers are in the NN model for a stable result (Haykin, 2009).
In this study, we consider only the simplest NN: a fully connected unidirectional NN, which is the model used in Florio et al. (2009); Marshall and English (2000); Zeinalizadeh et al. (2015). Therefore, we include the conditions of (10), (10, 10), and (10, 10, 10) for the NN. For instance, (10, 10) means there are two fully connected hidden layers with 10 neurons in each layer for the NN.
Based on our simulated dataset, a fully connected unidirectional neural network is sufficient for this prediction task. Yet, it should be mentioned that for more complex applications, advanced architectures are preferable: Convolutional Neural Networks (CNNs) excel with image and spatial data (Chua, 1997), while Recurrent Neural Networks (RNNs) are ideal for handling sequential data (Khaldi et al., 2023) and language and context modeling (Mikolov et al., 2011).

2.2.2. Learning Rate

The learning rate is a crucial hyperparameter in neural networks that determines the step size during the optimization process. A suitable value ensures efficient convergence during training, whereas values that are too large or too small can hinder model performance by causing overshooting or slow convergence, respectively (Yu & Zhu, 2020). We include three levels of learning rate in this search, 0,0001, 0.001, and 0.01 (Ranganath et al., 2013; Wu et al., 2019), using the Adam gradient descent method (Kingma & Ba, 2014).

2.2.3. Patience

Patience determines the number of epochs the training process should wait without observing improvement in a chosen metric before halting the training (Terry et al., 2021). This prevents overfitting and can potentially reduce training time. We include three levels of patience in this search: 5, 10, and 15 (Chen et al., 2022; Franchini et al., 2022; Ho et al., 2021).

2.2.4. Batch Size

Batch size refers to the number of training examples utilized in one iteration. It plays a pivotal role in optimizing the training process, influencing the model’s generalization ability, training speed, and convergence (Yu & Zhu, 2020). While smaller batches can provide a regularizing effect and lower generalization error, larger batches can accelerate the learning process by leveraging computational efficiencies. We include three levels of batch size in this search: 32, 64, and 128 (Kandel & Castelli, 2020; Peng et al., 2018; Yong et al., 2020; You et al., 2017).
An orthogonal design was used in this search, which means there were 3 × 3 × 3 × 3 = 81 combinations of hyperparameters in the search. With cross-validation, the combination of hyperparameters with the best performance on the validation dataset was selected. As a result, this is a computationally intensive task. For each combination of hyperparameters, ten NN models will be fitted, and each replication in a single simulation involves 81 hyperparameter combinations.

2.3. Criteria for Adequate Sample Size in Neural Network Model Fitting

In this study, we use two criteria for determining the MSSR: we want the majority of the predictive performances of the NN in 1000 replications to be close to the maximum performance, and we want the majority of the predictive performances of the NN in 1000 replications to outperform linear regression.

2.3.1. The Criterion Based on the Theoretical Maximum Performance

There are existing Monte Carlo simulation studies aiming to estimate the influence of sample size on the performance of linear regression in psychological studies. Linear regression is a commonly used prediction method for continuous IVs in psychology. As supervised NNs are also used for prediction, the criteria of these existing studies can be used as a reference.
For example, Riley et al. (2019) proposed four criteria for a stable performance of linear regressions: (i) small optimism in predictor effect estimates as defined by a global shrinkage factor of larger than 0.9; (ii) small absolute difference of less than 0.05 in the apparent and adjusted R 2 ; (iii) precise estimation (a margin of error less than 10% of the true value) of the model’s residual standard deviation; and, similarly, (iv) precise estimation of the mean predicted outcome value (model intercept). Except for the second criterion, none of the other criteria can be applied to the performance of the NN.
Therefore, we propose a criterion for this study based on the design discussed in Riley et al. (2019), with some modifications for Monte Carlo simulation. We expanded the absolute value and doubled the tolerance range from 0.05 to 0.1. For an adequate sample size, we expect that 95% of NNs in the simulation will have a performance that will meet the criterion of
R 97.5 % 2 R 2.5 % 2 < 0.1
Here, R 97.5 % 2 is the 97.5% percentile of R 2 in the simulation; R 2.5 % 2 is the 2.5% percentile of R 2 in the simulation.
With this design, we aim to find an MSSR for the NN that ensures that most of the NN models neither overperform nor underperform in relation to R t 2 . As noted earlier, R t 2 is computed using the model that simulated the DV. Consequently, R t 2 represents the theoretical upper bound of performance. In the following paragraphs, we will refer to this criterion as the criterion based on the theoretical maximum performance.

2.3.2. The Criterion Based on Outperforming of the Linear Model

We also propose another criterion: the NN model should outperform the linear regression model. We propose this metric for these two reasons:
1: Based on the Occam’s razor principle, if a complex model like an NN cannot outperform the simple linear model, the NN model should not be chosen, as the linear regression model offers a simpler interpretation. Psychological researchers also suggest that the performance of the supervised machine learning fitted model should be compared with the performance of the linear model in psychological research to make a binary decision on whether to use the supervised machine learning method in a study (Rosenbusch et al., 2021).
2: Given that there are both linear and nonlinear relationships between the IVs and DV in this simulation study, the NN should outperform the linear model.
Therefore, we aim to find an MSSR such that 80% of the NN models can outperform the linear model, which is calculated by including all the items linearly in a regression. The performance of the linear model is calculated with a sample size of 100,000. An adequate sample size should make 80% of R l 2 > R N N 2 . In the following paragraph, we will call this criterion the criterion based on outperforming the linear model.

2.4. General Simulation Design

The whole simulation study was conducted in Python (Pilgrim & Willison, 2009), with the packages Tensorflow (Abadi et al., 2016) and Keras (Chollet, 2023). We included sample sizes of 1000, 2500, 5000, 10,000, and 20,000. If either of the two criteria could not be met with a sample size of 20,000, we then tested additional sample sizes of 25,000, 30,000, 35,000, 40,000, 45,000, and 50,000, one after another. If a sample size of 50,000 was still insufficient to meet either of the two criteria, we considered the sample size requirement for this criterion in this condition as unattainable.
After the dataset was simulated, 80% of the simulated data were randomly assigned to the training dataset, and the rest were assigned to the testing dataset. In total, 20% of the training dataset was randomly assigned as the validation dataset for the hyperparameters selection by grid search. All the performance reported in the study is based on the performance of the model on the testing dataset.

3. Results

The full results for all sample sizes are provided in the supplementary document. A selection of these results is presented in Table 1. The MSSR based on the theoretical maximum performance is identified, and the MSSR based on outperforming the linear model is provided in Table 2.
Below, we will discuss the MSSR results provided by the two criteria. We will report Spearman’s correlation coefficients between the MSSR and the rank factors included in this simulation study. Both the correlation coefficients and the p-values will be provided. However, we emphasize that these p-values should be viewed as a reference only, given the low sample size in these tests (Cumming, 2014).

3.1. Criterion Based on the Theoretical Maximum Performance

Based on my findings, an MSSR can be determined for most of the conditions in this study, using the theoretical maximum performance as the criterion. The conditions where stable performance close to R t 2 in replication could not be found are mainly those with a low R t 2 (e.g., condition 25: R t 2 = 0.0273 ).
For the conditions where an MSSR has been found using this criterion (N = 96), the MSSR is positively correlated with error level (r = 0.5392, p < 0.001) and number of IVs (r = 0.1035, p = 0.3181). The MSSR is negatively correlated with linear coefficients (r = −0.066, p = 0.5255), complexity level (r = −0.136, p = 0.1886), and nonlinear coefficients (r = −0.2611, p = 0.0106, and R t 2 (r = −0.832, p < 0.001). Moreover, if a researcher aims to restrict the R t 2 to within a margin of ± 0.05 % using a sample size of 1000, the neural network would need to explain 80% of the variance—a level of explanatory power that is rarely achieved in psychological research.

3.2. Criterion Based on the Outperforming of the Linear Model

We have found unexpected results for the MSSR with this criterion. In all the conditions simulated in this study, there are both linear and nonlinear relationships between the IVs and DV. We expected the performance of the NN to be better than the linear model, as the linear model can only fit the linear relationship between the IVs and DV. However, we found that in only 89 out of 108 conditions, we could find an MSSR of less than 50,000 to ensure that 80% of the results provided by the NN were better than those provided by linear regression. While previous research has shown that high measurement error levels in the independent variable can linearize nonlinear relationships (Jacobucci & Grimm, 2020), we did not anticipate that the measurement error inherent in the Likert process would be sufficient to cause neural network performance to mirror that of a linear model. Furthermore, this measurement error may affect hyperparameter selection, thereby further impacting the overall performance of the model (Tsamardinos et al., 2015).
The conditions where an MSSR could not be identified using this criterion all had an R t 2 < 0.1 . These conditions also tended to have lower complexity and smaller nonlinear coefficients.
For the conditions where an MSSR was found using this criterion (N = 90), MSSR was positively correlated with error level (r = 0.37, p < 0.001), number of IVs (r = 0.14, p = 0.5299), and linear coefficients (r = 0.11, p = 0.1956). MSSR was negatively correlated with complexity level (r = −0.32, p = 0.002), nonlinear coefficients (r = −0.35, p < 0,0001), and r t (r = −0.72, p < 0,0001).
In the meantime, the variance can be explained by the model (i.e., R 2 ), which should be inversely proportional to the width of the interval, based on both theoretical proof and empirical evidence (Eng, 2003; Hazra, 2017). However, a universal constant based on this relationship cannot be found across all conditions. Therefore, we cannot provide a rule-of-thumb formula for researchers to use for NN sample size planning based on the simulation results of this study.

4. Discussion, Limitations, and Further Directions

This study was expected to offer guidance regarding the necessary sample size for fitting NN models with psychometric data. However, we found that meeting both criteria is challenging unless the sample size is quite large: a stable result close to the maximum theoretical performance cannot be achieved even with a sample size of 10,000 in some conditions when using the NN. In other conditions, even a stable performance cannot be reached with a sample size of 50,000. This indicates that the NN requires a very large sample size to learn some common nonlinear relationships in psychology. Therefore, based on the results of this study, we do not recommend using an NN for at least some prediction tasks with ordinal IVs.
Specifically, NNs should not be applied to datasets with a high level of inexplicable noise. In these datasets, a sample size of 50,000 is insufficient to meet the criteria proposed in this study. We advise against fitting NN models to datasets where only low performance is achievable. This recommendation is not because an NN model cannot outperform classical regression in these conditions; in fact, good performance can sometimes be observed due to sampling error. However, this instability is precisely why we advise against this application. While an NN model may dramatically outperform the linear regression model in the testing dataset due to sampling error, we cannot confidently expect a replication study with the same sample size (i.e., 1000) to yield a similar performance advantage.
By making this statement, we are critiquing the reproducibility of NNs with psychometric data unless the sample size is large enough. While psychologists consider performance estimation from independent testing datasets with cross-validation as more reliable and view it as a potential solution to the replication crisis (Koul et al., 2018; Rooij & Weeda, 2020), this study found that the results provided by the NN may not be replicable even when the training/testing dataset division is applied, unless an adequate sample size is available. However, if a necessary sample size can be recruited for a study, researchers can identify nonlinear relationships using developed methods like Jaccard et al. (1990), and a regression model with nonlinear terms can be conducted, which eliminates the necessity of using an NN.
To the best of our knowledge, researchers may still want to apply NNs to predict factors of interest in studies with psychometric IVs if they believe that the relationships between IVs and DVs are complex, cannot be captured by an analytical formula, or are far from linear relationships, such as exponential ( e x ) relationships (Guastello, 2001). In the meantime, NNs may still be useful to provide stable performance in the case that the reliability of IVs is high (i.e., the measurement error of IVs is at a low level). Yet, more research should be conducted in this field. In addition, CNNs or RNNs can still be useful when analyzing high-dimensional data, such as images, as regression cannot deal with this kind of high-dimensional data.
Regarding limitations and future directions, the most significant limitation of this study is the restricted range of conditions examined. Future research should encompass a broader variety of conditions for NN models and dataset designs. It is acknowledged that simulating every possible condition of NN models in a single study is unfeasible. Nevertheless, future studies should explore additional factors, such as different types of relationships between IVs and DVs and varying sample size thresholds. Moreover, future research should also consider scenarios involving binary or multinomial DVs. Unlike multinomial logistic regression, NN algorithms do not necessitate the assumption of a linear relationship between the IVs and the logit transformation of the DV. This characteristic could be a potential advantage of NN models, offering a more flexible approach to handling various data types and relationship dynamics.
However, researchers should be aware that this study utilized GPU boost with multi-threading in simulation, employing 16 threads on a high-performance server, and still required five days to complete the simulation. Therefore, similar tasks can be time-consuming, as we have mentioned above.
The discovery that NNs cannot provide stable results has sparked new considerations regarding the choice of supervised machine learning methods, particularly for low-dimensional ordinal IVs. While NNs are sophisticated and hold great potential for predicting outcomes with high-dimensional IVs, psychologists might also explore other advanced methods rooted in regression. Specifically, penalized linear regression algorithms, which balance bias and variance for improved performance, are noteworthy alternatives. Lasso (Tibshirani, 1996), Ridge (Hoerl & Kennard, 1970), and Elastic Net (Zou & Hastie, 2005) are three penalized linear regression models known to often surpass traditional linear regression in performance. While these models have fewer hyperparameters and coefficients, they might need a smaller sample size for a stable performance near the true value.
According to the simulation results of this study, in scenarios with measurement errors and high noise levels—even in the presence of nonlinear relationships—these advanced linear regression methods can yield robust performance and are highly recommended. Another advantage of these methods is their higher interpretability compared to NN models, a factor contributing to their popularity among psychologists (Bainter et al., 2023; Doornenbal, 2021; Yoo, 2018). For quantitative psychologists, an additional research avenue could involve gathering empirical evidence to assist applied psychologists in planning sample sizes for penalized regression methods. This approach could significantly enhance the accuracy and efficacy of research in psychology.

Author Contributions

Y.C. performed the simulation and wrote the main manuscript text. K.V.P. and J.L. provided suggestions about methodology in the simulation and revised the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This study has not received any funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original contributions presented in this study are included in the link of in https://osf.io/ptzkj/, accessed on 6 February 2025. Further inquiries can be directed to the corresponding author(s).

Conflicts of Interest

The authors have declared that no competing interests exist.

Appendix A. Formula of Simulations

Table A1. Glossary of variables.
Table A1. Glossary of variables.
Variable NamesMeaning of the Variables
x i The simulated continuous IVs, which serves as the true value of IVs,
 in which i stands for 1 , 2 , 3 ,
X i The Likertized IVs from x_i, which serves as the observed value of IVs,
 in which i stands for 1 , 2 , 3 ,
R t 2 theoretical maximum explainable variance
R n 2 explainable variance by NN model
Note: the glossary of variables.

References

  1. Abadi, M., Barham, P., Chen, J., Chen, Z., Davis, A., Dean, J., Devin, M., Ghemawat, S., Irving, G., Isard, M., Kudlur, M., Levenberg, J., Monga, R., Moore, S., Murray, D. G., Steiner, B., Tucker, P., Vasudevan, V., Warden, P., . . . Zheng, X. (2016). Tensorflow: A system for large-scale machine learning. In 12th usenix symposium on operating systems design and implementation (osdi 16) (pp. 265–283). USENIX Association. [Google Scholar]
  2. Almeida, J. S. (2002). Predictive non-linear modeling of complex data by artificial neural networks. Current Opinion in Biotechnology, 13(1), 72–76. [Google Scholar] [CrossRef]
  3. Alwosheel, A., van Cranenburgh, S., & Chorus, C. G. (2018). Is your dataset big enough? Sample size requirements when using artificial neural networks for discrete choice analysis. Journal of Choice Modelling, 28, 167–182. [Google Scholar] [CrossRef]
  4. Bainter, S. A., McCauley, T. G., Fahmy, M. M., Goodman, Z. T., Kupis, L. B., & Rao, J. S. (2023). Comparing bayesian variable selection to lasso approaches for applications in psychology. Psychometrika, 88(3), 1032–1055. [Google Scholar] [CrossRef]
  5. Bland, J. M., & Altman, D. G. (1997). Statistics notes: Cronbach’s alpha. BMJ, 314(7080), 572. [Google Scholar] [CrossRef] [PubMed]
  6. Bonett, D. G., & Wright, T. A. (2015). Cronbach’s alpha reliability: Interval estimation, hypothesis testing, and sample size planning. Journal of Organizational Behavior, 36(1), 3–15. [Google Scholar] [CrossRef]
  7. Chen, J., Wolfe, C., Li, Z., & Kyrillidis, A. (2022, May 23–27). Demon: Improved neural network training with momentum decay. Icassp 2022–2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) (pp. 3958–3962), Singapore. [Google Scholar]
  8. Cho, J., Lee, K., Shin, E., Choy, G., & Do, S. (2015). How much data is needed to train a medical image deep learning system to achieve necessary high accuracy? arXiv, arXiv:1511.06348. [Google Scholar]
  9. Choi, R. Y., Coyner, A. S., Kalpathy-Cramer, J., Chiang, M. F., & Campbell, J. P. (2020). Introduction to machine learning, neural networks, and deep learning. Translational Vision Science & Technology, 9(2), 14. [Google Scholar]
  10. Chollet, F. (2023). Keras: The python deep learning library. Available online: https://keras.io/ (accessed on 6 February 2025).
  11. Chua, L. O. (1997). CNN: A vision of complexity. International Journal of Bifurcation and Chaos, 7(10), 2219–2425. [Google Scholar] [CrossRef]
  12. Cumming, G. (2014). The new statistics: Why and how. Psychological Science, 25(1), 7–29. [Google Scholar] [CrossRef] [PubMed]
  13. Cybenko, G. (1989). Approximation by superpositions of a sigmoidal function. Mathematics of Control, Signals and Systems, 2(4), 303–314. [Google Scholar] [CrossRef]
  14. Daoud, M., & Mayo, M. (2019). A survey of neural network-based cancer prediction models from microarray data. Artificial Intelligence in Medicine, 97, 204–214. [Google Scholar] [CrossRef] [PubMed]
  15. Dawson, J. F., & Richter, A. W. (2006). Probing three-way interactions in moderated multiple regression: Development and application of a slope difference test. Journal of Applied Psychology, 91(4), 917. [Google Scholar] [CrossRef] [PubMed]
  16. Donda, K., Zhu, Y., Merkel, A., Wan, S., & Assouar, B. (2022). Deep learning approach for designing acoustic absorbing metasurfaces with high degrees of freedom. Extreme Mechanics Letters, 56, 101879. [Google Scholar] [CrossRef]
  17. Doornenbal, B. (2021). Big Five personality as a predictor of health: Shortening the questionnaire through the elastic net. Current Issues in Personality Psychology, 9(2), 159–164. [Google Scholar] [CrossRef] [PubMed]
  18. Dufour, M.-M., Lanovaz, M. J., & Cardinal, P. (2020). Artificial intelligence for the measurement of vocal stereotypy. Journal of the Experimental Analysis of Behavior, 114(3), 368–380. [Google Scholar] [CrossRef] [PubMed]
  19. Dwyer, D. B., Falkai, P., & Koutsouleris, N. (2018). Machine learning approaches for clinical psychology and psychiatry. Annual Review of Clinical Psychology, 14, 91–118. [Google Scholar] [CrossRef]
  20. Einfeld, S. L., & Tonge, B. J. (1995). The developmental behavior checklist: The development and validation of an instrument to assess behavioral and emotional disturbance in children and adolescents with mental retardation. Journal of Autism and Developmental Disorders, 25(2), 81–104. [Google Scholar] [CrossRef]
  21. Eng, J. (2003). Sample size estimation: How many individuals should be studied? Radiology, 227(2), 309–313. [Google Scholar] [CrossRef] [PubMed]
  22. Erdogan, Erten, G., Bozkurt Keser, S., & Yavuz, M. (2021). Grid search optimised artificial neural network for open stope stability prediction. International Journal of Mining, Reclamation and Environment, 35(8), 600–617. [Google Scholar] [CrossRef]
  23. Florio, T., Einfeld, S., Tonge, B., & Brereton, A. (2009). Providing an independent second opinion for the diagnosis of autism using artificial intelligence over the internet. Couns, Psycho Health Use Technol Mental Health, 5, 232–248. [Google Scholar]
  24. Franchini, G., Verucchi, M., Catozzi, A., Porta, F., & Prato, M. (2022). Biomedical image classification via dynamically early stopped artificial neural network. Algorithms, 15(10), 386. [Google Scholar] [CrossRef]
  25. Freedman, L. S., Schatzkin, A., & Wax, Y. (1990). The impact of dietary measurement error on planning sample size required in a cohort study. American Journal of Epidemiology, 132(6), 1185–1195. [Google Scholar] [CrossRef] [PubMed]
  26. Freund, Y. (1998). An introduction to boosting based classification. In Proceedings of the AT&T conference on quantitative analysis (pp. 80–91). Citeseer. [Google Scholar]
  27. Fritz, M. S., & MacKinnon, D. P. (2007). Required sample size to detect the mediated effect. Psychological Science, 18(3), 233–239. [Google Scholar] [CrossRef] [PubMed]
  28. Guastello, S. J. (2001). Nonlinear dynamics in psychology. Discrete Dynamics in Nature and Society, 6(1), 11–29. [Google Scholar] [CrossRef]
  29. Haykin, S. (2009). Neural networks and learning machines, 3/e. Pearson Education India. [Google Scholar]
  30. Hazra, A. (2017). Using the confidence interval confidently. Journal of Thoracic Disease, 9(10), 4125. [Google Scholar] [CrossRef]
  31. Ho, Y.-W., Rawat, T. S., Yang, Z.-K., Pratik, S., Lai, G.-W., Tu, Y.-L., & Lin, A. (2021). Neuroevolution-based efficient field effect transistor compact device models. IEEE Access, 9, 159048–159058. [Google Scholar] [CrossRef]
  32. Hoerl, A. E., & Kennard, R. W. (1970). Ridge regression: Biased estimation for nonorthogonal problems. Technometrics, 12(1), 55–67. [Google Scholar] [CrossRef]
  33. Jaccard, J., Wan, C. K., & Turrisi, R. (1990). The detection and interpretation of interaction effects between continuous variables in multiple regression. Multivariate Behavioral Research, 25(4), 467–478. [Google Scholar] [CrossRef]
  34. Jacobucci, R., & Grimm, K. J. (2020). Machine learning and psychological research: The unexplored effect of measurement. Perspectives on Psychological Science, 15(3), 809–816. [Google Scholar] [CrossRef] [PubMed]
  35. Kandel, I., & Castelli, M. (2020). The effect of batch size on the generalizability of the convolutional neural networks on a histopathology dataset. ICT Express, 6(4), 312–315. [Google Scholar] [CrossRef]
  36. Kavzoglu, T., & Mather, P. M. (2003). The use of backpropagating artificial neural networks in land cover classification. International Journal of Remote Sensing, 24(23), 4907–4938. [Google Scholar] [CrossRef]
  37. Khaldi, R., El Afia, A., Chiheb, R., & Tabik, S. (2023). What is the best RNN-cell structure to forecast each time series behavior? Expert Systems with Applications, 215, 119140. [Google Scholar] [CrossRef]
  38. Kingma, D. P., & Ba, J. (2014). Adam: A method for stochastic optimization. arXiv, arXiv:1412.6980. [Google Scholar]
  39. Koul, A., Becchio, C., & Cavallo, A. (2018). Cross-validation approaches for replicability in psychology. Frontiers in Psychology, 9, 1117. [Google Scholar] [CrossRef] [PubMed]
  40. Kühberger, A., Fritz, A., & Scherndl, T. (2014). Publication bias in psychology: A diagnosis based on the correlation between effect size and sample size. PLoS ONE, 9(9), e105825. [Google Scholar] [CrossRef]
  41. Levin, J. R., & Subkoviak, M. J. (1977). Planning an experiment in the company of measurement error. Applied Psychological Measurement, 1(3), 331–338. [Google Scholar] [CrossRef]
  42. Li, J. C.-H. (2018). Curvilinear moderation—A more complete examination of moderation effects in Behavioral sciences. Frontiers in Applied Mathematics and Statistics, 4, 7. [Google Scholar] [CrossRef]
  43. Liddy, E. D. (2001). Natural language processing. Syracuse University. [Google Scholar]
  44. Liu, L., Preotiuc-Pietro, D., Samani, Z. R., Moghaddam, M. E., & Ungar, L. (2016). Analyzing personality through social media profile picture choice. In Proceedings of the tenth international AAAI conference on web and social media. AAAI Press. [Google Scholar]
  45. Liu, B., Wei, Y., Zhang, Y., & Yang, Q. (2017). Deep neural networks for high dimension, low sample size data. IJCAI, 2017, 2287–2293. [Google Scholar]
  46. Marsh, H. W., Balla, J. R., & McDonald, R. P. (1988). Goodness-of-fit indexes in confirmatory factor analysis: The effect of sample size. Psychological Bulletin, 103(3), 391. [Google Scholar] [CrossRef]
  47. Marshall, D. B., & English, D. J. (2000). Neural network modeling of risk assessment in child protective services. Psychological Methods, 5(1), 102. [Google Scholar] [CrossRef]
  48. Matel, E., Vahdatikhaki, F., Hosseinyalamdary, S., Evers, T., & Voordijk, H. (2022). An artificial neural network approach for cost estimation of engineering services. International Journal of Construction Management, 22(7), 1274–1287. [Google Scholar] [CrossRef]
  49. Mathieu, J. E., Aguinis, H., Culpepper, S. A., & Chen, G. (2012). Understanding and estimating the power to detect cross-level interaction effects in multilevel modeling. Journal of Applied Psychology, 97(5), 951. [Google Scholar] [CrossRef] [PubMed]
  50. Maxwell, S. E. (2000). Sample size and multiple regression analysis. Psychological Methods, 5(4), 434. [Google Scholar] [CrossRef] [PubMed]
  51. Maxwell, S. E., Kelley, K., & Rausch, J. R. (2008). Sample size planning for statistical power and accuracy in parameter estimation. Annual Review of Psychology, 59, 537–563. [Google Scholar] [CrossRef] [PubMed]
  52. Mikolov, T., Kombrink, S., Burget, L., Černockỳ, J., & Khudanpur, S. (2011, May 22–27). Extensions of recurrent neural network language model. 2011 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) (pp. 5528–5531), Prague, Czech Republic. [Google Scholar]
  53. Pecher, B., Srba, I., & Bielikova, M. (2024). A survey on stability of learning with limited labelled data and its sensitivity to the effects of randomness. ACM Computing Surveys, 57(1), 1–40. [Google Scholar]
  54. Peng, C., Xiao, T., Li, Z., Jiang, Y., Zhang, X., Jia, K., Yu, G., & Sun, J. (2018). Megdet: A large mini-batch object detector. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 6181–6189). IEEE. [Google Scholar]
  55. Pilgrim, M., & Willison, S. (2009). Dive into python 3 (Vol. 2). Springer. [Google Scholar]
  56. Rajput, D., Wang, W.-J., & Chen, C.-C. (2023). Evaluation of a decided sample size in machine learning applications. BMC bioinformatics, 24(1), 48. [Google Scholar] [CrossRef] [PubMed]
  57. Ranganath, R., Wang, C., David, B., & Xing, E. (2013). An adaptive learning rate for stochastic variational inference. In International conference on machine learning (pp. 298–306). PMLR. [Google Scholar]
  58. Rawat, W., & Wang, Z. (2017). Deep convolutional neural networks for image classification: A comprehensive review. Neural Computation, 29(9), 2352–2449. [Google Scholar] [CrossRef]
  59. Richardson, J., Paxton, A., & Kuznetsov, N. (2017). Nonlinear methods for understanding complex dynamical phenomena in psychological science. Psychological Science Agenda, 31, 1–9. [Google Scholar]
  60. Riley, R. D., Snell, K. I., Ensor, J., Burke, D. L., Harrell, F. E., Jr., Moons, K. G., & Collins, G. S. (2019). Minimum sample size for developing a multivariable prediction model: Part I–Continuous outcomes. Statistics in Medicine, 38(7), 1262–1275. [Google Scholar] [CrossRef] [PubMed]
  61. Romanyuk, V. V. (2016). Training data expansion and boosting of convolutional neural networks for reducing the MNIST dataset error rate. Research Bulletin of the National Technical University of Ukraine “Kyiv Politechnic Institute”, 6, 29–34. [Google Scholar] [CrossRef]
  62. Rooij, M., & Weeda, W. (2020). Cross-validation: A method every psychologist should know. Advances in Methods and Practices in Psychological Science, 3(2), 248–263. [Google Scholar] [CrossRef]
  63. Rosenbusch, H., Soldner, F., Evans, A. M., & Zeelenberg, M. (2021). Supervised machine learning methods in psychology: A practical introduction with annotated R code. Social and Personality Psychology Compass, 15(2), e12579. [Google Scholar] [CrossRef]
  64. Schmidt, F. L., & Hunter, J. E. (1996). Measurement error in psychological research: Lessons from 26 research scenarios. Psychological Methods, 1(2), 199. [Google Scholar] [CrossRef]
  65. Schönbrodt, F. D., & Perugini, M. (2013). At what sample size do correlations stabilize? Journal of Research in Personality, 47(5), 609–612. [Google Scholar] [CrossRef]
  66. Terry, J. K., Jayakumar, M., & De Alwis, K. (2021). Statistically significant stopping of neural network training. arXiv, arXiv:2103.01205. [Google Scholar]
  67. Tibshirani, R. (1996). Regression shrinkage and selection via the lasso. Journal of the Royal Statistical Society Series B: Statistical Methodology, 58(1), 267–288. [Google Scholar] [CrossRef]
  68. Tripathi, M. (2021). Facial image noise classification and denoising using neural network. Sustainable Engineering and Innovation, 3(2), 102–111. [Google Scholar] [CrossRef]
  69. Tsamardinos, I., Rakhshani, A., & Lagani, V. (2015). Performance-estimation properties of cross-validation-based protocols with simultaneous hyper-parameter optimization. International Journal on Artificial Intelligence Tools, 24(05), 1540023. [Google Scholar] [CrossRef]
  70. Wei, M., Heppner, P. P., Mallen, M. J., Ku, T.-Y., Liao, K. Y.-H., & Wu, T.-F. (2007). Acculturative stress, perfectionism, years in the United States, and depression among Chinese international students. Journal of Counseling Psychology, 54(4), 385. [Google Scholar] [CrossRef]
  71. Witten, I. H., Frank, E., Hall, M. A., Pal, C. J., & DATA, M. (2005). Practical machine learning tools and techniques. In Data mining (Vol. 2). Elsevier. [Google Scholar]
  72. Wu, Y., Liu, L., Bae, J., Chow, K.-H., Iyengar, A., Pu, C., Wei, W., Yu, L., & Zhang, Q. (2019, December 9–12). Demystifying learning rate policies for high accuracy training of deep neural networks. 2019 IEEE International Conference on Big Data (Big Data) (pp. 1971–1980), Los Angeles, CA, USA. [Google Scholar]
  73. Xiong, Y., & Schneider, J. B. (1992). Transportation network design using a cumulative genetic algorithm and neural network. Transportation Research Record, 37–44. [Google Scholar]
  74. Yarkoni, T., & Westfall, J. (2017). Choosing prediction over explanation in psychology: Lessons from machine learning. Perspectives on Psychological Science, 12(6), 1100–1122. [Google Scholar] [CrossRef]
  75. Yong, H., Huang, J., Meng, D., Hua, X., & Zhang, L. (2020). Momentum batch normalization for deep learning with small batch size. In Computer vision–ECCV 2020: 16th european conference, Glasgow, UK, August 23–28, 2020, proceedings, part XII 16 (pp. 224–240). Springer International Publishing. [Google Scholar]
  76. Yoo, J. E. (2018). TIMSS 2011 student and teacher predictors for mathematics achievement explored and identified via elastic net. Frontiers in Psychology, 9, 317. [Google Scholar] [CrossRef] [PubMed]
  77. You, Y., Gitman, I., & Ginsburg, B. (2017). Scaling sgd batch size to 32k for imagenet training. arXiv, arXiv:1708.03888. [Google Scholar]
  78. Youyou, W., Kosinski, M., & Stillwell, D. (2015). Computer-based personality judgments are more accurate than those made by humans. Proceedings of the National Academy of Sciences, 112(4), 1036–1040. [Google Scholar] [CrossRef] [PubMed]
  79. Yu, T., & Zhu, H. (2020). Hyper-parameter optimization: A review of algorithms and applications. arXiv, arXiv:2003.05689. [Google Scholar]
  80. Zalake, N., & Naik, G. (2019, January 8). Generative chat bot implementation using deep recurrent neural networks and natural language understanding. Proceedings 2019: Conference on Technologies for Future Cities, Navi Mumbai, India. [Google Scholar]
  81. Zeinalizadeh, N., Shojaie, A. A., & Shariatmadari, M. (2015). Modeling and analysis of bank customer satisfaction using neural networks approach. International Journal of Bank Marketing, 33(6), 717–732. [Google Scholar] [CrossRef]
  82. Zhang, H., Weng, T.-W., Chen, P.-Y., Hsieh, C.-J., & Daniel, L. (2018). Efficient neural network robustness certification with general activation functions. Advances in Neural Information Processing Systems, 31, 4944–4953. [Google Scholar]
  83. Zou, H., & Hastie, T. (2005). Regularization and variable selection via the elastic net. Journal of the Royal Statistical Society Series B: Statistical Methodology, 67(2), 301–320. [Google Scholar] [CrossRef]
Table 1. A selection of full results.
Table 1. A selection of full results.
Cond119263898
Lp10000.93920.78840.42690.95900.1774
Lp25000.96180.89280.64770.97500.5197
Lp50000.97260.91970.77600.98150.6566
Lp100000.98130.94540.84860.98820.7479
lp200000.98820.96200.88850.99240.8455
Lp25000 0.96490.9023 0.8424
Lp30000 0.96670.9111 0.8538
Lp35000 0.97140.9090 0.8593
Lp40000 0.97060.9258 0.8808
Lp45000 0.97480.9208 0.8697
Lp50000 0.97650.9312 0.8761
Up10001.03251.14621.32201.01731.4159
Up25001.01971.09771.18371.01101.2486
Up50001.01551.06321.14131.00941.1807
Up100001.01241.04801.09291.00641.1430
Up200001.00841.03371.07101.00521.0994
Up25000 1.02761.0553 1.0679
Up30000 1.02721.0568 1.0578
Up35000 1.02481.0439 1.0480
Up40000 1.02211.0437 1.0488
Up45000 1.02211.0424 1.0457
Up50000 1.02131.0410 1.0463
Above10000.62600.27800.17100.58100.3970
Above25000.86400.31100.08800.88600.3190
Above50000.97200.42100.12500.97400.2350
Above100000.99500.48200.18300.99400.2050
Above200000.99800.58700.40500.99900.2250
Above25000 0.71100.6430 0.1660
Above30000 0.79500.7010 0.1460
Above35000 0.80000.7420 0.1500
Above40000 0.86300.7650 0.1400
Above45000 0.89100.8210 0.1670
Above50000 0.91200.8350 0.1540
Note: Cond refers to the condition selection as an example in this study, in which 1 stands for condition 1 in Table 2, 19 stands for condition 19 in Table 2, 26 stands for condition 26 in Table 2, 38 stands for condition 38 in Table 2, and 98 stands for condition 98 in Table 2; LPx, in which x is 1000, 2500, 5000, 10,000, 20,000, 30,000, 35,000, 40,000, 45,000, or 50,000, refers to the 2.5% percentile of the R n 2 R t 2 ; Upx, in which x is 1000, 2500, 5000, 10,000, 20,000, 30,000, 35,000, 40,000, 45,000, or 50,000, refers to the 97.5% percentile of the R n 2 R t 2 , and the criterion based on the theoretical maximum performance is calculated by the Lpx and Upx; and Abovex, in which x is 1000, 2500, 5000, 10,000, 20,000, 30,000, 35,000, 40,000, 45,000, or 50,000, refers to the percentage of R n 2 > R l 2 in 1000 replications. The missing value in sample sizes of 30,000, 35,000, 40,000, 45,000, and 50,000 means that a sample size that can satisfy both criteria can be found in a sample size of 1000, 2500, 5000, 10,000, or 20,000.
Table 2. Simulation results of the minimum sample size requirement and intervals of 1000 sample size.
Table 2. Simulation results of the minimum sample size requirement and intervals of 1000 sample size.
CondComplexLinearNonlinerErrorIVnumber R t 2 MSSRRMSSRA
1111130.719310002500
2211130.833510005000
3311130.876410001000
4112130.863910001000
5212130.900910001000
6312130.904210001000
7121130.816110005000
8221130.854610005000
9321130.896810001000
10122130.894110002500
11222130.905210002500
12322130.906610002500
13111430.242220,00020,000
14211430.378910,00020,000
15311430.714410001000
16112430.690110002500
17212430.791810002500
18312430.879210001000
19121430.322810,00040,000
20221430.4658500020,000
21321430.660010005000
22122430.684010005000
23222430.8456100020,000
24322430.877910001000
251111030.0273XX
262111030.1006X45,000
273111030.380110,0005000
281121030.203720,00020,000
292121030.592025005000
303121030.786210001000
311211030.0571XX
322211030.1021XX
333211030.178820,00035,000
341221030.272010,00020,000
352221030.4595500010,000
363221030.787410001000
37111150.6414250010,000
38211150.814010002500
39311150.887410001000
40112150.861510002500
41212150.899710002500
42312150.901910002500
43121150.789210005000
44221150.8416100010,000
45321150.880910002500
46122150.877210002500
47222150.898810002500
48322150.903410001000
49111450.0901XX
50211450.209820,00030,000
51311450.549050002500
52112450.516250005000
53212450.747210002500
54312450.865510002500
55121450.165020,000X
56221450.4475250020,000
57321450.584825005000
58122450.5867250010,000
59222450.782310002500
60322450.875810002500
611111050.0236XX
622111050.100950000X
633111050.291010,00010,000
641121050.167825,00020,000
652121050.4803500010,000
663121050.754610001000
671211050.0450XX
682211050.0875XX
693211050.250020,00020,000
701221050.240520,00020,000
712221050.4895500010,000
723221050.764910002500
731111100.5467250020,000
742111100.816310005000
753111100.853910002500
761121100.873510002500
772121100.892010002500
783121100.901710001000
791211100.7371500050,000
802211100.8356100010,000
813211100.865210005000
821221100.859610005000
832221100.896510005000
843221100.903910001000
851114100.0725XX
862114100.349510,00025,000
873114100.559350005000
881124100.4726500010,000
892124100.762510005000
903124100.851710002500
911214100.219920,000X
922214100.382910,00040,000
933214100.6038250010,000
941224100.5092500020,000
952224100.7743100010,000
963224100.875910002500
9711110100.0180XX
9821110100.0509XX
9931110100.158840,000X
10011210100.097025,000X
10121210100.4152500020,000
10231210100.664025002500
10312110100.0379XX
10422110100.1086XX
10532110100.157435,000X
10612210100.187520,000X
10722210100.4790500010,000
10832210100.631925005000
Note: ‘Cond’ refers to the condition in the simulation; ‘Complex’ indicates the complexity of the nonlinear relationship between IVs and the DV, where ‘1’ signifies simple complexity, ‘2’ medium complexity, and ‘3’ high complexity; ‘Linear’ denotes the categorical linear coefficient level, with ‘1’ representing coefficients simulated from the range of 0.1 to 0.3, and ‘2’ from 0.5 to 0.1; ‘Nonlinear’ refers to the categorical nonlinear coefficient level, with ‘1’ for coefficients from 0.1 to 0.3, and ‘2’ for coefficients from 0.5 to 0.1; ‘IVnumber’ refers to the number of independent variables; ‘ R t 2 ’ refers to the maximum variance that can be theoretically explained; ‘MSSRR’ refers to the minimum sample size required for the stability of the performance criterion; and ‘MSSRA’ refers to =the minimum sample size for the criterion of outperforming the linear model, with an ‘X’ indicating that the necessary sample size to meet this criterion cannot be found
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Cheng, Y.; Petrides, K.V.; Li, J. Estimating the Minimum Sample Size for Neural Network Model Fitting—A Monte Carlo Simulation Study. Behav. Sci. 2025, 15, 211. https://doi.org/10.3390/bs15020211

AMA Style

Cheng Y, Petrides KV, Li J. Estimating the Minimum Sample Size for Neural Network Model Fitting—A Monte Carlo Simulation Study. Behavioral Sciences. 2025; 15(2):211. https://doi.org/10.3390/bs15020211

Chicago/Turabian Style

Cheng, Yongtian, Konstantinos Vassilis Petrides, and Johnson Li. 2025. "Estimating the Minimum Sample Size for Neural Network Model Fitting—A Monte Carlo Simulation Study" Behavioral Sciences 15, no. 2: 211. https://doi.org/10.3390/bs15020211

APA Style

Cheng, Y., Petrides, K. V., & Li, J. (2025). Estimating the Minimum Sample Size for Neural Network Model Fitting—A Monte Carlo Simulation Study. Behavioral Sciences, 15(2), 211. https://doi.org/10.3390/bs15020211

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop