Next Article in Journal
Examining the Interaction Between Perceived Neighborhood Disorder, Positive Peers, and Self-Esteem on Adolescent Prosocial Behavior: A Study of Chinese Adolescents
Previous Article in Journal
The Therapeutic Benefits of Outdoor Experiences in India
Previous Article in Special Issue
Biased and Biasing: The Hidden Bias Cascade and Bias Snowball Effects
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

On the Continuum of Foundational Validity: Lessons from Eyewitness Science for Latent Fingerprint Examination

by
Adele Quigley-McBride
* and
T. L. Blackall
Department of Psychology, Simon Fraser University, Burnaby, BC V5A 1S6, Canada
*
Author to whom correspondence should be addressed.
Behav. Sci. 2025, 15(9), 1145; https://doi.org/10.3390/bs15091145
Submission received: 28 June 2025 / Revised: 1 August 2025 / Accepted: 15 August 2025 / Published: 22 August 2025
(This article belongs to the Special Issue Forensic and Legal Cognition)

Abstract

Whether forensic disciplines have established foundational validity—sufficient empirical evidence that a method reliably produces a predictable level of performance—has become a question of growing interest among scientists and legal professionals. This paper evaluates the foundational validity of two sources of forensic evidence relied upon in criminal cases: eyewitness identification decisions and latent fingerprint examiners’ conclusions. Importantly, establishing foundational validity and estimating accuracy are conceptually and functionally different. Though eyewitnesses can often be mistaken, identification procedures recommended by researchers are grounded in decades of programmatic research that justifies the use of methods that improve the reliability of eyewitness decisions. In contrast, latent print research suggests that expert examiners can be very accurate, but foundational validity in this field is limited by an overreliance on a handful of black-box studies, the dismissal of smaller-scale, yet high-quality, research, and a tendency to treat foundational validity as a fixed destination rather than a continuum. Critically, the lack of a standardized method means that any estimates of examiner performance are not tied to any specific approach to latent print examination. Despite promising early work, until the field adopts and tests well-defined procedures, foundational validity in latent print examination will remain a goal still to be achieved.

1. On the Continuum of Foundational Validity: Lessons from Eyewitness Science for Latent Fingerprint Examination

After decades of a comfortable reputation among investigators, legal professionals, and lay persons alike, forensic science faced increasing scrutiny in the 1990s and 2000s. The 2009 National Research Council report (NRC, 2009) is considered by many researchers and practitioners to be a turning point, as it clearly laid out several serious shortcomings in forensic science that had previously escaped formal scrutiny. The report was especially critical of pattern-matching disciplines—fields in which practitioners visually compare patterns or markings between two samples (e.g., latent fingerprint examination). In these disciplines, experts determine whether two pieces of evidence are sufficiently similar to conclude they share a common source (e.g., a suspect). Many researchers and practitioners acknowledged that the report exposed fundamental weaknesses in the scientific foundations of many forensic disciplines, prompting a wave of new research to address these concerns in forensic science.
There is now substantially more high-quality foundational research that can speak to reliability and the likelihood of error across a range of forensic disciplines. As a result, forensic service providers (FSPs) and legal professionals who use the results produced by forensic laboratories have started to ask how much research is enough. In particular, when considering court admissibility requirements for this type of information, the test in Frye v. United States (1923) requires “general acceptance in the relevant field”, the test from Daubert v. Merrel Dow Pharmaceuticals, Inc. (1993) requires knowledge of error rates, and the recently updated Federal Rules of Evidence, Rule 702 (2023) requires “reliable principles and methods”, generally in that discipline and as applied in that case. Thus, whether a forensic discipline has sufficient foundational research can, in theory, determine the extent to which that type of evidence can be relied upon in court cases.
In 2016, the President’s Council of Advisors on Science and Technology (PCAST, 2016) published a review of several forensic disciplines—single-source DNA, DNA mixtures, bitemarks, latent fingerprints, firearms, footwear, and hair analysis. Their mandate was to determine whether these disciplines had demonstrated foundational validity, defined as the extent to which a method has been empirically shown to produce accurate and consistent results, based on peer-reviewed, published studies (PCAST, 2016). Specifically, PCAST evaluated whether each discipline’s procedures had been tested for repeatability (within examiner), reproducibility (across examiner), and accuracy under conditions representative of actual casework (see also the discussion in Stern et al., 2019).
Importantly, PCAST emphasized that foundational validity is a property of the specific method under consideration, rather than a property of performance outcomes. That is, a discipline can lack foundational validity even when examiners achieve accurate results provided that their success cannot be attributed to a clearly defined and consistently applied method that can be independently replicated (see also Wilson et al., 2020). Given these implications, it is unsurprising that methodological precision and compliance have received increasing attention in recent years (Swofford et al., 2024). Without a clear and consistently applied method, results from studies designed to observe performance metrics reflect the accuracy achieved by an undefined mix of examiner strategies that cannot be meaningfully linked to any particular approach and are, consequently, difficult to interpret, predict, or replicate.
Of the disciplines reviewed, only single-source DNA, DNA mixtures with no more than three contributors, and latent print examination (LPE) passed the muster. However, although the PCAST report cited other studies, this declaration about LPE was based on a total of two black-box studies1 (Pacheco et al., 2014; Ulery et al., 2011), only one of which (Ulery et al.) had been published in a peer-reviewed journal. PCAST concluded that LPE could be considered foundationally valid if its limitations were clearly communicated when presenting results. Since the report, one additional LPE black-box study has been published (Hicklin et al., 2025), resembling those cited by PCAST. If black-box studies are indeed the primary evidence required to establish foundational validity, as the report implies, then the field remains in a similar position nearly a decade later.
So, is the existing body of empirical research on LPE sufficient to establish its foundational validity? In experimental psychology, most researchers would likely agree that three studies conducted under a narrow set of conditions may be promising and worthy of further exploration, but they are insufficient for making broad policy recommendations about the practices being evaluated. At the same time, researchers in psychology would not necessarily dismiss studies that fall outside of the strict black-box criteria outlined by PCAST (2016); a diverse range of empirical work can offer valuable insights. These considerations raise important questions: How much research is enough to consider a method foundationally valid? At what point can the latent fingerprint community be confident that their methods yield consistent results—both between (reproducible) and within examiners (repeatable; Stern et al., 2019)?
One way to answer this question is to look at other scientific fields that rely on expert perceptual judgments, such as radiology. Like forensic examiners, radiologists make judgments based on their perception and interpretation of complex, visually presented data to identify medical issues. However, radiology and other fields operating within medical systems benefit from methods that are well standardized, institutional safeguards designed to catch and correct errors, and formal oversight mechanisms. In contrast, many forensic disciplines, including LPE, operate under loosely defined frameworks and often lack the systemic infrastructure seen in fields like radiology. These differences, considered alongside the limited body of rigorous scientific research, make comparisons between forensic and medical disciplines difficult and potentially unproductive.
A better comparison might be found in the collection and use of a different type of forensic evidence: eyewitness identification. Like forensic pattern-matching disciplines, eyewitness evidence has faced, and continues to face, significant scrutiny over its reliability. Eyewitness evidence depends heavily on human perception and judgment, both on the part of the eyewitness and the investigators handling the case. Eyewitness data are typically collected under circumstances where decision-makers have wide discretion but are also often lacking the resources and personnel needed to adhere to research-based best practices (Wells et al., 2020). Both types of evidence have a strong influence on end-users—a confident eyewitness is highly persuasive (Garrett et al., 2021; Lindsay et al., 1989), much like a forensic expert is when they are testifying to their results during a trial (Cooper et al., 1996; Cramer et al., 2011). Many eyewitness identification studies focus on developing procedures that preserve the eyewitness’s memory for the suspect and appropriately test the extent to which the police suspect matches the eyewitness’s memory. The ultimate goal, as in pattern-matching disciplines, is to link a specific individual to a crime event.
Yet, eyewitnesses are mistaken at an alarming rate. Even when identification procedures follow best practices, approximately one-third of real eyewitnesses identify a known-innocent person (an intentionally placed filler; Wells et al., 2015). So, although research shows that eyewitnesses are less accurate than LPEs, there is a robust body of empirical research supporting the methods recommended for use in practice, which is important for foundational validity. This makes eyewitness identification a compelling example of how foundational validity can be achieved even in the face of known performance limitations.
There are also clear differences between eyewitness identification and LPE (see Table 1). For instance, eyewitness identification science examines a memory-based task rather than a perceptual comparison task. However, the cognitive processes and legal actors involved in establishing and using eyewitness evidence and forensic pattern-matching disciplines are remarkably similar (Quigley-McBride & Wells, 2018). Standardized eyewitness identification procedures are also supported by decades of research, beginning in the 1970s, clarifying how and why various factors affect eyewitness reliability.
For instance, there are multiple studies demonstrating that the quality of the initial memory will determine whether the eyewitness is able to make a correct identification later and how that is related to their certainty about that identification decision (Wixted & Wells, 2017). This is akin to the impact of the quality of the trace evidence collected from the crime scene, as this will limit what conclusions can be made by LPEs and the associated certainty (Dror & Kukucka, 2021). Research also shows that the methods used to collect eyewitness accounts can change both an eyewitness’s willingness to make an identification, their accuracy, and their confidence in their decision (Wells et al., 2020). This can be likened to the methods and standard operating procedures maintained by FSPs used to collect, examine, and evaluate forensic evidence (Swofford et al., 2024). Finally, any additional information the witness receives between the crime and the recalled event can affect their accuracy and confidence (Loftus, 2005), similar to the effect of various types of potentially biasing information on forensic examiners (Kassin et al., 2013).
In this paper, eyewitness identification science is presented as an example of a programmatic, rigorous body of cumulative science that sheds light on the strengths and weaknesses of different approaches to gathering forensically-relevant evidence. To be clear, though, the goal is not to present eyewitness science as the standard against which to compare forensic disciplines to determine foundational validity. Nor do we dispute studies that show low error rates among LPEs, suggesting that they can accurately sort between same- and different-source latent–exemplar pairs under the conditions that have been tested (e.g., Ulery et al., 2011; Hicklin et al., 2025). As noted, however, accuracy is not synonymous with foundational validity.
Foundational validity is not a “destination”. Forensic disciplines lie on a continuum from no foundational validity through to the aspirational goal of understanding everything about how and why a method or technique produces favorable outcomes under some circumstances and not in others. Eyewitness science has come a long way along this continuum, but the journey has not always been linear, and there is still much to test and understand. In this article, we compare and contrast the current state of eyewitness identification science with that of latent print examination—not because LPE lacks scientific grounding but because, aside from DNA analysis, it is the only other forensic discipline that has been recognized as possessing some degree of foundational validity by both practitioners and scientists (PCAST, 2016).

2. Where Does Eyewitness Science Lie on the Foundational Validity Continuum?

Eyewitness identification evidence is frequently used in the criminal legal system and has a broad body of research guiding best practices. Of course, this was not always the case. Eyewitness testimony also has a much longer history as a source of evidence, perhaps even since the very first person accused another of a misdeed, than physical forensic evidence, such as fingerprints. The emergence of systematic programs of research on eyewitness memory and identification procedures began in the 1960s and 1970s but really started to surge in the 1980s and 1990s (Wells & Quigley-McBride, 2016).
During this time, foundational work established what are now considered highly replicable, basic findings. For instance, early work by Professor Elizabeth Loftus and colleagues demonstrated how easy it was for information encountered after a crime event to alter memory, often without the eyewitness realizing it (Loftus & Greene, 1980), and, later, that entirely false memories could be implanted (Loftus & Pickrell, 1995). In the 1970s, Professor Gary Wells introduced a key conceptual distinction between system variables—factors that could be controlled by police and other agents in the legal system—and estimator variables—factors associated with the crime event, culprit, or witness that cannot be controlled, guiding applied research in this area (Wells, 1978). Wells and his colleagues demonstrated that the use of properly administered lineup procedures—system variables—was a reliable way for police to ensure an inaccurate eyewitnesses did not derail an investigation or result in the conviction of the wrong person. Throughout the 1970s, 1980s, and 1990s, eyewitness researchers ran studies to clarify best practices surrounding the creation and administration of lineups to ensure these procedures reduced the risk to innocent suspects while maintaining the ability to obtain an identification of the correct person as much as possible (e.g., Wells et al., 1979a, 1979b, 1993). By the late 1990s, eyewitness researchers had built up a robust collection of findings that had been replicated many times.
Representing a culmination of this work, the American Psychology-Law Society (AP-LS) produced a Scientific Review Paper (SRP) in 1998 (Wells et al., 1998). This document outlined four scientifically grounded recommendations that had been well-studied and replicated. At the time, there was a broad consensus among eyewitness identification researchers that each of these recommendations had sufficient empirical support to justify its use in real cases. Even if future research found reasons to refine these recommendations, the points in the 1998 SRP could be relied upon. Figure 1 shows the types of citations used to justify the recommendations in 1998. Of the 152 total citations, 59 were controlled, experimental studies, 30 were review papers drawing on the wider literature, and five were meta-analyses that combined across studies to more accurately estimate the effect of variables or procedures. While this reference list is not a comprehensive list of all studies conducted prior to 1998, it represented what experts in the discipline considered to be key contributions to the science.
Although there was enough work to make concrete recommendations for testing eyewitness memory in real cases, this area continued to grow post-1998. In fact, the area grew so much that in 2020, another SRP was released by the AP-LS. The 2020 SRP includes nine recommendations to aid in the collection, preservation, and evaluation of eyewitness evidence in the field, and for many common real-world situations, researchers can now draw on several studies that can help evaluate the reliability of eyewitness evidence under similar circumstances. The 2020 SRP (Wells et al., 2020) was supported by more than 200 citations, including 145 experimental studies—115 of which were published after the 1998 SRP (see Figure 2). The report also drew on a larger number of meta-analyses (13 total, 11 post-1998), review papers (33 total, 29 post-1998), and other empirical and non-empirical publications (87 total, with 54 published after 1998; see Figure 2). Of course, the raw number of studies cited in a review paper is a crude measure of scientific progress. However, these citations were selected by experts to support formal policy recommendations and vetted through several rounds of comments, suggesting they meet a standard of quality broadly accepted by researchers in the field.
As with any scientific discipline, not all avenues pursued by eyewitness researchers have withstood the test of time. Some effects—particularly those associated with estimator variables like lighting, exposure duration, weapon focus, and cross-race identification—have proven robust in laboratory settings where such factors can be carefully controlled. However, these same factors can be difficult to assess in real cases, as documentation is sparse and often relies on subjective eyewitness recollections, resulting in mixed success replicating these effects in archival and real-world samples (Horry et al., 2012). Even findings once thought to be well established, such as own-race bias (Meissner & Brigham, 2001), have shown greater variability in more recent research (Lee & Penrod, 2022).
Several system variable recommendations have also been revised or withdrawn in light of new evidence and do not appear in the 2020 SRP. For example, lineup administrators were once encouraged to include an “appearance change instruction,” warning witnesses that some of the culprit’s features may have changed since the crime (Charman & Wells, 2007). However, later research showed that this instruction increased eyewitnesses’ willingness to choose even when they were uncertain, increasing false identifications without boosting correct decisions (Molinaro et al., 2013). Similarly, sequential lineups—where photos are shown one at a time—were once preferred over simultaneous presentation (Lindsay & Wells, 1985). But a large field study found that, when other best practices were followed, there were only minimal differences in outcomes observed when lineups were presented sequentially versus simultaneously (Wells et al., 2015). Despite this, sequential lineups remain widely used in U.S. police departments (Greenspan et al., 2024), likely due to the broad success of disseminating this research in the past. Of course, it is clearly preferable to avoid retracted recommendations, but these examples demonstrate the self-correcting nature of science.
As a result of the body of research amassed by eyewitness researchers, there have been significant efforts to standardize the procedures used in the legal system by disseminating research-based recommendations to police and other policymakers. For example, the 2020 SRP provides guidance to police at every stage of the investigation. There are recommendations regarding the collection of eyewitness accounts after a crime event, how best to preserve eyewitness memory between the crime event and any subsequent recall session or identification procedure (e.g., asking witnesses not to discuss the event or search social media), how best to conduct an identification procedure (e.g., using a fair lineup with one suspect and five known-innocent fillers, administered by someone who does not know the suspect’s identity; (Wells et al., 2020)), and observations that can indicate whether an eyewitness’s identification was accurate or not (reflector variables; Wells, 2020). Thus, eyewitness science has developed programmatic research identifying specific, empirically supported methods for collecting and evaluating eyewitness evidence—methods shown to improve identification outcomes.
Since the 1998 SRP, there have been several field studies demonstrating the benefits of the recommended procedures with real eyewitnesses in real cases. These studies demonstrate the number of filler identifications (decisions were known to be inaccurate in real cases; Wells et al., 2015), the effect of a biased lineup (Steblay & Wells, 2020), and indicators of eyewitness accuracy (Quigley-McBride & Wells, 2023; Seale-Carlisle et al., 2019) in real cases. As a result, the evidence supporting these procedures is no longer limited to laboratory contexts, strengthening the credibility of this guidance. Of course, changes to police policies and common law take time, and standardization of methods remains inconsistent across agencies and jurisdictions in the USA (Greenspan et al., 2024). There will also always be real-world cases that are so unusual or unique that there is no existing research relevant to those case circumstances. Some real-world cases are also so unusual that no existing research is directly relevant. Still, so long as an agency documents the identification procedures used, there is ample empirical research addressing when and why eyewitness evidence is likely to be reliable—or not.
Though more research will always be necessary, eyewitness science has made substantial progress in identifying methods that produce reliable outcomes and using that research to bring about policy change. The nine recommendations made in the 2020 SRP have reached a level of foundational validity that almost every eyewitness researcher agrees with (Seale-Carlisle et al., 2024). Other recommendations enjoyed similar levels of empirical support and expert consensus but fell just short of the threshold for inclusion in the 2020 SRP—placing them slightly lower on the continuum of foundational validity.
Thus, eyewitness researchers have amassed enough evidence to recommend specific procedures and methods to preserve eyewitness evidence. Still, there are many areas of eyewitness work where researchers’ understanding is still limited or developing, such as the study of reflector variables (Wells, 2020). These are observations that can be made during or immediately after a lineup that change in predictable ways when an eyewitness is correct versus incorrect in any particular case. Foundational validity, or how best to collect and use this information, is still progressing in this area, though. One example is an eyewitness’s confidence in their decision—a highly confident eyewitness is more likely to be accurate than one who lacks confidence. However, researchers continue to debate how best to elicit, record, and interpret confidence and how to analyze it in experimental studies (Wixted & Wells, 2017). So, foundational validity can vary across topics even within a single field. Scientific progress requires continuous empirical testing—both for established procedures and for those still gaining empirical support.

3. Where Does Latent Fingerprint Examination Lie on the Foundational Validity Continuum?

Latent fingerprint examination (LPE), a well-known and generally accepted forensic pattern-matching discipline, has made significant strides in creating a body of programmatic research to provide a scientific foundation to the discipline. In fact, several topics relevant to LPE are supported by a growing body of peer-reviewed publications. These include, but are not limited to, the following:
In addition to peer-reviewed studies, considerable work has been developed outside traditional scientific channels—such as government reports, process maps, and consensus-based documents. Much of this growth has occurred since the 2009 NRC report.
While these developments contribute to the broader foundation of LPE, they are less relevant to the specific focus of this paper: the foundational validity of the human decision-making process in LPE. Just as the foundational validity of eyewitness identification depends on the methods used to collect and interpret witness memory, our focus is on the validity of the methods used by LPEs and the associated outcomes. This is related to, but distinct from, the question of LPE accuracy since there could be multiple foundationally valid methods that produce different levels of reliability and accuracy among LPEs that use them.
Controlled studies directly examining LPE judgments are fairly limited, with only a handful of black-box type studies (e.g., Busey et al., 2021; Hicklin et al., 2025; Koehler & Liu, 2021; Pacheco et al., 2014; Ulery et al., 2011), a small body of other work examining specific questions about examiner performance (e.g., Langenburg et al., 2012; Neumann et al., 2015; Tangen et al., 2011), and blind-proficiency testing efforts (Gardner et al., 2021). These studies report mixed findings. Some suggest low false positive error rates (<1%; Ulery et al., 2011; Hicklin et al., 2025) while others report higher rates of error (15.9% and 18.1%; Koehler & Liu, 2021), depending on the conditions and materials tested.
Although fingerprint examination has better empirical support than many other forensic disciplines, is the existing body of research sufficient to establish the foundational validity of LPE judgments? Some researchers, practitioners, and government entities have suggested that it is. For instance, as noted here in the introduction, the PCAST report (2016) concluded that “latent fingerprint analysis is a foundationally valid subjective methodology—albeit with a false positive rate that is substantial and likely to be higher than expected by many jurors based on longstanding claims about the infallibility of fingerprint analysis” (p. 9). However, this conclusion was based on “only two properly designed studies of the foundational validity and accuracy of latent fingerprint analysis” (p. 101), as other studies measuring error rates among LPEs failed to meet PCAST’s criteria for methodological rigor. Although it is standard in science to interpret results cautiously based on study limitations, the studies dismissed by PCAST are not without value. In fact, subsequent peer-reviewed publications referring to the report have critiqued PCAST’s treatment of the LPE literature and the question of foundational validity (e.g., Hicklin et al., 2025; Koehler & Liu, 2021).
Though limited, the existing research provides valuable insights into LPEs’ abilities. Under optimal conditions—such as when LPEs are not rushed, fatigued, stressed, or exposed to task-irrelevant information—false positive rates are generally below 1%, although false negative rates tend to be higher (e.g., 4.2% in Hicklin et al., 2025; 7.5% in Ulery et al., 2011). Furthermore, exclusion decisions appear to involve qualitatively different decision processes than identification decisions (Ulery et al., 2017). LPEs find fingerprint comparisons more difficult when the latent fingerprint is of poor quality (e.g., smudged or missing information), and they are more likely to find such prints to lack value for comparison (Ulery et al., 2013, 2014).
Expert LPEs consistently outperform laypersons, especially when the latent and the exemplar fingerprints appear highly similar but are not from the same source (a “close non-match”; Tangen et al., 2011). However, these close non-matches are also associated with elevated error rates among expert LPEs (Koehler & Liu, 2021). Research shows that LPEs can expect to examine close non-matches increasingly often due to improvements in database search algorithms that draw on databases with many potential candidates, increasing the chance that a very similar fingerprint from a different person is identified as a candidate (Busey et al., 2014; Li et al., 2021).
Studies also demonstrate that LPE relies heavily on subjective human judgment, evidenced by both intra- and inter-examiner inconsistencies (Ulery et al., 2013, 2014, 2015, 2016). Emerging evidence suggests that individual differences in pattern-matching abilities might be trait-like, with some people naturally more adept at these tasks than others (Growns et al., 2022, 2023). Finally, forensic examiners may behave more conservatively (i.e., less likely to make same-source judgment) when they know they are being tested, which is consistent with similar findings in other disciplines and is referred to as the Hawthorne effect (McCambridge et al., 2014; Risinger et al., 2002; Scurich et al., 2025).
Despite these advances, important limitations and knowledge gaps remain. Many findings have not been replicated across a variety of LPE samples—a significant portion of the error-rate literature originates from a single research group working with a dataset collected through a Noblis–FBI partnership (e.g., Hicklin et al., 2025; Ulery et al., 2011, 2012, 2013, 2014, 2015, 2016). While these studies are high-quality and influential, scientific fields benefit from research conducted by a diverse range of research teams with different approaches and perspectives. Moreover, existing error rate studies tend to rely on fingerprint impressions that are already processed to some degree, higher quality than latent fingerprints lifted from crime scenes, and examined under conditions where the LPEs know that they are being tested by researchers who will compare their judgments to the correct answers. There are also questions about whether the rate of same and different source latent–exemplar pairs in studies matches what occurs in regular casework. These are important deviations from casework that are known to impact how people make decisions—generally and specifically in LPE (Growns & Kukucka, 2021; Haber & Haber, 2014; Scurich et al., 2025; see also critiques of fingerprint proficiency testing in (Kelley et al., 2020); Koertner & Swofford, 2018). Thus, the error rates in operational settings may be different than those reported in the literature—particularly when considering the prevalence of close non-matches (Koehler & Liu, 2021) and the influence of task-irrelevant, biasing contextual information (Dror & Kukucka, 2021; Quigley-McBride et al., 2022) in real cases, which tend to change decision-making patterns.
Even if there were more error rate studies to draw on, though, we argue that concluding that LPE is foundationally valid would still be premature. Though the literature demonstrates that LPEs are able to distinguish between same- and different-source latent–exemplar fingerprint pairs at a much higher rate than novices and lay individuals, determining accuracy levels is different from determining whether a discipline is foundationally valid. Therefore, the most significant barrier to advancing foundational validity in LPE is the lack of a consistent methodological approach to examinations. Without consensus on a sufficiently detailed set of procedures and criteria—or even several clearly defined approaches—foundational validity cannot be established.
This is not to say that there is no consistency at all among LPEs. Most LPEs are trained to follow a procedural framework known as “ACE-V”, which outlines an order of operations. The examiner first Analyzes the latent fingerprint obtained from the crime scene to determine if there is a sufficient number of usable features such that the impression could be reasonably compared with another impression. Next, if the latent fingerprint is deemed to have analytic value, the examiner will perform a visual Comparison between the latent fingerprint and a fingerprint from a suspect or database search. Subsequently, the LPE will Evaluate the data and discern whether it is their opinion that these impressions are from the same source (“identification”), a different source (“exclusion”), or that there is not enough information to decide either way (“inconclusive”). Finally, that decision will be Verified by another examiner (Langenburg et al., 2009; Vanderkolk, 2011).
Although widely adopted, ACE-V prescribes only a general sequence of steps but does not standardize techniques or decision thresholds, leaving considerable room for examiner discretion. For example, ACE-V does not define how many or what types of features are required for comparison or identification, despite wide variability in these judgments (Ulery et al., 2013, 2014). Such subjectivity can lead to inconsistencies both across examiners (reproducibility) and within the same examiner under different conditions (repeatability; Ulery et al., 2012). Moreover, ACE-V includes a “Verification” step, but does not specify how often verification should occur and for what types of decisions, or how that verification procedure should look. In particular, there is a lack of guidance on the importance of using a blind verifier—another examiner that does not know the original examiner’s conclusions. The difference in the potential for error with blind versus open verification methods, broadly defined, could be significant (Kukucka & Dror, 2023), but both are acceptable when using ACE-V (Vanderkolk, 2011) according to existing standards (e.g., ANSI/ASB BPR 144, (AAFS Standards Board, 2022)).
Consequently, the ACE-V process, in its current form, does not ensure uniformity in decision-making or provide sufficient safeguards against examiner error. This flexibility also means that error rates reported in black box studies are not anchored to any standardized procedure and examiners participating in these studies may have employed vastly different internal criteria, despite ostensibly operating “within” the ACE-V framework. The literature currently suggests that false positive error rates could be less than 1% under some conditions and greater than 18% under others. Yet, it remains unclear how much of this variation is attributable to differences in the procedures and decision criteria used by each LPE.
In a recent publication, Swofford and colleagues (2024) highlighted the important distinction between method performance and method conformance in the development of forensic disciplines. Once a standardized method is implemented among LPEs, the performance of that method can be assessed in terms of reproducibility, repeatability, and false positive and negative error rates. The question of decision outcomes and performance, however, is separate from examining whether a given LPE in any particular case adhered to that standardized method. If an examiner deviates from the prescribed procedure, any resulting error may reflect non-conformance rather than a flaw in the method itself. Accordingly, empirical efforts should assess not only the reliability of standardized methods under controlled conditions but also the effects of deviations from those methods on the accuracy and defensibility of forensic conclusions.
This issue is not unique to fingerprint examination; it is a pervasive, systemic issue across forensic sciences. In the United States, even when consensus-based standards and best practice recommendations are published by organizations, such as the Academy Standards Board (ASB) or the Organization of Scientific Area Committees (OSAC), adherence is not mandatory, and oversight mechanisms are minimal or nonexistent. As a result, an FSP may claim to have adopted a standard, but there is no guarantee that it is being consistently followed in practice. Consequently, the protocols used by LPEs in the USA often vary by jurisdiction, laboratory, and even among individual LPEs—sometimes substantially. This reduces the applicability of the few existing studies on error rates among LPEs. Without knowing which methods, criteria, or procedures were used by the LPEs who participated in these studies, it remains an open question whether their reported performance metrics generalize to the method used in any given case.

4. Conclusions

This comparative analysis demonstrates that latent fingerprint examination currently occupies a lower position on the foundational validity continuum than eyewitness identification science. We arrive at this conclusion based on two overarching factors: (1) the current state of the empirical literature and (2) the lack of a standardized methodology in LPE.

4.1. What Is Known About Latent Fingerprint Examination Is Based on a Nascent Empirical Literature

A key difference between the two disciplines examined here is the breadth, depth, and strength of their respective evidence bases. When discussing the reliability of LPE, researchers, practitioners, and policymakers typically rely on just a few studies that meet the criteria for a “black-box” study (e.g., Pacheco et al., 2014; Ulery et al., 2011; Hicklin et al., 2025). However, the performance of eyewitnesses during identification procedures can be estimated by referencing many studies—over 200 cited in the 2020 SRP—conducted by a variety of research groups. Individual studies might vary in methodological rigor or realism, but the cumulative literature should be considered as a whole—ideally, some studies will feature the same limitations, but others will have addressed them.
This is not to say that eyewitness science has a perfect record. Some studies have been undermined by subsequent findings, resulting in updates to policy recommendations. Some subtopics remain underdeveloped and still lack the ability to comprehensively explain the reasons for and meaning of particular eyewitness behaviors and decisions. Applying the current body of research in practice is also challenging due to ethical and practical difficulties associated with obtaining data from real eyewitnesses in real cases. Nevertheless, there is a substantial and consistent body of evidence showing that eyewitness accuracy improves when police use a double-blind, fairly constructed lineup with clear, unbiased instructions—enough to support strong confidence in the real-world benefits of these practices (Wells et al., 2020).
Before LPE can reach a comparable level of foundational validity, the field must develop a similarly systematic and reproducible evidence base. Moreover, we argue that researchers in the LPE discipline move away from relying on large-scale, black-box studies. Though such studies are critical, a range of high-quality empirical work would contribute meaningfully to this literature, such as controlled experiments, field research, and mixed methods approaches that examine how specific procedures and contextual factors influence examiner performance. No single study can answer every question about how easy it is to use and adhere to a forensic method, or the associated performance of that method—but a diverse, cumulative body of research can gradually build a coherent understanding, despite the limitations of individual projects.

4.2. Latent Fingerprint Examination Does Not Have a Standardized Methodology

Perhaps the most important issue relevant to the foundational validity of LPE is the absence of a standardized methodology that is used in practice. While studies show that LPEs can be accurate when comparing latent and exemplar prints, the field is missing guidance on the circumstances, procedures, and decision thresholds that guarantee those levels of accuracy. To be clear, the goal is not to eliminate professional discretion, but the amount of leeway given to examiners in their work should be intentional and, ideally, based on empirical research.
Without first determining how LPEs should approach examinations given evidence of different levels of quality and different case circumstances, assessing the reliability of the discipline is impossible. In contrast, eyewitness research has long focused on two core procedures—the showup (presentation of a single suspect) and the lineup (presentation of a suspect embedded among fillers known to be innocent). Both procedures have been evaluated extensively under different conditions, enabling researchers to offer empirically grounded recommendations about how specific procedural choices influence outcomes—such as lineup size, instructions, or selection criteria.
Foundational validity, by definition, requires a defined method and evidence of how that method performs under known conditions. The lack of standardization in LPE creates a serious barrier to advancing latent fingerprint examination along the foundational validity continuum. In the United States, LPE practices vary widely across jurisdictions, FSPs, and even individual examiners, with not universally applied decision thresholds or procedures. Although multiple validated methods could be acceptable, standardization should be preceded by empirical testing of potential approaches to ensure they support inter- and intra-examiner consistency. Without, defined, consistently applied methods, estimating the potential for error in any particular case is scientifically indefensible. Importantly, this challenge is not unique to LPE—it reflects a broader problem across many forensic science disciplines.

4.3. Lessons from Eyewitness Reform and the Path Forward for Latent Prints

Eyewitness errors remain the most common contributing factor to wrongful convictions in DNA exoneration cases, and data suggests that these errors have contributed to more than 60% of Innocence Project exonerations (Innocence Project, 2025). Eyewitness evidence is especially influential—both during investigations and in court proceedings (Garrett et al., 2021). That said, misapplied or invalidated forensic science, including latent print evidence, follows closely behind, contributing to roughly half of all DNA exoneration cases (Innocence Project, 2025). This is especially troubling in light of other evidence that laypeople (Chin & Ibaviosa, 2022; Eldridge, 2019) and legal professionals (de Keijser & Elffers, 2012) tend to struggle with evaluating the reliability and probative value of forensic evidence. Given the real-world consequences of unreliable evidence, a clear understanding of the circumstances under which these forms of evidence can be relied upon is critical to safeguarding the integrity of the justice system.
When the Innocence Project began using DNA to exonerate wrongfully convicted individuals and identify sources of error, the eyewitness literature was already equipped with a strong empirical foundation (Wells et al., 1998) and expert consensus (Kassin et al., 1989, 2001). As a result, when reforms were needed, the field was prepared to offer concrete, evidence-based recommendations. In contrast, other forensic disciplines—particularly the pattern-matching sciences—were not similarly prepared, a fact underscored by the findings in the 2009 National Research Council report.
This does not mean that LPE decision-making cannot achieve a level of scientific rigor comparable to eyewitness identification. The field has already shown potential, with early research providing important insights. These accomplishments could be effectively highlighted through large-scale, critical reviews, similar to the two eyewitness SRPs discussed here, to synthesize what is well established and identify areas needing further study. What is concerning, however, is the growing inclination to treat a small set of studies as sufficient to establish foundational validity, or to dismiss targeted, well-designed studies simply because they do not meet PCAST’s strict criteria for black-box research. Furthermore, LPE—and forensic science disciplines generally—would benefit from reconceptualizing foundational validity as a continuum rather than a fixed destination (Wilson et al., 2020). To advance, the field must commit to ongoing programmatic research efforts, the development and validation of standardized methods, and a commitment to identifying the features of methods that facilitate accurate LPE judgments—and those that do not.

Author Contributions

Conceptualization, A.Q.-M. and T.L.B.; formal analysis, A.Q.-M.; investigation, A.Q.-M.; writing—original draft preparation, A.Q.-M.; writing—review and editing, A.Q.-M. and T.L.B.; visualization, A.Q.-M. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data were derived from coding the reference lists in Wells et al. (1998, 2020). Our coded data can be accessed on OSF (https://osf.io/68srz/ (accessed on 15 June 2025)).

Acknowledgments

We would like to thank Fiona John, a Research Assistant in our lab, for her work coding the references in the Wells et al. (1998, 2020) Scientific Review Papers.

Conflicts of Interest

The authors declare no conflicts of interest.

Note

1
Black-box studies test forensic examiner performance only—the decision processes and reasoning of the examiners are not evaluated, just the final decisions. PCAST stated that these were studies in which “many examiners are given the same test items and asked to reach conclusions, without information about the ground truth or about the responses of other examiners. The study measures how often examiners reach correct conclusions, erroneous conclusions, and inconclusive results.” (PCAST, 2016, p. 6). PCAST also specified that the materials and procedures should closely resemble casework.

References

  1. AAFS Standards Board. (2022). ANSI/ASB best practice recommendation 144: Best practice recommendations for the verification component in friction ridge examination. American Academy of Forensic Sciences. [Google Scholar]
  2. Ashbaugh, D. R. (1999). Quantitative-qualitative friction ridge analysis: An introduction to basic and advanced ridgeology. CRC Press. [Google Scholar]
  3. Busey, T., Heise, N., Hicklin, R. A., Ulery, B. T., & Buscaglia, J. (2021). Characterizing missed identifications and errors in latent fingerprint comparisons using eye-tracking data. PLoS ONE, 16(5), e0251674. [Google Scholar] [CrossRef]
  4. Busey, T., Silapiruiti, A., & Vanderkolk, J. (2014). The relation between sensitivity, similar non-matches and database size in fingerprint database searches. Law, Probability, and Risk, 13, 151. [Google Scholar] [CrossRef]
  5. Charman, S. D., & Wells, G. L. (2007). Eyewitness lineups: Is the appearance-change instruction a good idea? Law and Human Behavior, 31(1), 3–22. [Google Scholar] [CrossRef] [PubMed]
  6. Chin, J. M., & Ibaviosa, C. M. (2022). Beyond CSI: Calibrating public beliefs about the reliability of forensic science through openness and transparency. Science & Justice, 62(3), 272–283. [Google Scholar] [CrossRef] [PubMed]
  7. Cooper, J., Bennett, E. A., & Sukel, H. L. (1996). Complex scientific testimony: How do jurors make decisions? Law and Human Behavior, 20(4), 379–394. [Google Scholar] [CrossRef]
  8. Cramer, R. J., DeCoster, J., Harris, P. B., Fletcher, L. M., & Brodsky, S. L. (2011). A confidence-credibility model of expert witness persuasion: Mediating effects and implications for trial consultation. Consulting Psychology Journal: Practice and Research, 63(2), 129–137. [Google Scholar] [CrossRef]
  9. Daubert v. Merrel Dow Pharmaceuticals, Inc. (1993). 509 U.S.  579.
  10. de Keijser, J., & Elffers, H. (2012). Understanding of forensic expert reports by judges, defense lawyers and forensic professionals. Psychology, Crime & Law, 18(2), 191–207. [Google Scholar] [CrossRef]
  11. Dhaneshwar, R., Kaur, M., & Kaur, M. (2021). An investigation of latent fingerprinting techniques. Egyptian Journal of Forensic Science, 11(1), 33. [Google Scholar] [CrossRef]
  12. Dror, I. E., & Kukucka, J. (2021). Linear Sequential Unmasking-Expanded (LSU-E): A general approach for improving decision making as well as minimizing noise and bias. Forensic Science International: Synergy, 3, 100161. [Google Scholar] [CrossRef]
  13. Eldridge, H. (2019). Juror comprehension of forensic expert testimony: A literature review and gap analysis. Forensic Science International: Synergy, 1, 24–34. [Google Scholar] [CrossRef]
  14. Federal Rules of Evidence, Rule 702. (2023). Federal rules of evidence, testimony by expert witnesses, 702. Available online: https://www.uscourts.gov/sites/default/files/2025-02/federal-rules-of-evidence-dec-1-2024_0.pdf (accessed on 15 June 2025).
  15. Frye v. United States. (1923). 293 F. 1013 (D.C. Cir. 1923).
  16. Gardner, B. O., Kelley, S., & Neuman, M. (2021). Latent print comparisons and examiner conclusions: A field analysis of case processing in one crime laboratory. Forensic Science International, 319, 110642. [Google Scholar] [CrossRef]
  17. Garrett, B. L., Liu, A., Kafadar, K., Yaffe, J., & Dodson, C. S. (2021). Factoring the role of eyewitness evidence in the courtroom. Journal of Empirical Legal Studies, 17(3), 556–579. [Google Scholar] [CrossRef]
  18. Greenspan, R. L., Quigley-McBride, A., Bluestine, M. B., & Garrett, B. (2024). Psychological science from research to policy: Eyewitness identifications in Pennsylvania police agencies. Psychology, Public Policy, and Law, 30(4), 462–478. [Google Scholar] [CrossRef]
  19. Growns, B., Dunn, J. D., Mattijssen, J. A. T., Quigley-McBride, A., & Towler, A. (2022). Match me if you can: Evidence for a domain-general visual comparison ability. Psychonomic Bulletin & Review, 29, 866–881. [Google Scholar] [CrossRef]
  20. Growns, B., & Kukucka, J. (2021). The prevalence effect in fingerprint identification: Match and non-match base-rates impact misses and false alarms. Applied Cognitive Psychology, 35(3), 751–760. [Google Scholar] [CrossRef]
  21. Growns, B., Towler, A., & Martire, K. A. (2023). The novel object-matching test (NOM test): A psychometric measure of visual comparison ability. Behavior Research Methods, 56, 680–689. [Google Scholar] [CrossRef]
  22. Haber, R. N., & Haber, L. (2014). Experimental results of fingerprint comparisons validity and reliability: A review and critical analysis. Science & Justice, 54(5), 375–389. [Google Scholar] [CrossRef]
  23. Hicklin, R. A., Richetelli, N., Taylor, A., & Buscaglia, J. (2025). Accuracy and reproducibility of latent print decision on comparisons from searches of an automated fingerprint identification system. Forensic Science International, 370, 112457. [Google Scholar] [CrossRef] [PubMed]
  24. Holder, E. H., Robinson, L. O., & Laub, J. H. (2011). The fingerprint sourcebook (Chapters 9.1–9.26). Office of Justice Programs, National Institute of Justice.
  25. Horry, R., Memon, A., Wright, D. B., & Milne, R. (2012). Predictors of eyewitness identification decisions from video lineups in England: A field study. Law and Human Behavior, 36(4), 257–265. [Google Scholar] [CrossRef]
  26. Innocence Project. (2025). Eyewitness misidentification. Available online: https://innocenceproject.org/eyewitness-misidentification/ (accessed on 15 June 2025).
  27. Kassin, S. M., Dror, I. E., & Kukucka, J. (2013). The forensic confirmation bias: Problems, perspectives, and proposed solutions. Journal of Applied Research in Memory and Cognition, 2(1), 42–52. [Google Scholar] [CrossRef]
  28. Kassin, S. M., Ellsworth, P. C., & Smith, V. L. (1989). The “general acceptance” of psychological research on eyewitness testimony: A survey of the experts. American Psychologist, 44(8), 1089–1098. [Google Scholar] [CrossRef]
  29. Kassin, S. M., Tubb, V. A., Hosch, H. M., & Memon, A. (2001). On the “general acceptance” of eyewitness testimony research. A new survey of the experts. The American Psychologist, 56(5), 405–416. [Google Scholar] [CrossRef] [PubMed]
  30. Kelley, S., Gardner, B. O., Murrie, D. C., Pan, K. D., & Kafadar, K. (2020). How do latent print examiners perceive proficiency testing? An analysis of examiner perceptions, performance, and print quality. Science & Justice, 60(2), 120–127. [Google Scholar] [CrossRef] [PubMed]
  31. Koehler, J. J., & Liu, S. (2021). Fingerprint error rate on close non-matches. Journal of Forensic Sciences, 66(1), 129–134. [Google Scholar] [CrossRef] [PubMed]
  32. Koertner, A. J., & Swofford, H. J. (2018). Comparison of latent print proficiency tests with latent prints obtain in routine casework using automatic and objective quality metrics. Journal of Forensic Identification, 68(3), 379. [Google Scholar]
  33. Kukucka, J., & Dror, I. E. (2023). Human factors in forensic science: Psychological causes of bias and error. In D. DeMatteo, & K. C. Scherr (Eds.), The Oxford handbook of psychology and law (pp. 621–642). Oxford University Press. [Google Scholar] [CrossRef]
  34. Langenburg, G., Champod, C., & Genessay, T. (2012). Informing the judgments of fingerprint analysts using quality metric and statistical assessment tools. Forensic Science International, 219(1–3), 183–198. [Google Scholar] [CrossRef]
  35. Langenburg, G., Champod, C., & Wertheim, P. (2009). Testing for potential contextual bias effects during the verification stage of the ACE-V methodology when conducting fingerprint comparisons. Journal of Forensic Sciences, 54(3), 571–582. [Google Scholar] [CrossRef]
  36. Lee, J., & Penrod, S. D. (2022). Three-level meta-analysis of the other-race bias in facial identification. Applied Cognitive Psychology, 36(5), 1106–1130. [Google Scholar] [CrossRef]
  37. Li, K., Wu, D., Ai, L., & Luo, Y. (2021). The influence of close non-match fingerprints similar in delta regions of whorls on fingerprint identification. Journal of Forensic Sciences, 66(4), 1482–1494. [Google Scholar] [CrossRef]
  38. Lindsay, R. C., & Wells, G. L. (1985). Improving eyewitness identifications from lineups: Simultaneous versus sequential lineup presentation. Journal of Applied Psychology, 70(3), 556–564. [Google Scholar] [CrossRef]
  39. Lindsay, R. C., Wells, G. L., & O’Connor, F. J. (1989). Mock-juror belief of accurate and inaccurate eyewitnesses: A replication and extension. Law and Human Behavior, 13(3), 333–339. [Google Scholar] [CrossRef]
  40. Loftus, E. F. (2005). Planting misinformation in the human mind: A 30-year investigation of the malleability of memory. Learning & Memory, 12(4), 361–366. [Google Scholar] [CrossRef] [PubMed]
  41. Loftus, E. F., & Greene, E. (1980). Warning: Even memory for faces may be contagious. Law and Human Behavior, 4(4), 323–334. [Google Scholar] [CrossRef]
  42. Loftus, E. F., & Pickrell, J. E. (1995). The formation of false memories. Psychiatric Annals, 25(12), 720–725. [Google Scholar] [CrossRef]
  43. McCambridge, J., Witton, J., & Elbourne, D. R. (2014). Systematic review of the Hawthorne effect: New concepts are needed to study research participation effects. Journal of Clinical Epidemiology, 67(3), 267–277. [Google Scholar] [CrossRef]
  44. Meissner, C. A., & Brigham, J. C. (2001). Thirty years of investigating the own-race bias in memory for faces: A meta-analytic review. Psychology, Public Policy, and Law, 7(1), 3–35. [Google Scholar] [CrossRef]
  45. Molinaro, P. F., Arndorfer, A., & Charman, S. D. (2013). Appearance-change instruction effects on eyewitness lineup identification accuracy are not moderated by amount of appearance change. Law and Human Behavior, 37(6), 432–440. [Google Scholar] [CrossRef] [PubMed]
  46. National Research Council. (2009). Strengthening forensic science in the United States: A path forward. National Academies Press. [Google Scholar]
  47. Neumann, C., Champod, C., Yoo, M., Genessay, T., & Langenburg, G. (2015). Quantifying the weight of fingerprint evidence through the spatial relationship, directions and types of minutiae observed on fingermarks. Forensic Science International, 248, 154–171. [Google Scholar] [CrossRef]
  48. Pacheco, I., Cerchiai, B., & Stoiloff, S. (2014). Miami-Dade research study for the reliability of the ACE-V process: Accuracy & precision in latent fingerprint examinations [Unpublished Report]. U.S. Department of Justice Office of Justice Programs.
  49. President’s Council of Advisors on Science and Technology. (2016). Forensic science in criminal courts: Ensuring scientific validity of feature-comparison methods. Executive Office of the President.
  50. Quigley-McBride, A., Dror, I. E., Roy, T., Garrett, B. L., & Kukucka, J. (2022). A practical tool for information management in forensic decisions: Using Linear Sequential Unmasking-Expanded (LSU-E) in casework. Forensic Science International: Synergy, 4, 100216. [Google Scholar] [CrossRef]
  51. Quigley-McBride, A., & Wells, G. L. (2018). Fillers can help control for contextual bias in forensic comparison tasks. Law and Human Behavior, 42(4), 295–305. [Google Scholar] [CrossRef]
  52. Quigley-McBride, A., & Wells, G. L. (2023). Eyewitness confidence and decision time reflect identification accuracy in actual police lineups. Law and Human Behavior, 47(2), 333–347. [Google Scholar] [CrossRef]
  53. Risinger, D. M., Saks, M. J., Thompson, W. C., & Rosenthal, R. (2002). The Daubert/Kumho implications of observer effects in forensic science: Hidden problems of expectation and suggestion. California Law Review, 90(1), 1–56. [Google Scholar] [CrossRef]
  54. Scurich, N., Albright, T. D., Stout, P., Eudaley, D., Neuman, M., & Hundl, C. (2025). The Hawthorne effect in studies of firearm and toolmark examiners. Journal of Forensic Sciences, 70(4), 1329–1337. [Google Scholar] [CrossRef] [PubMed]
  55. Seale-Carlisle, T. M., Colloff, M. F., Flowe, H. D., Wells, W., Wixted, J. T., & Mickes, L. (2019). Confidence and response time as indicators of eyewitness identification accuracy in the lab and in the real world. Journal of Applied Research in Memory and Cognition, 8(4), 420–428. [Google Scholar] [CrossRef]
  56. Seale-Carlisle, T. M., Quigley-McBride, A., Teitcher, J. E. F., Crozier, W. E., Dodson, C. S., & Garrett, B. L. (2024). New insights on expert opinion about eyewitness memory research. Perspectives on Psychological Science, 17456916241234837, advance online publication. [Google Scholar] [CrossRef]
  57. Steblay, N. K., & Wells, G. L. (2020). Assessment of bias in police lineups. Psychology, Public Policy, and Law, 26(4), 393–412. [Google Scholar] [CrossRef]
  58. Stern, H. S., Cuellar, M., & Kaye, D. (2019). Reliability and validity of forensic science evidence. Significance, 16(2), 21–24. [Google Scholar] [CrossRef]
  59. Swofford, H. J., Koertner, A. J., Zemp, F., Ausdemore, M., Liu, A., & Salyards, M. J. (2018). A method for the statistical interpretation of friction ridge skin impression evidence: Method development and validation. Forensic Science International, 287, 113–126. [Google Scholar] [CrossRef] [PubMed]
  60. Swofford, H. J., Lund, S., Iyer, H., Butler, J., Soons, J., Thompson, R., Desiderio, V., Jones, J. P., & Ramotowski, R. (2024). Inconclusive decisions and error rates in forensic science. Forensic Science International: Synergy, 8, 100472. [Google Scholar] [CrossRef]
  61. Tangen, J. M., Thompson, M. B., & McCarthy, D. J. (2011). Identifying fingerprint expertise. Psychological Science, 22(8), 995–997. [Google Scholar] [CrossRef]
  62. Ulery, B. T., Hicklin, R. A., Buscaglia, J., & Roberts, M. A. (2011). Accuracy and reliability of forensic latent fingerprint decisions. Proceedings of the National Academy of Sciences, 108(19), 7733–7738. [Google Scholar] [CrossRef]
  63. Ulery, B. T., Hicklin, R. A., Buscaglia, J., & Roberts, M. A. (2012). Repeatability and reproducibility of decisions by latent fingerprint examiners. PLoS ONE, 7(3), e32800. [Google Scholar] [CrossRef]
  64. Ulery, B. T., Hicklin, R. A., Kiebuzinski, G. I., Roberts, M. A., & Buscaglia, J. (2013). Understanding the sufficiency of information for latent fingerprint value determinations. Forensic Science International, 230(1–3), 99–106. [Google Scholar] [CrossRef]
  65. Ulery, B. T., Hicklin, R. A., Roberts, M. A., & Buscaglia, J. (2014). Measuring what latent fingerprint examiners consider sufficient information for individualization determinations. PLoS ONE, 9(11), e110179. [Google Scholar] [CrossRef]
  66. Ulery, B. T., Hicklin, R. A., Roberts, M. A., & Buscaglia, J. (2015). Changes in latent fingerprint examiners’ markup between analysis and comparison. Forensic Science International, 247, 54–61. [Google Scholar] [CrossRef] [PubMed]
  67. Ulery, B. T., Hicklin, R. A., Roberts, M. A., & Buscaglia, J. (2016). Interexaminer variation of minutia markup on latent fingerprints. Forensic Science International, 264, 89–99. [Google Scholar] [CrossRef]
  68. Ulery, B. T., Hicklin, R. A., Roberts, M. A., & Buscaglia, J. (2017). Factors associated with latent fingerprint exclusion determinations. Forensic Science International, 275, 65–75. [Google Scholar] [CrossRef] [PubMed]
  69. Vanderkolk, J. R. (2011). Chapter 9: Examination process. In E. H. Holder, L. O. Robinson, & J. H. Laub (Eds.), The fingerprint sourcebook (Chapters 9.1–9.26). Office of Justice Programs, National Institute of Justice, USA. Available online: https://www.ojp.gov/pdffiles1/nij/225320.pdf (accessed on 15 June 2025).
  70. Varga, D. (2022). No-reference image quality assessment with convolutional neural networks and decision fusion. Applied Sciences, 12(1), 101. [Google Scholar] [CrossRef]
  71. Wells, G. L. (1978). Applied eyewitness-testimony research: System variables and estimator variables. Journal of Personality and Social Psychology, 36(12), 1546–1557. [Google Scholar] [CrossRef]
  72. Wells, G. L. (2020). Psychological science on eyewitness identification and its impact on police practices and policies. American Psychologist, 75(9), 1316–1329. [Google Scholar] [CrossRef]
  73. Wells, G. L., Kovera, M. B., Douglass, A. B., Brewer, N., Meissner, C. A., & Wixted, J. T. (2020). Policy and procedure recommendations for the collection and preservation of eyewitness identification evidence. Law and Human Behavior, 44(1), 3–36. [Google Scholar] [CrossRef]
  74. Wells, G. L., Leippe, M. R., & Ostrom, T. M. (1979a). Guidelines for empirically assessing the fairness of a lineup. Law and Human Behavior, 3(4), 285–293. [Google Scholar] [CrossRef]
  75. Wells, G. L., Lindsay, R. C. L., & Ferguson, T. J. (1979b). Accuracy, confidence, and juror perceptions in eyewitness identification. Journal of Applied Psychology, 64(4), 440–448. [Google Scholar] [CrossRef]
  76. Wells, G. L., & Quigley-McBride, A. (2016). Applying eyewitness identification research to the legal system: A glance at where we have been and where we could go. Journal of Applied Research in Memory and Cognition, 5(3), 290–294. [Google Scholar] [CrossRef]
  77. Wells, G. L., Rydell, S. M., & Seelau, E. P. (1993). The selection of distractors for eyewitness lineups. Journal of Applied Psychology, 78(5), 835–844. [Google Scholar] [CrossRef]
  78. Wells, G. L., Small, M., Penrod, S., Malpass, R. S., Fulero, S. M., & Brimacombe, C. A. E. (1998). Eyewitness identification procedures: Recommendations for lineups and photospreads. Law and Human Behavior, 22(6), 603–647. [Google Scholar] [CrossRef]
  79. Wells, G. L., Steblay, N. K., & Dysart, J. E. (2015). Double-blind photo lineups using actual eyewitnesses: An experimental test of a sequential versus simultaneous lineup procedure. Law and Human Behavior, 39(1), 1–14. [Google Scholar] [CrossRef]
  80. Wilson, B. M., Harris, C. R., & Wixted, J. T. (2020). Science is not a signal detection problem. Proceedings of the National Academy of Science, 117(11), 5559–5567. [Google Scholar] [CrossRef] [PubMed]
  81. Wixted, J. T., & Wells, G. L. (2017). The relationship between eyewitness confidence and identification accuracy: A new synthesis. Psychological Science in the Public Interest, 18(1), 10–65. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Types of citations in the 1998 and 2020 Scientific Review Papers on eyewitness identification science.
Figure 1. Types of citations in the 1998 and 2020 Scientific Review Papers on eyewitness identification science.
Behavsci 15 01145 g001
Figure 2. Citations in the 2020 Scientific Review Paper on eyewitness identification science published pre- and post-1998.
Figure 2. Citations in the 2020 Scientific Review Paper on eyewitness identification science published pre- and post-1998.
Behavsci 15 01145 g002
Table 1. Comparison of the different elements of eyewitness identification procedures and latent fingerprint examination and the use of that evidence.
Table 1. Comparison of the different elements of eyewitness identification procedures and latent fingerprint examination and the use of that evidence.
Elements of ProceduresEyewitness IdentificationLatent Print Examination
Crime Scene Data?Memory of people present during or around the time of a crime event.Latent fingerprint collected in connection with a crime.
Recommended Method?Fair, double-blind lineup procedure with unbiased instructions.ACE-V and any local standard operating procedures.
Expertise Required?Yes—but laypersons are a type of face recognition/matching “expert”.Yes—special training and experience required.
Goal?Link a particular person to a crime event or criminal act.Link a particular person to a crime event or criminal act.
End-Users?Police, lawyers, people accused of crimes, judges, and jurors.Police, lawyers, people accused of crimes, judges, and jurors.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Quigley-McBride, A.; Blackall, T.L. On the Continuum of Foundational Validity: Lessons from Eyewitness Science for Latent Fingerprint Examination. Behav. Sci. 2025, 15, 1145. https://doi.org/10.3390/bs15091145

AMA Style

Quigley-McBride A, Blackall TL. On the Continuum of Foundational Validity: Lessons from Eyewitness Science for Latent Fingerprint Examination. Behavioral Sciences. 2025; 15(9):1145. https://doi.org/10.3390/bs15091145

Chicago/Turabian Style

Quigley-McBride, Adele, and T. L. Blackall. 2025. "On the Continuum of Foundational Validity: Lessons from Eyewitness Science for Latent Fingerprint Examination" Behavioral Sciences 15, no. 9: 1145. https://doi.org/10.3390/bs15091145

APA Style

Quigley-McBride, A., & Blackall, T. L. (2025). On the Continuum of Foundational Validity: Lessons from Eyewitness Science for Latent Fingerprint Examination. Behavioral Sciences, 15(9), 1145. https://doi.org/10.3390/bs15091145

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop