Understanding the Role of Objectivity in Machine Learning and Research Evaluation
Abstract
:1. Introduction
- What is knowledge?
- Can knowledge be acquired from data?
- What is a good teacher?
- How to distinguish true scientific theory from false one?
- How to form good inductive theories?
2. Brief Exposition on Objectivity
Longino’s Viewpoint on Dissent Bringing Objectivity
- Social Knowledge (Accept/Reject):Objectivity neither belongs to one single person nor can it be ensured by an individual: it is a community’s practice. Any knowledge conceived or presented, by an individual, moulds into social knowledge when it passes through evaluation by others. Such evaluations often improve and reshape the knowledge into even more valuable and productive work. Even a rejected hypothesis serves a role in future research, as the research community becomes more and more aware of the shortcomings of certain biased results as well.
- CriticismIt is crucial for objectivity to reduce the influence of subjective preference of individual background beliefs or assumptions. Although criticism may not completely eliminate the influence of subjective preferences, especially in fields like ML where certain parameters/settings have to be optimised and controlled, it provides the means for evaluating how much influence they have over the formation of the scientific contribution. Criticism should cover every relevant aspect of research.
- Shared StandardsRelevant criticism appeals to what is accepted by those concerned. Standards held on to by a community make members of such a community responsible for the goals of the community.
- Recognised AvenuesResearch communities form different platforms, based on collective interests, and these platforms are responsible for verifying and validating their members’ work. This helps to maintain worthy standards. Platforms, like public forums, conferences, peer-review journals, also compete among one another based on the credibility and social benefits they bring [11]. Part of the purpose of all these activities is to extract effective criticism from community members and prevent idiosyncratic values from shaping knowledge.
- Role of the CommunityThe role of the community is quite significant as they are the receiver of the outcome of research and may be affected by it. As difficult as determining community response may be due to its qualitative nature, some of the ways through which we can quantify the response and evaluate research may be through grants scored, number of publications, contribution to textbook contents and awards by a scientific avenue. Such quantitative measures provide relatively simple ways of comparing factors [21]. However, there are limitations to quantitative measures [22]. For example, the number of publications does not tell the whole story, as there are other important factors, like the quality of the medium/journal.
3. Methodology in ML
- Collect data and analyse,
- Split the data so as to have a final set to confirm predictions,
- Apply a suitable model to the data,
- Iterate over attributes of the model that gives best fit
- Then validate the model on the held-out split data
4. Discussion
Scrutinizing the Role of Bias against Objectivity
- Creative Bias:
- Qualitative Bias:
- Procedural Bias:This is a type of inductive bias that refers to the influence of the order of a set of steps defined in the presented model of any work [40].
- Ground Truth Bias:Ground truth refers to the held-out data that is often used as the basis for comparison in ML tasks. Custom tailoring of the designed model or even the training data to optimise results for a given ground truth is a bias that is simply unacceptable because of the potential dangers during application [41].
- Socially Ethical:Ethics plays a major role in scientific work and various guidelines for research ethics are specified by research communities [42].
- Credible References:
- Peer Review:
5. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Acknowledgments
Conflicts of Interest
Abbreviations
ANN | Artificial Neural Network |
BLEU | Bilingual Evaluation Understudy |
ML | Machine Learning |
NLP | Natural Language Processing |
References
- Reiss, J.; Sprenger, J. Scientific Objectivity. In The Stanford Encyclopedia of Philosophy, Winter 2017 ed.; Zalta, E.N., Ed.; Metaphysics Research Lab, Stanford University: Stanford, CA, USA, 2017. [Google Scholar]
- Parry, W.T.; Hacker, E.A. Aristotelian Logic; Suny Press: Albany, NY, USA, 1991. [Google Scholar]
- Cherkassky, V.; Mulier, F.M. Learning from Data: Concepts, Theory, and Methods; John Wiley & Sons: Hoboken, NJ, USA, 2007. [Google Scholar]
- Adewumi, T.P.; Liwicki, F.; Liwicki, M. Conversational Systems in Machine Learning from the Point of View of the Philosophy of Science—Using Alime Chat and Related Studies. Philosophies 2019, 4, 41. [Google Scholar] [CrossRef] [Green Version]
- Stevens, E.; Antiga, L. Deep Learning with PyTorch: Essential Excerpts; Manning Publications: Shelter Island, NY, USA, 2019. [Google Scholar]
- Thagard, P. Philosophy and machine learning. Can. J. Philos. 1990, 20, 261–276. [Google Scholar] [CrossRef]
- Creath, R. Logical Empiricism. In The Stanford Encyclopedia of Philosophy, Summer 2020 ed.; Zalta, E.N., Ed.; Metaphysics Research Lab, Stanford University: Stanford, CA, USA, 2020. [Google Scholar]
- Hartman, J. Nature and Naturalism: A Philosophy Primer; lulu.com: Morrisville, NC, USA, 2013. [Google Scholar]
- Kazmier, L.J. Schaum’s Outline of Theory and Problems of Business Statistics; McGraw-Hill: New York, NY, USA, 1976. [Google Scholar]
- Cherkassky, V.; Friedman, J.H.; Wechsler, H. From Statistics to Neural Networks: Theory and Pattern Recognition Applications; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2012; Volume 136. [Google Scholar]
- Longino, H.E. Science as Social Knowledge: Values and Objectivity in Scientific Inquiry; Princeton University Press: Princeton, NJ, USA, 1990. [Google Scholar]
- Shamoun, D.; Calabrese, E. On Objective Risk. 2015. Available online: https://www.mercatus.org/system/files/Shamoun-On-Objective-Risk_0.pdf (accessed on 15 March 2021).
- Daston, L.; Galison, P. The image of objectivity. Representations 1992, 40, 81–128. [Google Scholar] [CrossRef]
- Zhang, Y.; Sun, S.; Galley, M.; Chen, Y.C.; Brockett, C.; Gao, X.; Gao, J.; Liu, J.; Dolan, B. Dialogpt: Large-scale generative pre-training for conversational response generation. arXiv 2019, arXiv:1911.00536. [Google Scholar]
- Haraway, D. Situated knowledges: The science question in feminism and the privilege of partial perspective. Fem. Stud. 1988, 14, 575–599. [Google Scholar] [CrossRef]
- White, M.D. 40 Immanuel Kant. In Handbook of Economics and Ethics; Edward Elgar Pub: Cheltenham, UK, 2009; p. 301. ISBN 9780198793991. [Google Scholar]
- Shamoo, A.E.; Resnik, D.B. Responsible Conduct of Research; Oxford University Press: Oxford, UK, 2009. [Google Scholar]
- Jukola, S. Longino’s theory of objectivity and commercialized research. In Empirical Philosophy of Science; Springer: Berlin, Germany, 2015; pp. 127–143. [Google Scholar]
- Longino, H.E. The Fate of Knowledge; Princeton University Press: Princeton, NJ, USA, 2002. [Google Scholar]
- Segerstrale, U.; Longino, H. Book review: Science as Social Knowledge: Values and Objectivity in Scientific Inquiry. Contemp. Sociol. 1992, 21, 138. [Google Scholar] [CrossRef]
- Bornmann, L.; Mutz, R. Growth rates of modern science: A bibliometric analysis based on the number of publications and cited references. J. Assoc. Inf. Sci. Technol. 2015, 66, 2215–2222. [Google Scholar] [CrossRef] [Green Version]
- Zahedi, Z.; Costas, R.; Wouters, P. How well developed are altmetrics? A cross-disciplinary analysis of the presence of ‘alternative metrics’ in scientific publications. Scientometrics 2014, 101, 1491–1513. [Google Scholar] [CrossRef] [Green Version]
- Hagan, M.T.; Demuth, H.B.; Beale, M.; De Jesus, O. Neural Network Design, 2nd ed.; PWS Publishing Co.: Boston, MA, USA, 2000. [Google Scholar]
- Papineni, K.; Roukos, S.; Ward, T.; Zhu, W.J. BLEU: A method for automatic evaluation of machine translation. In Proceedings of the 40th Annual Meeting of the Association for Computational Linguistics, Stroudsburg, PA, USA, 7–12 July 2002; pp. 311–318. [Google Scholar]
- Indurkhya, N.; Damerau, F.J. Handbook of Natural Language Processing; CRC Press: Boca Raton, FL, USA, 2010; Volume 2. [Google Scholar]
- Reimers, N.; Gurevych, I. Why comparing single performance scores does not allow to draw conclusions about machine learning approaches. arXiv 2018, arXiv:1803.09578. [Google Scholar]
- Douglas, C.M. Design and Analysis of Experiments; Wiley: Hoboken, NJ, USA, 2001; ISBN 978-1-119-63542-0. [Google Scholar]
- McDermott, M.; Wang, S.; Marinsek, N.; Ranganath, R.; Ghassemi, M.; Foschini, L. Reproducibility in machine learning for health. arXiv 2019, arXiv:1907.01463. [Google Scholar]
- Olorisade, B.K.; Brereton, P.; Andras, P. Reproducibility in Machine Learning-Based Studies: An Example of Text Mining. In Proceedings of the MLWorkshop at the 34th International Conference on Machine Learning (ICML2017), Sydney, Australia, 6–11 August 2017; Available online: https://www.semanticscholar.org/paper/Reproducibility-in-Machine-Learning-Based-Studies%3A-Olorisade-Brereton/2502db705985d6736b1cc9c8e3f13f617a6a9a98 (accessed on 15 March 2021).
- Musgrave, K.; Belongie, S.; Lim, S.N. A metric learning reality check. arXiv 2020, arXiv:2003.08505. [Google Scholar]
- Adewumi, T.P.; Liwicki, F.; Liwicki, M. Exploring Swedish & English fastText Embeddings with the Transformer. arXiv 2020, arXiv:2007.16007. [Google Scholar]
- Mikolov, T.; Chen, K.; Corrado, G.; Dean, J. Efficient estimation of word representations in vector space. arXiv 2013, arXiv:1301.3781. [Google Scholar]
- Dictionary, C. Cambridge advanced learner’s dictionary. In PONS-Worterbucher, Klett Ernst Verlag GmbH; Cambridge University Press: Cambridge, UK, 2008. [Google Scholar]
- Blackwell, S.E.; Woud, M.L.; MacLeod, C. A question of control? Examining the role of control conditions in experimental psychopathology using the example of cognitive bias modification research. Span. J. Psychol. 2017, 20, E54. [Google Scholar] [CrossRef] [Green Version]
- Lavesson, N.; Davidsson, P. Evaluating learning algorithms and classifiers. Int. J. Intell. Inf. Database Syst. 2007, 1, 37–52. [Google Scholar] [CrossRef]
- de Jesus, S.N.; Rus, C.L.; Lens, W.; Imaginário, S. Intrinsic motivation and creativity related to product: A meta-analysis of the studies published between 1990–2010. Creat. Res. J. 2013, 25, 80–84. [Google Scholar] [CrossRef]
- Hennessey, B.A. The Creativity—Motivation Connection; Cambridge University Press: Cambridge, UK, 2010. [Google Scholar]
- Morse, J.M. Critical analysis of strategies for determining rigor in qualitative inquiry. Qual. Health Res. 2015, 25, 1212–1222. [Google Scholar] [CrossRef]
- Onwuegbuzie, A.J.; Leech, N.L. Validity and qualitative research: An oxymoron? Qual. Quant. 2007, 41, 233–249. [Google Scholar] [CrossRef]
- Gordon, D.F.; Desjardins, M. Evaluation and selection of biases in machine learning. Mach. Learn. 1995, 20, 5–22. [Google Scholar] [CrossRef] [Green Version]
- Lei, Y.; Bezdek, J.C.; Romano, S.; Vinh, N.X.; Chan, J.; Bailey, J. Ground truth bias in external cluster validity indices. Pattern Recognit. 2017, 65, 58–70. [Google Scholar] [CrossRef] [Green Version]
- Adler, E.; Clark, R. How It’s Done: An Invitation to Social Research; Cengage Learning: Boston, MA, USA, 2007. [Google Scholar]
- Rubin, H.R.; Redelmeier, D.A.; Wu, A.W.; Steinberg, E.P. How reliable is peer review of scientific abstracts? J. Gen. Intern. Med. 1993, 8, 255–258. [Google Scholar] [CrossRef] [PubMed]
- Ware, M. Peer Review: Benefits, Perceptions and Alternatives; Citeseer: Princeton, NJ, USA, 2008. [Google Scholar]
- Balci, O. Verification, validation, and accreditation. In Proceedings of the 1998 Winter Simulation Conference, Proceedings (Cat. No.98CH36274), Washington, DC, USA, 13–16 December 1998; Volume 1, pp. 41–48. [Google Scholar]
- Anney, V.N. Ensuring the quality of the findings of qualitative research: Looking at trustworthiness criteria. J. Emerg. Trends Educ. Res. Policy Stud. 2014, 5, 272–281. [Google Scholar]
Focus on Science As Practice | |
Refocus on Scientific Method Practiced by Social Group |
Degree | |||||
---|---|---|---|---|---|
Evaluation Key | 1—Least Applicable | 2 | 3 | 4 | 5—Strongly Applicable |
Socially Ethical | |||||
Credible References |
Degree | |||||
---|---|---|---|---|---|
Evaluation Key | 1—Least Applicable | 2 | 3 | 4 | 5—Strongly Applicable |
Creative Bias | |||||
Qualitative Bias |
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2021 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
Share and Cite
Javed, S.; Adewumi, T.P.; Liwicki, F.S.; Liwicki, M. Understanding the Role of Objectivity in Machine Learning and Research Evaluation. Philosophies 2021, 6, 22. https://doi.org/10.3390/philosophies6010022
Javed S, Adewumi TP, Liwicki FS, Liwicki M. Understanding the Role of Objectivity in Machine Learning and Research Evaluation. Philosophies. 2021; 6(1):22. https://doi.org/10.3390/philosophies6010022
Chicago/Turabian StyleJaved, Saleha, Tosin P. Adewumi, Foteini Simistira Liwicki, and Marcus Liwicki. 2021. "Understanding the Role of Objectivity in Machine Learning and Research Evaluation" Philosophies 6, no. 1: 22. https://doi.org/10.3390/philosophies6010022
APA StyleJaved, S., Adewumi, T. P., Liwicki, F. S., & Liwicki, M. (2021). Understanding the Role of Objectivity in Machine Learning and Research Evaluation. Philosophies, 6(1), 22. https://doi.org/10.3390/philosophies6010022