Responsible Research Assessment and Research Information Management Systems

Definition: In the context of open science,


What Is Research Assessment?
Research assessment (RA) encompasses a multifaceted process aimed at evaluating the quality, impact, and effectiveness of research endeavors.As highlighted by Robert K. Merton [1], "The activities of scientists are subject to rigorous policing", indicating the inherent scrutiny applied to scientific endeavors.This scrutiny is manifest in the monitoring and evaluation of research performance, which permeates through individual, institutional, and governmental levels within the scientific community.RA is conducted by a variety of entities, each with their own motivations and objectives, such as academic institutions (promotion and tenure committees, departments), funding agencies (grant review panels), research councils (advisory committees, program managers), or government bodies (national assessment programs).Their main reasons are resource allocation, accountability, quality assurance, strategic planning on the institutional level, and recognition and reward of individual researchers.
At its core, RA serves to establish and uphold standards of research quality.It plays a pivotal role in shaping the allocation of resources by contributing to the efficient and accountable funding of projects, programs, and research teams.It affects the scholarly career of researchers in various ways, insofar that it forms the basis for decisions related to promotions and tenure in academic institutions and influences the likelihood of securing grants and funding (often based on number of publications, citation counts, h-index, and journal impact factors); a strong record in RA can enhance a researcher's reputation in their field, lead to invitations to speak at conferences, collaborate on projects, and join editorial boards, further enhancing a researcher's career prospects and competitivity for academic positions.Moreover, RA functions as a governance tool, fostering improvements in the quality of scholarship by promoting rigorous inquiry and adherence to established methodologies [2,3].
The landscape of RA is dynamic, continually evolving in response to the shifting demands of the scientific enterprise [4].Emerging technologies, such as artificial intelligence (AI) and machine learning, offer opportunities to enhance assessment methodologies by facilitating the analysis of vast datasets with greater precision and efficiency [5].Additionally, there is a continual refinement of metrics and indicators to encompass the diverse dimensions of research output and impact.Multidimensional assessment approaches have gained traction, recognizing the multifaceted nature of research performance and accounting for various aspects beyond traditional bibliometric measures [6].
Collaboration among stakeholders, including researchers, institutions, and funding agencies, is instrumental in advancing RA practices.By fostering synergies and sharing best practices, collaboration promotes methodological advancements and ensures that assessment frameworks align with the evolving needs and aspirations of the scientific community [7].
Recent discourse in the field underscores the importance of responsible research assessment (RRA), which advocates for transparency, diversity, quality, and open metrics [8].This paradigm shift acknowledges the diverse contributions to research and underscores the imperative of qualitative evaluation while advocating for the judicious use of quantitative indicators.By embracing RRA principles, the scientific community can mitigate the unintended consequences of metric-driven assessment practices and foster a culture of responsible and equitable evaluation [9].

What Are Research Information Management Systems?
To address the challenge of RA, universities, research-performing organizations (RPOs) and research funding organizations (RFOs) make use of a large diversity of research information management (RIM) infrastructures [10].Such systems collect and store research information, i.e., "information (. ..) relating to the conduct and communication of research", including bibliographic metadata (titles, abstracts, references, author data, affiliation data, and data on publication venues), metadata on research software, research data, samples, and instruments, information on funding and grants, and information on organizations and research contributors [11].
While the necessity for RIM systems is widespread across the academic landscape, the specific solutions implemented vary considerably [12].These solutions encompass a broad spectrum of tools and technologies, including but not limited to current research information systems (CRIS), contract management software, bespoke in-house software solutions, expert finder tools, and other specialized systems.Each of these solutions exhibits distinct characteristics in terms of format, data model, and terminology, reflecting the diverse needs and contexts within which they are deployed [13].
RIM systems play a crucial role in facilitating not only the documentation and tracking of research activities but also in enabling comprehensive and systematic evaluation processes.By providing a centralized repository for research-related metadata, these systems support decision-making processes related to resource allocation, performance assessment, and strategic planning within academic institutions and research organizations [14].Furthermore, they contribute to enhancing transparency, collaboration, and efficiency across the research ecosystem.
The continuous evolution and refinement of RIM systems reflect ongoing efforts to adapt to the evolving needs and challenges of RA in the context of open science and responsible research evaluation practices [15].As such, further research and development efforts are warranted to optimize the functionality, interoperability, and usability of these systems, ensuring that they remain robust and responsive tools for supporting the advancement of scholarly inquiry and knowledge dissemination.

Which Are the New Challenges?
Traditional RA methods, primarily reliant on bibliometrics and peer reviews, face increasing scrutiny for their narrow focus and potential to distort research priorities.In response, there is a growing movement towards RRA, emphasizing inclusivity, broader evaluation criteria, and adherence to open science practices.
For decades, research performance has been assessed essentially through bibliometrics (e.g., number of publications and citations) and peer reviews.Yet, RA has come under increasing scrutiny and criticism; the criticism focuses, among other things, on the choice of data sources, on the methods of calculating indicators, on the assessment procedures, and on underlying infrastructures [16].For instance, RA systems have been criticized as "too narrow in what they measure (. ..)Existing approaches favour individuals or teams that secure large grants, publish in journals with high impact factors-such as Nature-or register patents, at the expense of high-quality research that does not meet these criteria (. ..)" [17].Even if the "practice of university appointment procedures shows that easily measurable quantitative indicators continue to be prioritized (. ..) the strong weighting of indicators of pure quantity of research output is problematic in the assessment and prediction of excellent scientific performance due to their questionable validity" [18].Compared to quantitative indicators (such as the number of publications, the journal impact factor, or h-index), the assessment of the quality, rigor, reliability, robustness, transparency, and innovativeness of scientific work appears to be more valid as predictors of excellence and good research practice.

What Is Responsible Research Assessment?
Within the framework of open science, the San Francisco Declaration on Research Assessment [19], UNESCO [20], the European Commission [21], and, more recently, the Paris Call on Research Assessment [22], have put RA reform at the top of the agenda.The International Network of Research Management Societies developed a framework to ensure that research evaluation is meaningful, responsible, and effective [23], and authorities have begun to bring their RA systems into line with these emerging principles and goals [24].
In France, for instance, the Academy of Sciences requests an RRA based on clear, objective, transparent, and predetermined criteria [25], and the national Plan for Open Science calls for the promotion of open science and the diversity of scientific productions in RA, to foster the quality, effectiveness, and societal impact of research [26].Consequently, universities, RPOs, and RFOs are reevaluating their RA strategies to align with evolving priorities and societal demands.This entails a shift towards open-access publishing, transparency, fairness, inclusivity, and ethical conduct in research, as integral components of their open science initiatives.Other countries, such as the UK, The Netherlands, Finland, Spain, Italy, Germany, or Austria, are moving in the same direction.
Along with criticisms of traditional metrics such as journal impact factors and citation counts, other factors contribute to this shift, like the emergence of alternative metrics, the drive for fairness and inclusivity, the emphasis on dedicating more time to research rather than evaluation, the importance of sustaining long-term research endeavors, and the demand for broader evaluation criteria such as societal impact, public engagement, reproducibility, and adherence to open science practices, as well as contributions to policy or industry.In Europe, the Coalition for Advancing Research Assessment, supported by the European University Association, the European Science Foundation, and Science Europe, started efforts to shape and implement this transition towards RRA [8] Yet, RRA is not a clearly defined concept.It refers to the ethical and equitable evaluation of research outputs, such as publications, datasets, and other scholarly contributions.It involves assessing the quality, impact, and significance of research in a fair and transparent manner, while also considering the broader societal implications and ethical considerations.An editorial of the prestigious journal Nature defines RRA as more inclusive, less focused on journal-and publication-based metrics, avoiding using rankings of universities and research organizations, rewarding more-qualitative factors and open science, including data sharing and collaboration [17].
Among the key principles of RRA are fairness and equity (ensuring that all researchers, regardless of background or affiliation, have equal opportunities for evaluation and recognition), transparency (making the assessment process clear and understandable to all stakeholders, including researchers, funding agencies, and the public), diversity and inclusion (recognizing and valuing contributions from researchers representing diverse perspectives, disciplines, and methodologies), and openness (supporting open access to research outputs and data, facilitating reproducibility, and promoting knowledge sharing).Other, more generic principles are quality (evaluating research based on its scientific merit, methodological soundness, and adherence to ethical standards), and accountability (holding individuals and institutions accountable for their assessment practices and ensuring that they align with ethical standards and best practices).

What Does This Mean for Systems?
The move towards RRA presents both challenges and opportunities for RIM infrastructures, necessitating their adaptation to meet the evolving needs and expectations of the research community and to ensure that assessment systems minimize harm.Advancements in technology have facilitated the creation of novel tools and methodologies for research assessment, like for instance machine learning algorithms for text and data mining, semantic analysis, and network analysis.Such infrastructures enable more advanced and automated approaches to assess research outputs and their impact, as demonstrated by the Global Research Assessment Platform for Open Science (GraspOS) project whose purpose is to evaluate infrastructures used for RIM [27,28].
The influence of RRA on RIM systems extends beyond technological considerations, impacting the integration of diverse data sources, the improvement of (meta)data standards (including the establishment of reliable and sustainable persistent identifiers), interoperability, data exchange with other institutional and external systems, and the promotion of open science practices.
We already mentioned above that RRA initiatives tend to reject journal-based metrics, especially those produced by corporate companies, preferring altmetrics based on social media, novel metrics based on nonconventional research output (e.g., grey literature, software, or research data), and qualitative procedures (reviews, storytelling. ..).Moreover, such initiatives generally insist on community control over RA, no vendor lock-in, open infrastructures as common goods, open-source systems, and open data.
For the RA systems and infrastructures, this means more diversity, different processing, and a different kind of governance.Based on ongoing work of CoARA which launched working groups on responsible metrics and indicators and on open infrastructures [8], and the GraspOS project which published a landscape study on existing tools and services that can facilitate the implementation of open-science-aware RRA practices [29], we can add some more details.
More diversity of data: Compliance with RRA requires systems that are able to ingest and handle more and different kinds of data on research, such as research data, software, working papers, public and social media, AV material, but also narratives, data related to ethical performance, and so on.This means, too, that systems are able to deal with unusual data sources, a larger variety of metadata, and other identifiers.
More diversity of metrics: Compliance with RRA requires systems that are able to produce more and different kinds of metrics, including, above all, qualitative indicators of research impact (peer review), a responsible use of quantitative indicators, and communityled curation and annotation.Moreover, metrics should not only reflect research output but also research practice, like professional and integer conduct of research and openness, the reproducibility of results and the originality of ideas; research should be assessed on its own merits, especially results beyond the state of the art.
Enhanced Processing: Ensuring compliance with RRA mandates systems that facilitate transparent processing of research data, particularly in terms of data collection and analysis.This transparency is essential to enable scholars, researchers, and other scientific personnel to scrutinize and validate the outcomes.It entails fostering open and explicit communication regarding assessment criteria, implementing controls and restrictions on system usage and metrics, promoting interactivity, and advocating for open data practices.
Revised Governance Approach: Adherence to RRA necessitates systems characterized by increased community oversight.This entails stewardship by research communities, RPOs, and RFOs, along with the adoption of open-source systems and community-driven development methodologies.Avoiding vendor lock-in is imperative to ensure a certain level of sustainability.
In the same spirit, the Barcelona Declaration on Open Research Information [11] assumes that "openness of information about the conduct and communication of research must be the new norm" and requests openness as the default for research information, and systems, services, and infrastructures supporting and enabling such openness.
The Barcelona Declaration further emphasizes the importance of customization and flexibility within RIM infrastructures.As RRA practices vary significantly across nations, disciplines, institutions, and funding agencies, RIM systems must be adaptable and flexible to accommodate these differences.This adaptability entails the integration of configurable data models, reporting templates, and user interfaces.By offering customization options, RIM systems can better align with diverse assessment criteria and workflows, meeting the unique needs and preferences of users.
Governments, universities, RPOs, and RFOs have implemented a wide array of systems and infrastructures, each with distinct architectures, standards, data models, business models, and governance structures.Some infrastructures are cloud-based; some are commercial and proprietary software while others are open source.Consequently, their transition towards RRA compliance will be non-uniform, as there is no one-size-fits-all solution.
Another challenge is user engagement and training: As RRA practices evolve, RIM systems must actively engage users, i.e., researchers, administrators, funders, and policymakers, to ensure that their requirements and expectations are addressed.This entails providing training and support resources, gathering feedback through user surveys and focus groups, and fostering a community of practice centered on RA and information management.

Ethical Issues
In practice, the human element, particularly the ethical dimension, is often overlooked in RIM projects.But the digitization of society prompts us to contemplate both the individual and societal implications within a digital context, while also recognizing the unique considerations of digital ethics [30].This entails two approaches: firstly, applying general ethical principles to establish behavioral standards in the digital realm; and secondly, examining the digital domain itself and deriving moral insights from its distinct characteristics.In doing so, digital ethics acknowledges the evolving cultural and value dynamics shaped by emerging technologies [31].
Digital ethics is the heir to computer ethics and merges with information ethics [32].While this field of research encompasses a broad spectrum of research areas (such as medical ethics, journalistic ethics, AI ethics, data ethics, etc.), it can be seen as reflecting on a new ethics, proposing a macro-ethics [33], taking an interest in digital infrastructures, and questioning online conflicts.Digital ethics is concerned not only with the design of systems (ethics by design) but also with the appropriation of these infrastructures, and it raises the question of the distributed moral responsibility of ethical infrastructures (infra-ethics) [34].
Ethics challenges RIM systems in two ways [35].On the one hand, they should be able to represent research ethics as part of scientific practices and outcomes, on the individual as well as on the institutional level, and their data model should avoid biased terminology.On the other hand, they should be compliant with the principles and values of good scientific practice and with legal frameworks, especially regarding the accessibility and usage of sensitive data.As mentioned above, the guiding principles of open science add other ethical imperatives, such as transparency and openness.

Practical Issues
Despite an expanding body of literature on digital ethics, few studies have examined the ethical aspects of RIM systems.For this reason, RIM operators should collaborate with ethics committees or research integrity officers [36].
Implementing RIM systems involves complex technical requirements like data integration, interoperability, and scalability.Addressing these challenges requires robust technical solutions and expertise in system architecture and integration [37].
Managing research data within RIM systems requires effective data management frameworks to ensure data quality, security, and compliance with regulations such as GDPR [38].Establishing clear data ownership rights, strict access controls, and processes for managing the data lifecycle are crucial to minimize risks and ensure data integrity.
Ensuring high data quality is essential for the effectiveness and reliability of RIM systems [39].This includes a careful choice of indicators, with special attention to labeling, categories, and granularity, and the selection of reliable sources of information.Moreover, this includes measures for validating, cleansing, and updating research data to minimize errors and inconsistencies [40].Additionally, continuous monitoring and improvement of data quality are necessary to ensure that the provided information is trustworthy and meaningful [41].
User acceptance is a key factor in the success of RIM systems.By involving users in the development process, considering their feedback, and continuously adapting to their needs and requirements, user acceptance and usability can be enhanced [42].This includes, for instance, the design of user-friendly interfaces and workflows and the provision of comprehensive user training and support.

Future Directions
As part of their open science policies, RPOs and RFOs highlight the importance of openness, transparency, and integrity of research.In this environment and as part of science studies, future research on RA should focus on at least three issues: • The awareness of these organizations regarding RRA and ethics, as well as their readiness for potential changes, especially on the level of related infrastructures, procedures, and tools.• The assessment of responsible RIM, i.e., the analysis of the development, implementation, governance, and usage of RIM systems from an ethical point of view.This ethical investigation should include the stakeholders' "distributed responsibility" and the security of RIM data, in the context of the vertical integration of research infrastructures [43].• The convergence of RIM systems and institutional repositories, with new challenges regarding the system governance and openness and the quality and usage of research information and data [44].
Additionally, future studies could explore the implications of open science on interdisciplinary research collaboration, knowledge dissemination, and academic career advancement.Understanding how open science practices influence these aspects can provide valuable insights into shaping policies and practices that promote collaboration, transparency, and equity in the research ecosystem.
In the context of big data and AI, information systems should be beneficial, respectful of people and the environment, robust and secure; they should value human autonomy, promote fairness, and be explainable, accountable, and understandable [45].AI already starts to affect RA procedures and tools.Future research is needed to assess the impact of AI on RRA, especially regarding the sources and quality of research information and the transparency of algorithms, but also the governance, accountability, and openness of the systems.
. As of 18 April 2024, 723 organizations have signed the CoARA agreement, mainly universities and research centers, representing nearly 50 countries.The European Commission, as the most important RFO in Europe, signed the CoARA Agreement on Reforming Research Assessment in 2022 and endorsed the San Francisco Declaration on Research Assessment [19].