Identifying New Promising Research Directions with Open Peer Reviews and Contextual Top2Vec
Abstract
1. Introduction
- Scientometric analysis relies on accurate, manually curated international citation databases, which can be costly to maintain and use. However, these databases often fail to capture local trends and needs. For instance, research in region-specific agriculture may hold significant practical value, yet these related papers are frequently not indexed or well-cited in international databases.
- There is a substantial gap between publishing the first innovative results and presenting the first robust signals of a new frontier in citation databases. For example, the preprint of the Vaswani et al. “Attention is all you need” paper [3] was published first in June 2017 (Figure 1), but even in 2018, citation databases could not reliably detect this new frontier (because all citation information for 2018 was available only in 2019).
- We have proposed and tested an end-to-end conceptual framework that combines several ideas in review-based ranking and text mining: use topic-level assessment to tackle reviewing bias [4], apply pre-trained classifiers to extract relevant fragments from the reviews, and employ the classifiers to obtain unified review scores [5]; therefore, scores from different sources can be considered together. This framework builds and detects novel research directions on preprint databases with Contextual Top2Vec [6], summarizes them with a Transformer-based sequence-to-sequence model, and automatically ranks them based on the related open reviews with a Transformer-based encoder model. This integrated framework enables the early identification and ranking of novel research directions without relying on citation databases, reducing bias and accelerating discovery.
- We have validated this end-to-end framework through experiments on retrospective data from the International Conference on Learning Representations (ICLR 2017–2019) [4] and arXiv.org (2017–2021) [7]. Results show that high-ranked research directions identified by our method achieve significantly better citation performance than others.
2. Related Work
2.1. Detection of Promising Research Directions
2.2. Topical Vector Models
2.3. Topic Summarization
3. Materials and Methods
3.1. General Framework
- –
- We cannot rely on the explicit score given by the reviewer, because different conferences, journals, and research funds use different scoring systems.
- –
- This rough (“accept”/”borderline”/”decline”) scoring leads to lower classification error; in addition, combining these scores for a whole topic would provide more versatile information anyway.
3.2. Phrase Generation for the Contextual Top2Vec
3.3. Research Direction Summarization
- Get the most relevant direction’s documents with the Contextual Top2vec model. In our experiments, we obtained the 10 most relevant papers for each direction.
- From each document’s text extract title, select the first M = 5 sentences (they are likely to be from the document’s abstract), and top K = 5 (as they follow in the paper’s text) sentences containing top keywords or key phrases from the topic.
- Use the extracted sentences as a prompt and summarize each paper with an LLM.
- Concatenate all the paper summarizations and use it as a prompt to summarize the whole direction.

4. Datasets
5. Experiment Results
6. Discussion
7. Conclusions
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Du, S.; Sanmugam, M. A Bibliometric Review of Studies on the Application of Augmented Reality to Cultural Heritage by Using Biblioshiny and CiteSpace. In Embracing Cutting-Edge Technology in Modern Educational Settings; Chee, K.N., Sanmugam, M., Eds.; IGI Global: Hershey, PA, USA, 2023; pp. 184–213. [Google Scholar]
- Park, I.; Yoon, B. Identifying Promising Research Frontiers of Pattern Recognition through Bibliometric Analysis. Sustainability 2018, 10, 4055. [Google Scholar] [CrossRef]
- Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Polosukhin, I. Attention is all you need. In Proceedings of the 30th Advances in Neural Information Processing Systems, Long Beach, CA, USA, 7–9 December 2017. [Google Scholar]
- Wang, G.; Peng, Q.; Zhang, Y.; Zhang, M. What have we learned from OpenReview? World Wide Web 2023, 26, 683–708. [Google Scholar] [CrossRef]
- Kumar, P.B.; Ranjan, S.; Ghosal, T.; Agrawal, M.; Ekbal, A. PEERAssist: Leveraging on paper-review interactions to predict peer review decisions. In Proceedings of the 23rd International Conference Towards Open and Trustworthy Digital Societies, Virtual Event, 1–3 December 2021; pp. 421–435. [Google Scholar]
- Angelov, D.; Inkpen, D. Topic modeling: Contextual token embeddings are all you need. In Proceedings of the Findings of the Association for Computational Linguistics: EMNLP, Miami, FL, USA, 12–16 November 2024; pp. 13528–13539. [Google Scholar]
- ArXiv.org. Open-Access Archive. Available online: https://arxiv.org (accessed on 21 October 2025).
- Huang, Y.; Liu, H.; Pan, J. Identification of data mining research frontier based on conference papers. Int. J. Crowd Sci. 2021, 5, 143–153. [Google Scholar] [CrossRef]
- Ye, G.; Wang, C.; Wu, C.; Peng, Z.; Wei, J.; Song, X.; Tan, Q.; Wu, L. Research frontier detection and analysis based on research grants information: A case study on health informatics in the US. J. Informetr. 2023, 17, 101421. [Google Scholar] [CrossRef]
- Albrekht, V.; Mukhamediev, R.I.; Popova, Y.; Muhamedijeva, E.; Botaibekov, A. Top2Vec Topic Modeling to Analyze the Dynamics of Publication Activity Related to Environmental Monitoring Using Unmanned Aerial Vehicles. Publications 2025, 13, 15. [Google Scholar] [CrossRef]
- Abbas, K.; Hasan, M.K.; Abbasi, A.; Mokhtar, U.A.; Khan, A.; Abdullah, S.N.H.S.; Ahmed, F.R.A. Predicting the future popularity of academic publications using deep learning by considering it as temporal citation networks. IEEE Access 2023, 11, 83052–83068. [Google Scholar] [CrossRef]
- Teplitskiy, M.; Peng, H.; Blasco, A.; Lakhani, K.R. Is novel research worth doing? Evidence from peer review at 49 journals. Proc. Natl. Acad. Sci. USA 2022, 119, e2118046119. [Google Scholar] [CrossRef] [PubMed]
- Severin, A.; Strinzel, M.; Egger, M.; Barros, T.; Sokolov, A.; Mouatt, J.V.; Müller, S. Relationship between journal impact factor and the thoroughness and helpfulness of peer reviews. PLoS Biol. 2023, 21, e3002238. [Google Scholar] [CrossRef] [PubMed]
- Danielle, L.; Agha, L. Big names or big ideas: Do peer-review panels select the best science proposals? Science 2015, 348, 434–438. [Google Scholar] [CrossRef] [PubMed]
- Yang, C.; Han, J. Revisiting citation prediction with cluster-aware text-enhanced heterogeneous graph neural networks. In Proceedings of the IEEE 39th International Conference on Data Engineering (ICDE 2023), Anaheim, CA, USA, 3–7 April 2023; pp. 682–695. [Google Scholar]
- Staudinger, M.; Kusa, W.; Piroi, F.; Hanbury, A. An analysis of tasks and datasets in peer reviewing. In Proceedings of the Fourth Workshop on Scholarly Document Processing (SDP 2024), Bangkok, Thailand, 16 August 2024; pp. 257–268. [Google Scholar]
- Fernandes, G.L.; Vaz-de Melo, P.O.S. Enhancing the examination of obstacles in an automated peer review system. Int. J. Digit. Libr. 2023, 25, 341–364. [Google Scholar] [CrossRef]
- Kang, D.; Ammar, W.; Dalvi, B.; Van Zuylen, M.; Kohlmeier, S.; Hovy, E.; Schwartz, R. A dataset of peer reviews (PeerRead): Collection, insights and NLP applications. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long Papers), New Orleans, LA, USA, 1–6 June 2018; pp. 1647–1661. [Google Scholar]
- Ribeiro, A.C.; Sizo, A.; Cardoso, H.L.; Reis, L.P. Acceptance decision prediction in peer review through sentiment analysis. In Proceedings of the 20th EPIA Conference on Artificial Intelligence, Virtual Event, 7–9 September 2021; pp. 766–777. [Google Scholar]
- Plank, B.; Van Dalen, R. CiteTracked: A Longitudinal Dataset of Peer Reviews and Citations. In Proceedings of the Joint Workshop on Bibliometric-Enhanced Information Retrieval and Natural Language Processing for Digital Libraries (ACM BIRNDL SIGIR), Paris, France, 25 July 2019. [Google Scholar]
- Szumega, J.; Bougueroua, L.; Gkotse, B.; Jouvelot, P.; Ravotti, F. The open review-based (ORB) dataset: Towards automatic assessment of scientific papers and experiment proposals in high-energy physics. arXiv 2023, arXiv:2312.04576. [Google Scholar]
- Dieng, A.B.; Ruiz, F.J.; Blei, D.M. Topic modeling in embedding spaces. Trans. Assoc. Comput. Linguist. 2020, 8, 439–453. [Google Scholar] [CrossRef]
- Angelov, D. Top2vec: Distributed representations of topics. arXiv 2022, arXiv:2008.09470. [Google Scholar]
- Grootendorst, M. BERTopic: Neural topic modeling with a class-based TF-IDF procedure. arXiv 2022, arXiv:2203.05794. [Google Scholar]
- Devlin, J.; Chang, M.W.; Lee, K.; Toutanova, K. Bert: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), Minneapolis, MN, USA, 2–7 June 2019; pp. 4171–4186. [Google Scholar]
- Raffel, C.; Shazeer, N.; Roberts, A.; Lee, K.; Narang, S.; Matena, M.; Liu, P.J. Exploring the limits of transfer learning with a unified text-to-text transformer. J. Mach. Learn. Res. 2020, 21, 1–67. [Google Scholar]
- Lewis, M.; Liu, Y.; Goyal, N.; Ghazvininejad, M.; Mohamed, A.; Levy, O.; Zettlemoyer, L. BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Online, 5–10 July 2020; pp. 7871–7880. [Google Scholar]
- Ranganathan, J.; Abuka, G. Text Summarization using Transformer Model. In Proceedings of the Ninth International Conference on Social Networks Analysis, Management and Security (SNAMS), Milan, Italy, 29 November–1 December 2022. [Google Scholar] [CrossRef]
- Premnath, P.; Yenumulapalli, V.O.; Mohankumar, P.; Sivanaiah, R. TechSSN at SemEval-2025 Task 10: A Comparative Analysis of Transformer Models for Dominant Narrative-Based News Summarization. In Proceedings of the 19th International Workshop on Semantic Evaluation (SemEval-2025), Vienna, Austria, 31 July–1 August 2025; pp. 2205–2212. [Google Scholar]
- Zhang, T.; Kishore, V.; Wu, F.; Weinberger, K.Q.; Artzi, Y. BERTScore: Evaluating Text Generation with BERT. In Proceedings of the International Conference on Learning Representations (ICLR), Virtual Event, 26 April–1 May 2020. [Google Scholar]
- Gilhuly, C.; Shahzad, H. Consistency Evaluation of News Article Summaries Generated by Large (and Small) Language Models. arXiv 2025, arXiv:2502.20647. [Google Scholar] [CrossRef]
- Onan, A.; Alhumyani, H.A. FuzzyTP-BERT: Enhancing extractive text summarization with fuzzy topic modeling and transformer networks. J. King Saud Univ. Comput. Inf. Sci. 2024, 36, 102080. [Google Scholar] [CrossRef]
- Wang, Z.; Duan, Z.; Zhang, H.; Wang, C.; Tian, L.; Chen, B.; Zhou, M. Friendly topic assistant for transformer based abstractive summarization. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), Online, 16–20 November 2020; pp. 485–497. [Google Scholar]
- ICORE Conference Portal. Available online: https://portal.core.edu.au/conf-ranks (accessed on 10 October 2025).
- Conneau, A.; Khandelwal, K.; Goyal, N.; Chaudhary, V.; Wenzek, G.; Guzmán, F.; Stoyanov, V. Unsupervised Cross-lingual Representation Learning at Scale. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Online, 5–10 July 2020; pp. 8440–8451. [Google Scholar]
- COntext TOp2Vec Frontiers. Available online: https://github.com/masterdoors/COTOF/tree/main (accessed on 21 October 2025).
- Bouma, G. Normalized (pointwise) mutual information in collocation extraction. In Proceedings of the Biennial GSCL Conference; Gunter Narr Verlag: Tübingen, Germany, 2009; Volume 30, pp. 31–40. [Google Scholar]
- ISA-NLP Linguistic Parser. Available online: https://github.com/IINemo/isanlp (accessed on 10 October 2025).
- Novel or Not Dataset. Available online: https://github.com/masterdoors/COTOF/blob/main/notebooks/new_or_not_eng.ipynb (accessed on 21 October 2025).
- Summarization Dataset. Available online: https://github.com/masterdoors/COTOF/tree/main/datasets/summarization (accessed on 21 October 2025).
- OCTIS. Available online: https://github.com/MIND-Lab/OCTIS (accessed on 21 October 2025).
- Review Segmentation Model. Available online: https://huggingface.co/Ryzhik22/rev-classif-twolangs (accessed on 21 October 2025).
- Kirman, C.R.; Simon, T.W.; Hays, S.M. Science peer review for the 21st century: Assessing scientific consensus for decision-making while managing conflict of interests, reviewer and process bias. Regul. Toxicol. Pharmacol. 2019, 103, 73–85. [Google Scholar] [CrossRef] [PubMed]



| PMI | SYNTAX | ATTENTION | |||||||
|---|---|---|---|---|---|---|---|---|---|
| Dataset | 20NG | Yahoo | ICLR | 20NG | Yahoo | ICLR | 20NG | Yahoo | ICLR |
| Cnpmi_phrase | −0.311 | −0.263 | −0.492 | −0.304 | −0.259 | −0.487 | −0.307 | −0.259 | −0.492 |
| Cnpmi_words | 0.074 | 0.073 | −0.025 | 0.094 | 0.066 | 0.005 | 0.084 | 0.070 | −0.008 |
| Cv_phrase | 0.389 | 0.318 | 0.564 | 0.374 | 0.319 | 0.563 | 0.385 | 0.316 | 0.564 |
| Cv_words | 0.601 | 0.560 | 0.421 | 0.616 | 0.546 | 0.457 | 0.606 | 0.561 | 0.443 |
| Cwe_words | 0.107 | 0.079 | 0.071 | 0.108 | 0.074 | 0.078 | 0.110 | 0.073 | 0.072 |
| CSBERT | 0.385 | 0.365 | 0.413 | 0.392 | 0.360 | 0.408 | 0.385 | 0.364 | 0.402 |
| CBERTScore | 0.650 | 0.585 | 0.638 | 0.649 | 0.589 | 0.638 | 0.650 | 0.589 | 0.640 |
| Model | P | R | F1 |
|---|---|---|---|
| xlm-roberta-base | 0.74 ± 0.05 | 0.77 ± 0.04 | 0.74 ± 0.05 |
| xlm-roberta-large | 0.74 ± 0.02 | 0.78 ± 0.02 | 0.74 ± 0.03 |
| bert-base-multilingual-uncased | 0.70 ± 0.11 | 0.73 ± 0.11 | 0.71 ± 0.11 |
| Model | P | R | F1 |
|---|---|---|---|
| bert-base-uncased | 0.93 ± 0.12 | 0.93 ± 0.13 | 0.93 ± 0.12 |
| bert-large-uncased | 0.93 ± 0.12 | 0.93 ± 0.12 | 0.93 ± 0.12 |
| distilbert-base-uncased | 0.91 ± 0.07 | 0.90 ± 0.06 | 0.90 ± 0.06 |
| distilroberta-base | 0.73 ± 0.06 | 0.74 ± 0.05 | 0.73 ± 0.06 |
| roberta-base | 0.81 ± 0.03 | 0.79 ± 0.02 | 0.80 ± 0.02 |
| roberta-large | 0.90 ± 0.10 | 0.87 ± 0.10 | 0.87 ± 0.12 |
| xlm-roberta-base | 0.71 ± 0.36 | 0.74 ± 0.21 | 0.70 ± 0.29 |
| xlm-roberta-large | 0.64 ± 0.31 | 0.63 ± 0.16 | 0.56 ± 0.23 |
| Model | TEXT | TEXT + QUESTION TAG | TEXT + QUESTION TAG + IS NEW | ||||||
|---|---|---|---|---|---|---|---|---|---|
| P | R | F1 | P | R | F1 | P | R | F1 | |
| bert-base- uncased | 0.66 ± 0.03 | 0.64 ± 0.02 | 0.64 ± 0.03 | 0.59 ± 0.03 | 0.61 ± 0.03 | 0.59 ± 0.03 | 0.60 ± 0.05 | 0.62 ± 0.04 | 0.60 ± 0.05 |
| bert-large- uncased | 0.62 ± 0.04 | 0.62 ± 0.03 | 0.61 ± 0.04 | 0.60 ± 0.03 | 0.61 ± 0.02 | 0.59 ± 0.04 | 0.54 ± 0.14 | 0.54 ± 0.13 | 0.53 ± 0.15 |
| distilbert-base- uncased | 0.56 ± 0.01 | 0.58 ± 0.01 | 0.57 ± 0.01 | 0.62 ± 0.00 | 0.62 ± 0.00 | 0.62 ± 0.00 | 0.37 ± 0.01 | 0.46 ± 0.01 | 0.40 ± 0.01 |
| distilroberta-base | 0.57 ± 0.00 | 0.60 ± 0.00 | 0.57 ± 0.00 | 0.60 ± 0.04 | 0.60 ± 0.01 | 0.59 ± 0.04 | 0.41 ± 0.17 | 0.48 ± 0.11 | 0.44 ± 0.14 |
| roberta-base | 0.63 ± 0.01 | 0.62 ± 0.03 | 0.62 ± 0.02 | 0.66 ± 0.02 | 0.65 ± 0.02 | 0.65 ± 0.02 | 0.54 ± 0.15 | 0.53 ± 0.08 | 0.51 ± 0.11 |
| roberta-large | 0.63 ± 0.02 | 0.62 ± 0.02 | 0.62 ± 0.01 | 0.64 ± 0.02 | 0.58 ± 0.08 | 0.58 ± 0.07 | 0.61 ± 0.04 | 0.63 ± 0.02 | 0.61 ± 0.05 |
| xlm-roberta-base | 0.61 ± 0.02 | 0.61 ± 0.01 | 0.60 ± 0.01 | 0.28 ± 0.24 | 0.42 ± 0.15 | 0.32 ± 0.20 | 0.36 ± 0.01 | 0.46 ± 0.01 | 0.40 ± 0.01 |
| xlm-roberta-large | 0.42 ± 0.20 | 0.50 ± 0.10 | 0.44 ± 0.15 | 0.32 ± 0.10 | 0.37 ± 0.04 | 0.28 ± 0.06 | 0.49 ± 0.12 | 0.50 ± 0.05 | 0.47 ± 0.09 |
| Model_Name | Avg Ime per Batch | Rouge1 | Rouge2 | Rougel | Bleu4 | Bertscore |
|---|---|---|---|---|---|---|
| philschmid/bart-large-cnn-samsum | 2.43 ± 0.16 | 0.33 ± 0.02 | 0.08 ± 0.00 | 0.19 ± 0.00 | 0.02 ± 0.01 | 0.60 ± 0.00 |
| Talina06/arxiv-summarization | 3.03 ± 0.51 | 0.26 ± 0.03 | 0.05 ± 0.00 | 0.15 ± 0.01 | 0.02 ± 0.01 | 0.56 ± 0.00 |
| facebook/bart-large-cnn | 2.88 ± 0.66 | 0.33 ± 0.02 | 0.08 ± 0.00 | 0.18 ± 0.00 | 0.03 ± 0.01 | 0.60 ± 0.00 |
| google/flan-t5-large | 9.72 ± 1.46 | 0.22 ± 0.02 | 0.04 ± 0.00 | 0.13 ± 0.01 | 0.01 ± 0.00 | 0.52 ± 0.01 |
| google/pegasus-cnn_dailymail | 3.09 ± 0.30 | 0.31 ± 0.03 | 0.06 ± 0.00 | 0.17 ± 0.01 | 0.02 ± 0.01 | 0.60 ± 0.00 |
| google/pegasus-xsum | 2.55 ± 1.00 | 0.23 ± 0.09 | 0.04 ± 0.02 | 0.14 ± 0.03 | 0.01 ± 0.01 | 0.54 ± 0.01 |
| mistralai/Mistral-7B-Instruct-v0.2 | 36.77 ± 3.93 | 0.39 ± 0.03 | 0.10 ± 0.02 | 0.19 ± 0.02 | 0.04 ± 0.01 | 0.61 ± 0.02 |
| openlm-research/open_llama_13b | 45.20 ± 12.64 | 0.25 ± 0.03 | 0.03 ± 0.00 | 0.14 ± 0.01 | 0.01 ± 0.00 | 0.54 ± 0.01 |
| Falconsai/text_summarization | 3.71 ± 0.62 | 0.28 ± 0.05 | 0.05 ± 0.01 | 0.15 ± 0.01 | 0.02 ± 0.01 | 0.54 ± 0.01 |
| NousResearch/Nous-Hermes-13b | 30.32 ± 19.02 | 0.22 ± 0.04 | 0.03 ± 0.01 | 0.12 ± 0.02 | 0.01 ± 0.00 | 0.54 ± 0.01 |
| Qwen/Qwen1.5-7B-Chat | 15.60 ± 11.63 | 0.20 ± 0.06 | 0.02 ± 0.01 | 0.11 ± 0.02 | 0.01 ± 0.01 | 0.53 ± 0.04 |
| Conference | Original Scores | Predicted Scores | ||
|---|---|---|---|---|
| GR | Cites | GR | Cites | |
| ICLR 2017 | 86.0 (p-value 0.002) | 95.0 (p-value 0.0001) | 82.0 (p-value 0.007) | 85.0 (p-value 0.003) |
| ICLR 2018 | 45.0 (p-value 0.658) | 93.0 (p-value 0.0002) | 47.0 (p-value 0.602) | 75.0 (p-value 0.031) |
| ICLR 2019 | 80.0 (p-value 0.011) | 89.0 (p-value 0.001) | 79.0 (p-value 0.014) | 75.5 (p-value 0.031) |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Devyatkin, D.; Sochenkov, I.V.; Popov, D.; Zubarev, D.; Ryzhova, A.; Abanin, F.; Grigoriev, O. Identifying New Promising Research Directions with Open Peer Reviews and Contextual Top2Vec. Big Data Cogn. Comput. 2025, 9, 319. https://doi.org/10.3390/bdcc9120319
Devyatkin D, Sochenkov IV, Popov D, Zubarev D, Ryzhova A, Abanin F, Grigoriev O. Identifying New Promising Research Directions with Open Peer Reviews and Contextual Top2Vec. Big Data and Cognitive Computing. 2025; 9(12):319. https://doi.org/10.3390/bdcc9120319
Chicago/Turabian StyleDevyatkin, Dmitry, Ilya V. Sochenkov, Dmitrii Popov, Denis Zubarev, Anastasia Ryzhova, Fyodor Abanin, and Oleg Grigoriev. 2025. "Identifying New Promising Research Directions with Open Peer Reviews and Contextual Top2Vec" Big Data and Cognitive Computing 9, no. 12: 319. https://doi.org/10.3390/bdcc9120319
APA StyleDevyatkin, D., Sochenkov, I. V., Popov, D., Zubarev, D., Ryzhova, A., Abanin, F., & Grigoriev, O. (2025). Identifying New Promising Research Directions with Open Peer Reviews and Contextual Top2Vec. Big Data and Cognitive Computing, 9(12), 319. https://doi.org/10.3390/bdcc9120319

