Next Article in Journal
Exploring Evaluation Methods for Interpretable Machine Learning: A Survey
Previous Article in Journal
Cyberbullying Detection on Social Media Using Stacking Ensemble Learning and Enhanced BERT
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Job Vacancy Ranking with Sentence Embeddings, Keywords, and Named Entities

Department of Software Engineering, Shamoon College of Engineering, Beer Sheva 84500, Israel
*
Author to whom correspondence should be addressed.
Information 2023, 14(8), 468; https://doi.org/10.3390/info14080468
Submission received: 16 June 2023 / Revised: 6 August 2023 / Accepted: 15 August 2023 / Published: 20 August 2023

Abstract

:
Resume matching is the process of comparing a candidate’s curriculum vitae (CV) or resume with a job description or a set of employment requirements. The objective of this procedure is to assess the degree to which a candidate’s skills, qualifications, experience, and other relevant attributes align with the demands of the position. Some employment courses guide applicants in identifying the key requirements within a job description and tailoring their experience to highlight these aspects. Conversely, human resources (HR) specialists are trained to extract critical information from numerous submitted resumes to identify the most suitable candidate for their organization. An automated system is typically employed to compare the text of resumes with job vacancies, providing a score or ranking to indicate the level of similarity between the two. However, this process can become time-consuming when dealing with a large number of applicants and lengthy vacancy descriptions. In this paper, we present a dataset consisting of resumes of software developers extracted from a public Telegram channel dedicated to Israeli hi-tech job applications. Additionally, we propose a natural language processing (NLP)-based approach that leverages neural sentence representations, keywords, and named entities to achieve state-of-the-art performance in resume matching. We evaluate our approach using both human and automatic annotations and demonstrate its superiority over the leading resume–vacancy matching algorithm.

1. Introduction

The resume–vacancy matching problem refers to the task of automatically matching job seekers’ resumes or CVs with job vacancies or job descriptions [1]. Its goal is to determine the degree of compatibility between a candidate’s skills, qualifications, and experience and the requirements and preferences specified by the employer in the vacancy description. The automatization of a resume–vacancy ranking can help streamline the selection process, save time for both job seekers and employers, and improve the overall efficiency of the job market [2].
This task is not straightforward but has many different aspects or facets to consider because it is a complex process that requires careful attention to various factors. It involves comparing and matching different elements from the candidate’s resume with the relevant parts of the job description. These elements usually include skills, experience, and keywords. Skills that the candidate possesses should align with the skills required for the job, and the candidate’s past work experience must be assessed to see if it matches the vacancy experience requirements. In many cases, job descriptions include specific keywords that indicate essential qualifications. Resume–vacancy matching usually requires identifying these keywords in the resume and seeing if they match the ones listed in the job description. However, in most cases, a perfect match may not be possible. Some aspects of the candidate’s qualifications might align with the job description, while others may not be an exact fit. Beyond simple matching, advanced techniques need to be applied to understand the semantics and context of the job requirements and the candidate’s qualifications [3].
To address the resume–vacancy ranking problem, various automated techniques have been applied in recent years, including natural language processing, machine learning, and information retrieval. The key issues for these systems are that resumes come in a variety of formats, making it difficult to accurately extract relevant information. Furthermore, the job market moves quickly, and matching systems must react to changing job criteria and skill expectations.
Rule-based systems are relatively easy to develop and implement because they rely on predefined rules and keywords to match resumes with job vacancies. Recruiters can tailor these algorithms to specific job requirements, resulting in more accurate matches for their company. On the other hand, rule-based systems lack the agility to handle more nuanced matching circumstances and may struggle with unusual or fast-changing work needs. If applicants use different languages or synonyms for the same talents, over-reliance on keywords may result in false positives or mismatches. For example, Resumix [4] is one of the earliest systems that used rule-based methods to match candidate resumes with job vacancies based on predefined keywords and criteria. Works [5,6] use taxonomies and ontologies as the means of defining rules for resume–vacancy matching.
NLP-based systems leverage natural language processing techniques to understand the context and semantics of resumes and job descriptions, leading to more accurate matches. They can identify related skills and experiences even when specific keywords are not explicitly mentioned, making them more robust in matching candidates. However, biased training data can influence these systems, perhaps discriminating against specific groups or preferring applicants with certain attributes. The development of these systems necessitates substantial computer resources, and fine tuning the models can be time-consuming. For example, paper [7] represents text data with the help of word embeddings and the bag-of-words (BOW) model, while the authors of [8,9] used the BOW model with term frequency–inverse document frequency (TF-IDF) weights [10]. Paper [11] represents text with word n-gram vectors [12].In general, all these methods use vector similarity for the chosen text representation to rank resumes or vacancies. Different similarity measures appear in these works, such as cosine similarity, Dice similarity [13], Jaccard similarity, and Jaccard overlap. Several advanced approaches use recurrent neural networks (RNNs) or convolutional neural networks (CNNs) to operate as feature extractors and a word embedding layer [14] to represent texts [15,16]. Our proposed method belongs to this category of resume-matching systems.
The next class of resume-matching systems is the general machine learning models. As more data are processed and input from recruiters is received, these models can continuously improve their matching accuracy. These systems can adjust to individual recruiters’ or organizations’ preferences, adapting matches to their own requirements. However, in order to work well, these systems require a vast amount of data, creating issues about data privacy and security. Some of the algorithms work as black boxes, making it difficult to comprehend how particular matches are formed, thus leading to biases. Examples of such systems include [17,18], which employ neural networks to create end-to-end matching models. Multiple works, such as [19,20,21,22], treat the task of resume matching as a recommender system that suggests the best vacancies for a given resume. The authors of [23] utilize historical data to facilitate the matching.
OKAPI BM25 [24] is a ranking algorithm used in information retrieval systems to rank documents based on their relevance to a given query. It is a good baseline for resume rating tests [25] and is one of the most accurate computer algorithms utilizing a bag-of-words paradigm [26].
Extractive summarization is an NLP task whose objective is to generate a summary of a given text by selecting and combining important sentences or phrases from the original document [27]. The summary is created by extracting sentences that are considered the most informative of the original content. Various statistical and machine learning techniques have been developed over the years; good surveys of the field are given in [27,28]. Modern top-line techniques for extractive summarization rely on transformers [29,30].
Keywords are specific words or phrases that carry significant meaning or relevance within a particular context or subject matter. In the context of resume matching or text analysis, keywords are often used to represent essential skills, qualifications, experience, or specific terms related to a job or topic. When analyzing text, including resumes or job vacancies, keywords play a vital role in identifying and matching relevant information.
The field of keyword extraction focuses on the task of automatically identifying and extracting important keywords or key phrases from a given text. This process plays a crucial role in various natural language processing (NLP) applications, such as information retrieval, document summarization, text classification, and topic modeling. The methods for keyword extraction range from tf-idf-based [31,32] to graph-based, such as TextRank [33], to co-occurrence-based methods, such as RAKE [34], YAKE [35], and others [36,37]. Finally, neural methods have taken center stage during the last years [38,39].
A named entity refers to a specific real-world object or concept that has a proper name or designation. It can be an individual, organization, location, date, time, product, or any other entity that is referred to by a particular name or label. Recognizing and extracting named entities from text is an important task in natural language processing and information retrieval (IR). Methods for named entity recognition (NER) use probabilistic models such as conditional random fields [40,41,42], recurrent neural networks [43], transformers [44,45], graph neural networks [46,47], and transfer learning [48].
In this paper, we propose a method for efficient resume–vacancy matching by using summarization and enhancing the produced summaries by keywords or named entities found in texts. We use statistical and semantic vector representations for the enhanced texts and compute their similarity to produce a vacancy ranking for every resume. The method is unsupervised and does not require training. We perform an extensive experimental evaluation of our method to show its validity.
We address the following research questions in this paper:
  • RQ1: does text enhancement with keywords or named entities improve ranking results?
  • RQ2: does a summarization of either resumes or vacancies improve ranking results?
  • RQ3: what text representation produces the best results?
  • RQ4: can our method distinguish between vacancies relevant and irrelevant for an applicant?
This paper is organized as follows: Section 2 describes the details of our method—text processing and text types, text enhancement with keywords and named entities, and text representations. In Section 3, we describe the baselines to which we compare our method. Section 4 describes how our datasets were created, processed, and annotated. Metrics used for experimental evaluations and the rationale behind them are reported in Section 5. Section 6 provides full details of the experimental evaluation. Finally, Section 8 contains conclusions drawn from the evaluation.

2. Method

Our method is based on vector similarities and is therefore called Vector Matching (VM) method. We provide its outline and details in the subsections below.

2.1. The Outline

As a first step, we select one of the following text types for resumes and vacancies:
  • Full texts;
  • Summarized texts (summaries);
  • Summaries enhanced by keywords;
  • Summaries enhanced by named entities.
The extraction of summaries, keywords, and name entities is described below in Section 2.2.
In the second step, we represent the text data as numeric vectors with one of the methods described below in Section 2.3. Thus, for a resume, R, represented and a vacancy, V, we have vectors V resume and V vacancy . Then, we compute an L1 distance between a resume and a vacancy vector of length n:
L 1 ( V resume , V vacancy ) = i = 1 n V resume [ i ] V vacancy [ i ]
The resulting ranking is inferred based on these distances, arranged in ascending order from the smallest to the largest. A pipeline of our approach is given in Figure 1.

2.2. Text Types

In this section, we describe the text types that our ranking method uses. An overview of data representations is provided in Figure 2, and a detailed description is given in the sections below.

2.2.1. Full Texts

As a first option, the VM method uses full texts of resumes and vacancies after basic preprocessing that includes removing numbers and converting them to lowercase.

2.2.2. Extractive Summaries

We use BERT-based extractive summarization [49] with a pre-trained English model bert-base-uncased to obtain summaries for vacancies and resumes. The method performs tokenization and then encodes tokens and feeds into a transformer that predicts the saliency scores of each sentence. The saliency scores indicate the importance or relevance of each sentence in the context of the overall document. We set the maximum size of a summary to 10 sentences, and the ELBOW method of BERT summarizer is used to determine the optimal summary length. The final summaries are computed from preprocessed full texts and contain at most 10 sentences.

2.2.3. Keyword-Enhanced Summaries

In the context of resume-matching systems, keywords from job vacancies are often compared with keywords extracted from resumes to determine the level of alignment between the candidate’s profile and the job requirements. The presence or absence of specific keywords can be used as a signal to assess the suitability of a candidate for a particular position [50].
To build keyword-enhanced summaries, we attach the keywords extracted from full texts to the summaries and pass them on to the text representation module. We do it because the extracted keywords may not be contained in a summary.
KeyBERT [39] is a method for keyword extraction that utilizes a BERT transformer-based language model to generate contextually relevant keywords. It aims to capture the semantic meaning of the text and identify key phrases that best represent the content. It employs a two-step procedure of embedding generation and keyword selection with a maximal marginal relevance (MMR) algorithm to select the most informative keywords from the generated embeddings. KeyBERT is a flexible method that can extract both single-word and multi-word keywords, and its implementation is available at https://github.com/MaartenGr/KeyBERT (accessed on 1 June 2023).
We applied KeyBERT and extracted multi-word keywords from resumes and vacancies with word numbers in the range [ 1 , 3 ] . Examples of keywords for both types of data appear in Table 1, together with their KeyBERT scores.

2.2.4. Named-Entity-Enhanced Summaries

To build this text representation, we append the NEs extracted from full texts to the summaries and pass them on to the text representation module. The details of NE extraction are provided below.
We have used the spaCy SW package [51] to perform the NE extraction to preprocessed resumes and vacancies. By leveraging a pre-trained model, spaCy’s named entity recognition (NER) module is capable of accurately recognizing and extracting named entities from text. The main types of spaCy NE are described in the Appendix A. We use the en_core_web_sm that is trained in English and, therefore, suits our data. Examples of NEs extracted from the data appear in Appendix A.
To construct data representations, we append named entities to the texts, whether in their complete form or as summaries, and then feed them into the text representation module.

2.3. Text Representation

Our vector matching method uses the following text representations:
  • Word n-grams with 1 n 3 . A word n-gram is a contiguous sequence of n words within a text or a sentence.
  • Character n-grams with 1 n 3 . A character n-gram is a contiguous sequence of n characters within a text.
  • BOW vectors with TF-IDF weights, where every resume and vacancy is treated as a separate document. A BOW vector is a numerical representation of a text document that captures the occurrence or frequency of words in the document.
  • Bidirectional encoder representations from transformers (BERT) sentence embeddings obtained from a pre-trained English BERT model “bert-base-uncased” [52]; this model was selected because all the resumes are written in English.
These representations are computed for every one of the text data forms described in Section 2.2.

3. Baselines

3.1. The OKAPI Algorithm

The Okapi BM25 (Best Matching 25) algorithm [53] is a ranking function commonly used in information retrieval systems to calculate the relevance of a document to a given query. It is an improved version of the original Okapi BM11 algorithm [24].
The algorithm takes into account factors such as term frequency, document length, and the overall distribution of terms in a collection of documents. It aims to estimate the probability that a document is relevant to a particular query. The algorithm performs tokenization and computes a score for a document–query pair using the following components: term frequency and inverse document frequency for every term (token), document length, query length, average document length, and two free parameters, k 1 and b, which control the impact of term frequency and document length normalization, respectively. The OKAPI-BM25 algorithm was shown to be a useful baseline for experiments and features for ranking [26].

3.2. BERT-Based Ranking

We use the method introduced in [54] as a neural baseline, which we denote by BERT-rank. This method applies a knowledge distillation approach that transfers knowledge from a larger and more accurate teacher model to a smaller student model. The knowledge transfer is performed across different neural architectures, allowing the student model to benefit from the teacher model’s insights. The proposed method employs multiple techniques, including soft target training, attention alignment, and feature distillation, to effectively transfer knowledge.

4. The Data

4.1. Vacancies Dataset

We downloaded job postings titled “software developer” in the United States during August 2019 [55]. These data were extracted using JobsPikr—a job data delivery platform that extracts job data from job boards across the globe [56]. JobsPikr crawls over 70,000 employer websites daily to process over one million job postings; its API provides access to a database of over 500 million jobs over the past 3 years [57]. The data were downloaded in .csv format, and they contain 10,000 vacancies with a combined size of 41.51 MB. The following data issues were found: (1) the same vacancy can appear more than once with a different unique id, and (2) portions of text in the job description field are repeated in many cases.
An example of a vacancy and dataset statistics are provided in Appendix A.
We extracted the names of vacancies using the “find-job-titles” library [58]. To find vacancies relevant to our applicants we select those that include the word “developer” or “full-stack” from the entire array of vacancies. Table 2 describes the column and context statistics of the data, providing the minimum, maximum, and average character length for every field, data type, and the number of non-empty entries for each field. Table 3 gives an example of a relevant vacancy as it appears in the vacancies dataset.
To find vacancies that are not relevant to our applicants, we downloaded them with JobsPikr [56] from the “10,000-data-scientist-job-postings-from-the-usa” directory (a partial example of such a vacancy is given in Table 4). Table 5 shows examples of NE and keywords that are extracted from this type of vacancy; while not all vacancies in this set are for data scientists, they are intended for people looking for senior positions only.

4.2. Resume Dataset

As we did not succeed in finding a good-quality resume dataset, we constructed our own by downloading real resumes from the Telegram [59] group “HighTech Israel Jobs” [60], which freely publishes resumes and vacancies.
All resumes in the group are written in a free format. Most often they consist of one page, but some have two or more pages. There are no clear structures, marks, or mandatory sections. In the original dataset, there are summaries in Russian and English; it happens that one person uploads several versions of their resume. The following data preprocessing steps were made:
  • Exclusion of duplicate records from the dataset.
We use the “hashlib” library [61] that determines how many bytes should be checked. Then, the function recursively checks all the specified paths. The output is duplicated addresses.
2.
Exclusion of resumes not written in English.
We use the “detect” function of the “langdetect” library [62]. The contents of the files were extracted and checked. This function returns the name of the language of the text. If it returned “en”, which is English, then we accept this file; otherwise, we delete it.
3.
Exclusion of resumes of non-developers.
More than 90% of the resumes were eliminated at the third stage because the dataset contained a lot of resumes of testers, project managers, etc. We used a custom-made list of keywords to search for in candidates’ resumes using the “flashtext” library [63]. A sample of the list containing job titles for Oracle developers is provided in Table 6.
4.
Anonymization.
Even though the resumes are published in an open Telegram group, we masked the names, phone numbers, and addresses of applicants.
Figure 3 gives an example of a resume from this dataset.

4.3. Human Annotation

Human rankings are essential for evaluating resume ranking systems because they provide a crucial benchmark for measuring the effectiveness and accuracy of automated resume–vacancy matching methods. To produce these rankings, we use the following pipeline:
  • First, we selected a subset of the resume dataset containing 30 resumes (vacancy rankings are time-consuming, and the availability of our annotators was limited).
  • Then, we selected a random subset of the vacancies dataset containing five relevant vacancies.
  • Two annotators with a computer science background were asked to rank the vacancies from the most relevant (rank 1) to the least relevant (rank 5) for every one of the 30 resumes.
    Our annotators received detailed guidelines from a manager of recruitment, organizational development, and welfare in an HR department of a large academic institution. Mrs. Yaarit Katzav (https://en.sce.ac.il/administration/human-resources1/staff (accessed on 1 June 2023)) is a senior HR manager with over 13 years of experience who is responsible, among other things, for recruiting information systems engineers at SCE Academic College. The full vacancy–resume matching guidelines provided by her appear in Table 7.
  • Finally, we received 30 rank arrays of length 5 from both of our annotators and computed an average rank array of length 5 for every resume. (The resume dataset and annotations are freely available at https://github.com/NataliaVanetik/vacancy-resume-matching-dataset).
We use these human rankings as the ground truth for evaluating the performance of our method and competing methods; we compare the ground-truth rankings with the rankings produced by automatic methods for the selected 30 resumes and 5 vacancies.

5. Metrics

To analyze the correlation between ground-truth ratings and ratings produced by evaluated methods, we consider several metrics that are suited for comparing rankings.
Krippendorff’s alpha [64] is a metric used for assessing the agreement or reliability of rankings or ordinal data among multiple raters or annotators. The metric takes into account both the observed agreement and the expected agreement between raters, providing a measure of inter-rater reliability. It considers the differences between ranks assigned by different raters and compares them to the expected differences based on chance:
α = 1 expected   agreement observed   agreement
The metric values range from −1 to 1, with higher values indicating greater agreement among raters.
Spearman’s rank correlation coefficient [65] is a statistical measure used to assess the strength and direction of the monotonic relationship between two ranked variables. When applied to ranking comparison, Spearman’s rank correlation coefficient quantifies the similarity or agreement between two sets of rankings. It measures how consistently the orderings of the items in the two rankings correspond. It is computed as
ρ = 1 6 d i 2 n ( n 2 1 )
where ρ is Spearman’s rank correlation coefficient, d i stands for the difference between the ranks of the corresponding pair of items, and n is the number of items or observations in each ranking.
Cohen’s kappa coefficient [66] is a statistical measure used to assess the agreement between two raters or annotators when working with categorical or ordinal data. When applied to a comparison of rankings, Cohen’s kappa coefficient quantifies the agreement between two sets of rankings, considering the chance agreement that could occur by random chance. It is computed as
κ = observed   agreement expected   agreement 1 expected   agreement
From our observations, the three metrics correlated perfectly in all our experiments (i.e., produced the same rankings of methods). However, Krippendorff’s alpha consistently displayed higher values than Cohen’s kappa and Spearman’s correlation. It can offer a more thorough measure of agreement and is made to handle situations with more than two raters or coders [67]. Cohen’s kappa focuses on the pairwise agreement between two raters and may be less suggestive when comparing rankings. Therefore, we chose not to report Cohen’s kappa values. We used Krippendorff’s alpha and Spearman’s correlation to evaluate the agreement among the annotators and the agreement of vacancy ranks produced by different systems with the ground truth.

6. Experimental Evaluation

6.1. Hardware and Software Setup

All experiments were performed on Google Colab [68] with Pro settings. We used spaCy [51] for sentence splitting, tokenization, and NE extraction. Sklearn [69] and SciPy [70] Python packages were used to compute evaluation metrics and text representations. For keyword extraction, we used the KeyBERT SW package [39] with default settings.

6.2. Methods

Our baselines for comparison are the original OKAPI BM25 algorithm [24] and the BERT-based method of [54] denoted by BERT-rank. Our vector matching method is denoted by VM, and we use all varieties of text data (described in Section 2.2) and text representation vectors (described in Section 2.3). We have a total of 4 text data formats (full texts, summaries, and summaries extended with either keywords or named entities) for both resumes and vacancies and 5 text representations (word and character n-grams, BERT sentence embeddings, tf-idf vectors, and concatenations of all of these vectors), which gives us a total of 80 options. Therefore, we only report the best-performing combinations in each case.

6.3. Metrics

We compare the rankings produced by every method to the ground-truth ranking and report Krippendorff’s alpha (computed with ReCal web service [71]) and Spearman’s correlation values for all of them. For the automatic tests, we also report accuracy (a full explanation is provided in Section 6.5).

6.4. Human-Annotated Data Evaluation

In these tests, we perform a ranking of 5 selected vacancies for all of the 30 resumes annotated by humans. The rankings are compared to the average ranking produced by our two human annotators. Table 8 shows the results of ranking our VM method and the results for our two baselines—OKAPI BM25 and BERT-based rank—as well. The best scores are marked with a gray background.
We only list the VM methods (sorted by decreasing Krippendorff’s alpha) that outperform both baselines. The difference between the top method and baselines is very significant. Furthermore, we see that the best method uses the full text of resumes and summarized vacancies and character n-grams as text representation.
Figure 4 shows Krippendorff’s alpha values for all text representations for the best text type setup (full texts for resumes and summaries for vacancies). We observe that the best scores are achieved by the two representations that incorporate character n-grams and that the difference from other text representations is significant. Because pre-trained BERT models are trained on generic texts rather than resumes and vacancies, BERT sentence vectors perform poorly as text representation. The tf-idf vectors also underperform because vacancies and resumes come from a variety of candidates and firms, and less significant terms may be given greater weights simply because they are uncommon, rather than because they are important.
We can also see that the top three methods use summarized texts for either vacancies or resumes or both. Indeed, from looking at the texts, we can see that vacancy descriptions often contain excessive information, and resumes contain parts that are too generic. Our HR consultant has confirmed that the same resume is often submitted to several positions, and it is not adapted to a specific vacancy. She also explained that, in most cases, only parts of the vacancy description are specific to the job, and the rest often contains general company information.
Figure 5 shows the behavior of the VM method when the same text type is chosen for both resumes and vacancies. We can see that character n-gram-based text representation provides the best performance regardless of what text type (full text, summary, or keyword- and named-entity-enhanced summaries) was chosen, and that tf-idf gives the worst performance in all cases.
We have tested the possibility of using cosine similarity for this task, but it resulted in inferior performance. It turns out that the L1 distance is more suitable for this specific task. The intuition for this phenomenon is that cosine similarity treats all features equally, while L1 distance allows for feature-specific weighting, enabling the prioritization of specific features in the similarity calculation.
In Table 9, we list the scores of the top 10 variations in the VM method (according to their Krippendorff’s alpha value) when cosine similarity is used. The best scores are marked with a gray background color. Although most of these scores are still better than the baselines’ results, their values are significantly lower than in Table 8. We do, however, observe that text representations that use character n-grams produce the best results in this setup as well.
We also performed an additional human evaluation for the relevance of results produced by the top competing model (OKAPI-BM25) and our top model (VM with full-text resumes, summarized vacancies, and character n-gram text representation). The top-ranked job openings selected by the two top techniques for each of the 66 resumes in the whole dataset were provided to the annotators. Every resume–vacancy pair was given a relevance score from 1 to 3, with 1 denoting that the position is unimportant, 2 denoting that it is somewhat relevant, and 3 denoting that the position is relevant to the applicant. Each resume and vacancy was assessed by two different annotators. Average relevancy scores for the two methods appear in Table 10. We can see that the best vacancies selected by the top VM method are more relevant than those selected by the OKAPI-BM25 algorithm. A pairwise two-tailed statistical significance test [72] showed that the difference in assessments for these methods is statistically significant with a p-value of 0.0455.

6.5. Automatically Annotated Data Evaluation

To be less dependent on human annotations, we have designed the following test that can be applied to all resumes in our dataset.
We have extracted one random relevant developer vacancy, D developer , from our dataset and one data scientist/senior vacancy, D data scientist . Because our resume dataset contains data for software developers and full-stack engineers only, this type of vacancy is not relevant for them. Therefore, the ground-truth ranking for all resumes has the form
[ D developer , D data scientist ]
to which we can compare the rankings produced by our and baseline methods. In this case, we can report binary classification accuracy as well because we consider the answer (1) correct and the answer [ D data scientist , D developer ] to be incorrect. Table 11 contains the results for baseline methods and all variations in the VM method that demonstrate non-negative Krippendorff’s alpha values. The best scores are marked with a gray background. We can see that the top 5 VM methods demonstrate better performance than any of our baselines in all metrics, including accuracy. We can see that scores of all the methods improve, meaning that this is an easier task than the ranking of five relevant vacancies. The best performance is achieved when NE and summary are used as text data and character n-grams are chosen as text representation.

6.6. Extended Human-Annotated Data Evaluation

Because human ranking is a time-demanding task, we suggest the following semi-automated test to further assess ranking methods. We expand the five annotated vacancies with a single random data scientist vacancy from JobsPikr to obtain six vacancies. Because the extra vacancy is not relevant for applicants in our 30 annotated resumes, its rank should be last for every single one of them. With this approach, we do not need to compute a new ground truth. Table 12 shows the results of ranking our VM method and the results for our two baselines—OKAPI BM25 and BERT-based rank—as well. The best scores are marked with a gray background. We only list the VM methods that outperform both baselines. We can see that VM methods perform much better than baselines and that the top method is the same configuration that produced the best scores in Table 8. Furthermore, we see that OKAPI performs significantly better when there is one irrelevant in the list than in the case where there is none (Table 8).

7. Limitations

7.1. Resume Type

The proposed method focuses solely on information technology (IT) resumes and vacancies due to the data availability for this specific domain. As a result, the findings and conclusions drawn in this study are limited to the context of IT-related job applications and may not be directly applicable to other industries or professions. The exclusion of data from other fields may restrict the generalizability of the research outcomes beyond the IT domain. Our future research will try to combine diverse datasets from other areas in order to provide a more thorough understanding of automatic resume–vacancy matching systems in various job sectors.

7.2. Resume Structure

The vacancy–resume matching method used in this study is limited to unstructured resumes alone, thereby limiting its application. Unstructured resumes lack a uniform format and can vary greatly in content and organization, making it difficult to reliably extract useful information; while unstructured resumes are common, many firms also receive structured resumes via online application portals where individuals enter their information into predetermined fields. However, more study and methodological changes would be required to broaden the application of our matching methodology.

8. Conclusions

In this paper, we presented a new annotated resume dataset for the task of a resume–vacancy ranking. Additionally, we have proposed a method based on the statistical and semantic vector representation of texts constructed with NLP techniques. The method is unsupervised and does not require training. We compared our VM method with two baselines—OKAPI BM25 and BERT-rank—using various text data formats and representations.
The results of our evaluation demonstrate that our method outperforms both baselines in terms of ranking accuracy. Specifically, we found that using the full text of resumes and summarized vacancies, along with character-n-gram-based text representation, yielded the best performance. This combination achieved a significantly higher Krippendorff’s alpha value compared to the baselines. Furthermore, we analyzed the impact of different text representations and text data types on the performance of our VM method. We observed that text representations incorporating character n-grams consistently produced the best results, while tf-idf-based representations yielded the lowest performance.
Additionally, we found that using the same text type for both resumes and vacancies, particularly character-n-gram-based representations, resulted in improved rankings. It also became apparent that the L1 distance was better suited for this specific ranking task; while the performance using cosine similarity surpassed the baselines, the scores were noticeably lower when compared to the L1-distance-based approach.
Overall, the summarization of vacancies does improve the ranking results; however, it is not the best text type choice for the resumes. Therefore, the answer to research question 2 is partially true. However, adding keywords does not produce the best results except for the case of automatic evaluation where an irrelevant vacancy is present. The answer to research question 1 concerning named entities depends on the data—we only benefit from named entities if we have vacancies unrelated to our applicants. We have found that the best text representations in every setup are the ones that use character n-grams, thus answering research question 3. Automatic and semi-automatic evaluations in Section 6.5 and Section 6.6 show that the performance of all the methods, including OKAPI, improves when we evaluate resumes against vacancies that are not relevant for the applicants. Moreover, the advantage of our approach is preserved in these cases too. Therefore, the answer to research question 4 is positive.   

Author Contributions

Conceptualization, N.V. and G.K.; methodology, N.V.; software, N.V.; validation, G.K.; formal analysis, N.V.; investigation, N.V.; resources, G.K.; data curation, G.K.; writing—original draft preparation, N.V. and G.K.; writing—review and editing, N.V.; visualization, G.K.; All authors have read and agreed to the published version of this manuscript.

Funding

This research received no external funding.

Data Availability Statement

The data used in this study is freely available at https://github.com/NataliaVanetik/vacancy-resume-matching-dataset.

Acknowledgments

We wish to express our deepest gratitude to Yaarit Katzav, a senior HR manager at SCE Academic College, for taking the time to explain to us and our annotators how to match vacancies to resumes. Her guidance and valuable insights have been extremely helpful.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

    The following abbreviations are used in this manuscript:
BERTBidirectional Encoder Representations from Transformers
BOWBag-of-Words
CVCurriculum Vitae
HRHuman Resources
IRInformation Retrieval
ITInformation Technology
NENamed Entity or Entities
NERNamed Entity Recognition
NLPNatural Language Processing
RQResearch Question
TF-IDFTerm Frequency-Inverse Document Frequency
VMVector Matching

Appendix A

Appendix A.1. Named Entity Recognition Details

The main types of NE that can be identified with spaCy appear in Table A1. Examples of named entities for both types of data appear in Table A2, together with their types (note that all NEs are lower-cased).
Table A1. NEs in spaCy.
Table A1. NEs in spaCy.
NEDescription
PERSONIndividuals, including names of people, fictional characters, or groups of people.
ORGOrganizations, institutions, companies, or agencies.
GPEGeo-political entity that includes countries, cities, states, provinces, etc.
LOCNon-GPE locations, such as mountains, bodies of water, or specific landmarks.
PRODUCTNamed products, including goods or commercial items.
EVENTNamed events, such as sports events, festivals, or conferences.
WORK_OF_ARTArtistic works, including books, paintings, songs, or movies.
LAWNamed laws, legislations, or regulations.
LANGUAGENamed languages or language-related terms.
DATEDates or periods of time.
TIMESpecific times or time intervals.
PERCENTPercentages or numerical ratios.
MONEYMonetary values, including currencies, amounts, or symbols.
QUANTITYMeasurements or quantities.
ORDINALOrdinal numbers.
CARDINALCardinal numbers or numerical values.
Table A2. Named entity examples.
Table A2. Named entity examples.
Data TypeNamed EntityType
ResumeArlingtonGPE
mdmORG
OverCARDINAL
500CARDINAL
TheDATE
FirstDATE
ThirtyDATE
DaysDATE
TheDATE
FirstDATE
Vacancy1990DATE
IsraelGPE
YearsDATE
HebrewLANGUAGE
MasaORG
EnglishNORP

Appendix A.2. Keyword Analysis for Developer and Non-Developer Vacancies

Table A3 provides the lists of the top 10 keywords for the two types of vacancies used in automated experiments of Section 6.5. As can be seen from the table, the keyword phrases are different, and there are only four common keyword phrases.
Table A3. Top ten keywords for developer and data science vacancies.
Table A3. Top ten keywords for developer and data science vacancies.
Developer Vacancies
Keywords Count
Software developer 25
Software development 6
Seeking software developer 6
Software developers 5
Senior software developer 4
Participate in scrum 4
Developer in Windsor 4
Development environment candidates 4
Agile development environment 4
Lead software developer 3
Data Science Vacancies
Keywords Count
Data scientist 35
Data scientist with 7
Data scientists 6
The data scientist 5
Seeking data scientist 4
Data analyst 3
Data scientist for 3
Senior data scientist 3
Data scientist role 3
Data engineers 3
Responsibilities kforce 1
Responsibilities kforce has 1
Software development engineer 1
Talented software development 1

Appendix A.3. The Effect of Different Sentence Embeddings

We have compared different pre-trained sentence embedding models on resumes with human rankings and report the results in Table A4. In addition to bert-base-uncased [52], we tested the nli-bert-large-max-pooling [73], longformer [74], and all-distilroberta-v1 [75] models. The top five results for every sentence embedding model are reported.
While none of these setups outperforms our best model, we can see that the scores improve significantly for nli-bert-large-max-pooling and longformer compared to our base model, bert-base-uncased. Therefore, we see the obvious benefits of using larger models when text representation is set to be sentence embedding.
Table A4. Comparing different sentence embedding representations.
Table A4. Comparing different sentence embedding representations.
Resume Text DataVacancy Text DataSentence Embedding ModelKrippendorff’s
Alpha
Spearman’s
Correlation
kw + summarySummarybert-base-uncased0.36950.3003
ne + summaryFull textbert-base-uncased0.34660.2773
ne + summarySummarybert-base-uncased0.31550.2403
Full textFull textbert-base-uncased0.30970.2340
Full textSummarybert-base-uncased0.29170.2140
Keywords + summarySummarynli-bert-large-max-pooling0.62110.5816
Keywords + summaryner + summarynli-bert-large-max-pooling0.61510.5749
ner + summaryKeywords + summarynli-bert-large-max-pooling0.61390.5779
Full textSummarynli-bert-large-max-pooling0.60700.5680
Full textKeywords + summarynli-bert-large-max-pooling0.60400.5646
ner + summarySummaryLongformer0.58720.5467
Keywords + summarySummaryLongformer0.58060.5406
Full textFull textLongformer0.55640.5138
ner + summaryKeywords + summaryLongformer0.55170.5007
Full textSummaryLongformer0.53620.4900
Full textKeywords + summaryall-distilroberta-v10.54030.4867
Keywords + summarySummaryall-distilroberta-v10.52820.4808
Keywords + summaryKeywords + summaryall-distilroberta-v10.49220.4408
ner + summaryKeywords + summaryall-distilroberta-v10.47870.4198
Full textSummaryall-distilroberta-v10.42450.3672

Appendix A.4. Extended Automatic Evaluation

To further extend our automatically annotated evaluation of Section 6.5, we extracted 100 random developer vacancies from our dataset and 100 data scientist/senior vacancies and constructed 100 pairs of vacancies. For all these pairs, the ground truth has the form described in (1). We have evaluated our baselines and methods for all of the resumes against every one of the 100 vacancy pairs. The results are reported in Table A5; the top 20 results for the VM method variations are reported. We can see that all the reported VM methods demonstrate better performance than any of our baselines in all metrics, including accuracy; in fact, 19 of them achieve a perfect accuracy score of 1.0.

Appendix A.5. Rankings Produced by ChatGPT API

ChatGPT [76] is a language model developed by OpenAI [77], specifically referred to as a large language model (LLM). Trained using machine learning techniques, it can generate human-like text based on the input it is given. It analyzes the structure and content of the text it has been trained on, learns patterns, and then applies this knowledge to create new, contextually relevant content.
We used OpenAI API at https://platform.openai.com/docs/api-reference (accessed on 1 August 2023) to supply the LLM used by ChatGPT with five vacancies to evaluate and posed a query asking to rank these vacancies from the most relevant to the least relevant for every one of the 30 resumes we have human annotations for. The results of the comparison to human rankings appear in Table A6. We can see that although there is a positive correlation with human judgment in all three metrics, these scores fall far below the scores of the OKAPI BM25 algorithm and our top models in Table 8.
Table A5. Extended automatic tests.
Table A5. Extended automatic tests.
BaselineKrippendorff’s
Alpha
Spearman’s
Correlation
Acc
OKAPI-BM25
BERT-rank
−0.0455
−0.2045
−0.3939
−0.6061
0.3030
0.1970
VM
Resume
text data
Vacancy
text data
Text
representation
Krippendorff’sSpearman’sacc
Keywords + summaryner + summarytfidf111
Keywords + summaryJob + descriptiontfidf + char ng + word ng + sbert111
Keywords + summaryJob + descriptiontfidf111
Keywords + summaryJob + descriptionChar ng111
Keywords + summaryJob + descriptionWord ng111
ner + summaryner + summarytfidf111
ner + summaryJob + descriptiontfidf + char ng + word ng + sbert111
ner + summaryJob + descriptiontfidf111
ner + summaryJob + descriptionChar ng111
ner + summaryJob + descriptionWord ng111
Summaryner + summarytfidf111
SummaryJob + descriptiontfidf + char ng + word ng + sbert111
SummaryJob + descriptiontfidf111
SummaryJob + descriptionChar ng111
SummaryJob + descriptionWord ng111
Docner + summarytfidf111
DocJob + descriptiontfidf111
DocJob + descriptionWord ng111
ner + summarySummarytfidf0.97730.96970.9848
Table A6. Extended automatic tests on 30 human-ranked resumes.
Table A6. Extended automatic tests on 30 human-ranked resumes.
BaselineKrippendorff’s
Alpha
Spearman
Correlation
ChatGPT0.14220.0430

References

  1. Do, D.Q.; Tran, H.T.; Ha, P.T. Job Vacancy and Résumé Analysis for Efficient Matching of Job Offers and Applicants. In Proceedings of the IEEE International Conference on Data Science and Advanced Analytics, Montreal, QC, Canada, 17–19 October 2016. [Google Scholar]
  2. Nocker, M.; Sena, V. Big data and human resources management: The rise of talent analytics. Soc. Sci. 2019, 8, 273. [Google Scholar] [CrossRef]
  3. Chala, S.A.; Ansari, F.; Fathi, M.; Tijdens, K. Semantic matching of job seeker to vacancy: A bidirectional approach. Int. J. Manpow. 2018, 39, 1047–1063. [Google Scholar]
  4. McKinney, A. Real Resumix & Other Resumes for Federal Government Jobs: Including Samples of Real Resumes Used to Apply for Federal Government Jobs; Prep Publishing: Fayetteville, NC, USA, 2003. [Google Scholar]
  5. Riabchenko, A. Taxonomy-Based Vacancy: CV Matching. Master’s Thesis, LUT University, Lappeenranta, Finland, 2022. [Google Scholar]
  6. Senthil Kumaran, V.; Sankar, A. Towards an automated system for intelligent screening of candidates for recruitment using ontology mapping (EXPERT). Int. J. Metadata Semant. Ontol. 2013, 8, 56–64. [Google Scholar] [CrossRef]
  7. Pudasaini, S.; Shakya, S.; Lamichhane, S.; Adhikari, S.; Tamang, A.; Adhikari, S. Scoring of Resume and Job Description Using Word2vec and Matching Them Using Gale-Shapley Algorithm. In Proceedings of the ICOECA 2021, Bangalore, India, 18–19 February 2021. [Google Scholar] [CrossRef]
  8. Tejaswini, K.; Umadevi, V.; Kadiwal, S.; Revanna, S. Design and Development of Machine Learning based Resume Ranking System. Glob. Transit. Proc. 2022, 3, 371–375. [Google Scholar] [CrossRef]
  9. Kawan, S.; Mohanty, M.N. Multiclass Resume Categorization Using Data Mining. Int. J. Electr. Eng. Technol. 2020, 11, 267–274. [Google Scholar]
  10. Sammut, C.; Webb, G.I. Encyclopedia of Machine Learning; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2011. [Google Scholar]
  11. Stefanovič, P.; Kurasova, O.; Štrimaitis, R. The N-Grams Based Text Similarity Detection Approach Using Self-Organizing Maps and Similarity Measures. Appl. Sci. 2019, 9, 1870. [Google Scholar] [CrossRef]
  12. Brown, P.F.; Della Pietra, V.J.; Desouza, P.V.; Lai, J.C.; Mercer, R.L. Class-based n-gram models of natural language. Comput. Linguist. 1992, 18, 467–480. [Google Scholar]
  13. Dice, L.R. Measures of the amount of ecologic association between species. Ecology 1945, 26, 297–302. [Google Scholar] [CrossRef]
  14. Mikolov, T.; Sutskever, I.; Chen, K.; Corrado, G.S.; Dean, J. Distributed representations of words and phrases and their compositionality. In Proceedings of the 27th Annual Conference on Neural Information Processing Systems, Lake Tahoe, NV, USA, 5–10 December 2013; Volume 26. [Google Scholar]
  15. Qin, C.; Zhu, H.; Xu, T.; Zhu, C.; Jiang, L. Enhancing person-job fit for talent recruitment: An ability-aware neural network approach. In Proceedings of the 41st International ACM SIGIR Conference on Research & Development in Information Retrieval, Ann Arbor, MI, USA, 8–12 July 2018; pp. 25–34. [Google Scholar]
  16. Zhu, C.; Zhu, H.; Xiong, H.; Ma, C.; Xie, F.; Ding, P.; Li, P. Person-job fit: Adapting the right talent for the right job with joint representation learning. ACM Trans. Manag. Inf. Syst. 2018, 9, 12.1–12.17. [Google Scholar] [CrossRef]
  17. Yuan, Z.; Lu, S. Application in person-job fit evaluation of BP neural network in knowledge workers. J. Wuhan Univ. Technol. 2010, 32, 515–518. [Google Scholar]
  18. Bing, Y.; Fei, J. Application in person-job fit evaluation of support vector machine. J. Cent. South Univ. For. Fechnology (Soc. Sci.) 2011, 5, 92–94. [Google Scholar]
  19. Alsaif, S.A.; Sassi Hidri, M.; Eleraky, H.A.; Ferjani, I.; Amami, R. Learning-Based Matched Representation System for Job Recommendation. Computers 2022, 11, 161. [Google Scholar] [CrossRef]
  20. Koh, M.F.; Chew, Y.C. Intelligent job matching with self-learning recommendation engine. Procedia Manuf. 2015, 3, 1959–1965. [Google Scholar] [CrossRef]
  21. Lu, Y.; Ingram, S.; Gillet, D. A recommender system for job seeking and recruiting website. In Proceedings of the 22nd International Conference on World Wide Web, Rio de Janeiro, Brazil, 13–17 May 2013; pp. 963–966. [Google Scholar]
  22. Gionis, A.; Gunopulos, D.; Koudas, N. Machine learned job recommendation. In Proceedings of the Fifth ACM Conference on Recommender Systems, Chicago, IL, USA, 23–27 October 2011; pp. 325–328. [Google Scholar]
  23. Wu, D.; Zhou, L.; Lin, H. Design and implementation of job recommendation system for graduates based on random walk. J. Guangxi Norm. Univ. 2011, 29, 179–185. [Google Scholar]
  24. Robertson, S.; Walker, S.; Jones, S.; Hancock-Beaulieu, M.; Gatford, M. Okapi at TREC-3; Nist Special Publication, Diane Publishing Co.: Darby, PA, USA, 1995; Volume 109, p. 109. [Google Scholar]
  25. Espenakk, E.; Knalstad, M.J.; Kofod-Petersen, A. Lazy learned screening for efficient recruitment. In Proceedings of the Case-Based Reasoning Research and Development: 27th International Conference, ICCBR 2019, Otzenhausen, Germany, 8–12 September 2019; Springer: Berlin/Heidelberg, Germany, 2019; pp. 64–78. [Google Scholar]
  26. Martinez-Gil, J.; Paoletti, A.; Pichler, M. A Novel Approach for Learning How to Automatically Match Job Offers and Candidate Profiles. Inf. Syst. Front. 2020, 22, 1265–1274. [Google Scholar] [CrossRef]
  27. Moratanch, N.; Chitrakala, S. A survey on extractive text summarization. In Proceedings of the 2017 International Conference on Computer, Communication and Signal Processing (ICCCSP), Chennai, India, 10–11 January 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 1–6. [Google Scholar]
  28. Widyassari, A.P.; Rustad, S.; Shidik, G.F.; Noersasongko, E.; Syukur, A.; Affandy, A. Review of automatic text summarization techniques & methods. J. King Saud Univ.—Comput. Inf. Sci. 2022, 34, 1029–1046. [Google Scholar]
  29. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is All You Need. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; pp. 5998–6008. [Google Scholar]
  30. Abdel-Salam, S.; Rafea, A. Performance study on extractive text summarization using BERT models. Information 2022, 13, 67. [Google Scholar] [CrossRef]
  31. Salton, G. Introduction to Modern Information Retrieval; McGraw-Hill: New York, NY, USA, 1983. [Google Scholar]
  32. Li, J.; Fan, Q.; Zhang, K. Keyword extraction based on tf/idf for Chinese news document. Wuhan Univ. J. Nat. Sci. 2007, 12, 917–921. [Google Scholar] [CrossRef]
  33. Mihalcea, R.; Tarau, P. Textrank: Bringing order into text. In Proceedings of the 2004 Conference on Empirical Methods in Natural Language Processing, Barcelona, Spain, 25–26 July 2004; pp. 404–411. [Google Scholar]
  34. Rose, S.; Engel, D.; Cramer, N.; Cowley, W. Automatic keyword extraction from individual documents. In Text Mining: Applications and Theory; John Wiley & Sons: Hoboken, NJ, USA, 2010; pp. 1–20. [Google Scholar]
  35. Campos, R.; Mangaravite, V.; Pasquali, A.; Jorge, A.; Nunes, C.; Jatowt, A. YAKE! Keyword extraction from single documents using multiple local features. Inf. Sci. 2020, 509, 257–289. [Google Scholar] [CrossRef]
  36. Hu, J.; Li, S.; Yao, Y.; Yu, L.; Yang, G.; Hu, J. Patent keyword extraction algorithm based on distributed representation for patent classification. Entropy 2018, 20, 104. [Google Scholar] [CrossRef]
  37. Huh, J.H. Big data analysis for personalized health activities: Machine learning processing for automatic keyword extraction approach. Symmetry 2018, 10, 93. [Google Scholar] [CrossRef]
  38. Sordoni, A.; Galley, M.; Auli, M.; Brockett, C.; Ji, Y.; Mitchell, M.; Nie, J.Y.; Gao, J.; Dolan, B. A neural network approach to context-sensitive generation of conversational responses. arXiv 2015, arXiv:1506.06714. [Google Scholar]
  39. Grootendorst, M. KeyBERT: Minimal Keyword Extraction with BERT. 2020. Available online: https://github.com/MaartenGr/KeyBERT (accessed on 1 June 2023). [CrossRef]
  40. Lafferty, J.; McCallum, A.; Pereira, F.C. Conditional Random Fields: Probabilistic Models for Segmenting and Labeling Sequence Data; University of Pennsylvania: Philadelphia, PA, USA, 2001. [Google Scholar]
  41. Das, A.; Garain, U. CRF-based named entity recognition@ icon 2013. arXiv 2014, arXiv:1409.8008. [Google Scholar]
  42. Huang, Z.; Xu, W.; Yu, K. Bidirectional LSTM-CRF models for sequence tagging. arXiv 2015, arXiv:1508.01991. [Google Scholar]
  43. Lample, G.; Ballesteros, M.; Subramanian, S.; Kawakami, K.; Dyer, C. Neural architectures for named entity recognition. arXiv 2016, arXiv:1603.01360. [Google Scholar]
  44. Syed, M.H.; Chung, S.T. MenuNER: Domain-adapted BERT based NER approach for a domain with limited dataset and its application to food menu domain. Appl. Sci. 2021, 11, 6007. [Google Scholar] [CrossRef]
  45. Luoma, J.; Pyysalo, S. Exploring cross-sentence contexts for named entity recognition with BERT. arXiv 2020, arXiv:2006.01563. [Google Scholar]
  46. Cetoli, A.; Bragaglia, S.; O’Harney, A.D.; Sloan, M. Graph convolutional networks for named entity recognition. arXiv 2017, arXiv:1709.10053. [Google Scholar]
  47. Zhou, R.; Xie, Z.; Wan, J.; Zhang, J.; Liao, Y.; Liu, Q. Attention and Edge-Label Guided Graph Convolutional Networks for Named Entity Recognition. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, Abu Dhabi, United Arab Emirates, 7–11 December 2022; pp. 6499–6510. [Google Scholar]
  48. Francis, S.; Van Landeghem, J.; Moens, M.F. Transfer learning for named entity recognition in financial and biomedical documents. Information 2019, 10, 248. [Google Scholar] [CrossRef]
  49. Miller, D. Bert-Extractive-Summarizer PyPI. 2019. Available online: https://pypi.org/project/bert-extractive-summarizer (accessed on 1 June 2023).
  50. Novoresume. Resume Keywords: How to Use Them. Novoresume. Available online: https://novoresume.com/?noRedirect=true (accessed on 12 June 2023).
  51. Honnibal, M.; Montani, I. spaCy: Industrial-Strength Natural Language Processing in Python. 2020. Available online: https://spacy.io (accessed on 4 June 2023).
  52. Devlin, J.; Chang, M.W.; Lee, K.; Toutanova, K. BERT; Pre-Training of Deep Bidirectional Transformers for Language Understanding. arXiv 2018, arXiv:1810.04805. [Google Scholar]
  53. Amati, G. BM25. In Encyclopedia of Database Systems; Liu, L., Özsu, M., Eds.; Springer: Boston, MA, USA, 2009; pp. 59–61. [Google Scholar] [CrossRef]
  54. Hofstätter, S.; Althammer, S.; Schröder, M.; Sertkan, M.; Hanbury, A. Improving Efficient Neural Ranking Models with Cross-Architecture Knowledge Distillation. In Proceedings of the 2021 Conference on Human Information Interaction and Retrieval, Canberra, Australia, 14–19 March 2021; ACM: New York, NY, USA, 2021; pp. 84–93. [Google Scholar]
  55. Jha, A.; Kumar, P. JobsPikr Datasets. 2017. Available online: https://data.world/jobspikr/software-developer-job-listings-from-usa (accessed on 20 February 2023).
  56. Jha, A.; Kumar, P. JobsPikr. 2017. Available online: https://www.jobspikr.com/ (accessed on 20 February 2023).
  57. Job Feed API. 2017. Available online: https://www.jobspikr.com/job-feed-api/ (accessed on 22 October 2022).
  58. Ahlmann, J. Find_Job_Titles Python SW Package. 2017. Available online: https://pypi.org/project/find-job-titles (accessed on 1 June 2023).
  59. Telegram FZ LLC; Telegram Messenger Inc. Telegram. Available online: https://telegram.org (accessed on 1 June 2023).
  60. Telegram FZ LLC; Telegram Messenger Inc. HighTech Israel Jobs Telegram Channel. Available online: https://tgstat.ru/en/channel/@israjobs (accessed on 5 September 2019).
  61. Docs.Python.Org. Hashlib—Secure Hashes and Message Digests. 2022. Available online: https://docs.python.org/3/library/hashlib.html (accessed on 22 October 2022).
  62. Danilak, M.M. Langdetect PyPI. 2022. Available online: https://pypi.org/project/langdetect/ (accessed on 22 October 2022).
  63. Singh, V. Flashtext PyPI. 2020. Available online: https://pypi.org/project/flashtext (accessed on 22 October 2022).
  64. Krippendorff, K. Computing Krippendorff’s Alpha-Reliability; University of Pennsylvania: Philadelphia, PA, USA, 2011. [Google Scholar]
  65. Myers, L.; Sirois, M.J. Spearman correlation coefficients, differences between. Encycl. Stat. Sci. 2004, 12. [Google Scholar] [CrossRef]
  66. Cohen, J. A coefficient of agreement for nominal scales. Educ. Psychol. Meas. 1960, 20, 37–46. [Google Scholar] [CrossRef]
  67. de Raadt, A.; Warrens, M.J.; Bosker, R.J.; Kiers, H.A. A comparison of reliability coefficients for ordinal rating scales. J. Classif. 2021, 38, 519–543. [Google Scholar] [CrossRef]
  68. Kim, D.; Chai, K.; Kim, J.; Lee, H.; Lee, J.; Kim, J. Colaboratory: An educational research environment for machine learning using Jupyter Notebooks. J. Educ. Resour. Comput. (JERC) 2017, 16, 1–10. [Google Scholar]
  69. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V.; et al. scikit-learn: Machine Learning in Python. 2011. Available online: https://scikit-learn.org (accessed on 4 June 2023).
  70. Virtanen, P.; Gommers, R.; Oliphant, T.E.; Haberland, M.; Reddy, T.; Cournapeau, D.; Burovski, E.; Peterson, P.; Weckesser, W.; Bright, J.; et al. SciPy 1.0: Fundamental Algorithms for Scientific Computing in Python. Nat. Methods 2020, 17, 261–272. [Google Scholar] [CrossRef]
  71. Freelon, D. ReCal: Intercoder reliability calculation as a web service. Int. J. Internet Sci. 2010, 5, 20–33. [Google Scholar]
  72. Woolson, R.F.; Wilcoxon signed-rank test. Wiley Encyclopedia of Clinical Trials; John Wiley & Sons: Hoboken, NJ, USA, 2007; pp. 1–3. [Google Scholar]
  73. Reimers, N.; Gurevych, I. Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing, Hong Kong, China, 3–7 November 2019; Association for Computational Linguistics: Stroudsburg, PA, USA, 2019; Volume 11. [Google Scholar]
  74. Beltagy, I.; Peters, M.E.; Cohan, A. Longformer: The Long-Document Transformer. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Online, 5–10 July 2020; pp. 3642–3652. [Google Scholar]
  75. Sanh, V.; Debut, L.; Chaumond, J.; Wolf, T. DistilBERT, a distilled version of BERT: Smaller, faster, cheaper and lighter. arXiv 2019, arXiv:1910.01108. [Google Scholar]
  76. Brown, T.B.; Mann, B.; Ryder, N.; Subbiah, M.; Kaplan, J.; Dhariwal, P.; Neelakantan, A.; Shyam, P.; Sastry, G.; Askell, A.; et al. Language Models are Few-Shot Learners. arXiv 2020, arXiv:2005.14165. [Google Scholar]
  77. OpenAI. OpenAI. 2023. Available online: https://openai.com/ (accessed on 30 July 2023).
Figure 1. A pipeline of the VM method.
Figure 1. A pipeline of the VM method.
Information 14 00468 g001
Figure 2. Text types used by the VM method.
Figure 2. Text types used by the VM method.
Information 14 00468 g002
Figure 3. Example of a resume from the resume dataset.
Figure 3. Example of a resume from the resume dataset.
Information 14 00468 g003
Figure 4. Effect of text representations (left) and text data choice (right) on the best model configuration (dotted lines indicate text representations.)
Figure 4. Effect of text representations (left) and text data choice (right) on the best model configuration (dotted lines indicate text representations.)
Information 14 00468 g004
Figure 5. Choosing the same text type for resumes and vacancies (dotted lines indicate text types.)
Figure 5. Choosing the same text type for resumes and vacancies (dotted lines indicate text types.)
Information 14 00468 g005
Table 1. Keyword examples.
Table 1. Keyword examples.
Data TypeKeywordsScore
VacancySoftware development0.5359
Using software development0.529
Purpose development and0.5233
Purpose development0.5101
Maintenance of software0.5049
ResumeJava backend development0.6156
Backend java developer0.6135
Java backend0.6066
Backend java0.5917
Now backend java0.58
Table 2. Vacancy Dataset Features.
Table 2. Vacancy Dataset Features.
Field NameTypeNon-Empty CountMin LengthMax LengthAvg Length
Crawl_timestampString10,000252525
URLURL10,000---
Job_titleString10,0001814,230.102
CategoryString916123818.828
Company_nameString10,000227727.285
City string986914110.141
CountryString10,000232.559
Inferred_cityString78293228.649
Inferred_stateString92074208.325
Inferred_countryString92293138.602
Post_dateDate10,000---
Job_descriptionString10,0006427,5763966.576
Job_typeString10,0008108.938
Salary_offeredString0---
Job_boardString10,0004137.551
Geo string9999333
CursorInteger10,000---
Contact_emailString0---
Contact_phone_numberString81916012.009
Uniq_idString10,000323232
Table 3. A sample developer vacancy from the vacancies dataset.
Table 3. A sample developer vacancy from the vacancies dataset.
Column NameContent
Crawl_timestamp6 February 2019 05:53:38 +0000
URLhttps://www.dice.com/jobs/detail/C%252B%252B-Software-Developer-SigmaWay-San-Jose-CA-95101/90994229/801699
Job_titleC++ Software Developer
CategoryComputer-or-internet
Company_nameSigmaWay
CitySan Jose
StateCA
CountryUSA
Inferred_citySan Jose
Inferred_stateCalifornia
Inferred_countryUSA
Post_date5 February 2019
Job_descriptionApply by Email/Direct Application at [email protected] C++ Software Developer Duration: 12+ Months Location: Sunnyvale, CA Top 3 Must Haves 1. C++ Software Experience
(5+ years) 2. Education in information technology, computer science, or related field. Masters is preferred. 3. Knowledge of Software Development and Design across multiple platforms Preferred but not required Having experience with robotics projects or systems is a plus Required Qualifications Bachelors degree in information technology, computer science, or related field. Masters is preferred. At least 5 years experience in C++ development Excellent knowledge in algorithms and data structures Experienced in professional software development using versioning systems (git), reviewing systems Experience in agile development projects Excellent teaming and communication skills Ability to work in cross-functional teams
Job_typeUndefined
Salary_offered
Job_boardDice
GeoUSA
Cursor1,549,436,429,136,812
Contact_emailEmpty 100%
Contact_phone_number(408) 627-7905
Uniq_id3b22ba3aa471681a909f878d8cec1b58
Table 4. An example of a data science vacancy.
Table 4. An example of a data science vacancy.
FieldContent
Job_descriptionData Scientist EnvoyIT is looking for a Data Scientist for a Full-time position with one of our Healthcare clients in Reston, VA You will be joining a new initiative. This is an outstanding employer with plenty of growth opportunity. Base plus bonus and benefits. PURPOSE: As a Data Scientist, you will: Join a brand new team of machine learning researchers with an extensive track record in both academia and industry. Bring a combination of mathematical rigor and innovative algorithm design to create recipes that extract relevant insights from billions of rows of data to effectively & efficiently improve health outcomes. Create thoughtful solutions that engage and empower members to make more informed decisions about their health Develop statistical applications that can be reproduced and deployed on enterprise platforms. Develop functional means for measuring the quality of healthcare members receive annually. Interact with and report to an audience that includes Directors, Vice-Presidents and the C-level executives. Build tools and support structures needed to analyze data, perform elements of data cleaning, feature selection and feature engineering and organize experiments in conjunction with best practices. Assess the effectiveness and accuracy of new data sources and data gathering techniques.
Job_titleData Scientist
Uniq_ideda91b88eb3096ed98bc1a5f6b5568df
Table 5. Keyword and NE examples for a data science vacancy.
Table 5. Keyword and NE examples for a data science vacancy.
KeywordScoreNamed EntityType
Synergy software design0.64313dCARDINAL
Development team synergy0.58122004DATE
Synergy software0.5392EachDATE
Com synergy software0.5264YearDATE
Team synergy software0.5223MaritimeFAC
PlazaFAC
Table 6. A list of job titles for Oracle developers.
Table 6. A list of job titles for Oracle developers.
Job Title
Oracle ADF Developer, Oracle Apex Developer, Oracle Applications Developer, Oracle BPM Developer, Oracle BRM Developer, Oracle Business Intelligence Developer, Oracle Data Warehouse Developer, Oracle Database Developer, Oracle Developer, Oracle E Business Developer, Oracle EBS Developer, Oracle ERP Developer, Oracle ETL Developer, Oracle Financial Application Developer, Oracle Financials Developer, Oracle Forms Developer, Oracle Fusion Developer, Oracle Fusion Middleware Developer, Oracle HRMS Developer, Oracle OBIEE Developer, Oracle PL SQL Developer, Oracle R12 Developer, Oracle Reports Developer, Oracle SOA Developer, Oracle SQL Developer, Oracle Technical Developer
Table 7. Ranking guidelines.
Table 7. Ranking guidelines.
ParameterExplanation
Work experienceCandidates should ensure that their work experience matches the level required for the job postings they are applying for.
QualificationsEducational and professional qualifications are also important considerations. Candidates should make sure that they have the necessary qualifications required for the job.
ClarityCandidates should look for job postings with clear and specific job requirements and job culture descriptions, and avoid applying for roles with vague or generic requirements. They should tailor their applications to the specific job posting, highlighting the most important requirements that match their skills and qualifications.
ConcisenessSome job postings might have a long list of requirements. Try to focus on the most important requirements that match your skills and qualifications.
KeywordsCandidates should focus on the most relevant keywords and requirements that match their skills and qualifications, and avoid irrelevant information.
Industry trendsCandidates should stay up to date with the latest industry trends and developments in the IT field. This can help them identify new job opportunities and position themselves as experts in their field.
Table 8. Evaluation results for baselines and the VM method on human-annotated data.
Table 8. Evaluation results for baselines and the VM method on human-annotated data.
Baseline Krippendorff’s
Alpha
Spearman’s
Correlation
OKAPI-BM25
BERT-rank
0.3055
−0.1779
0.2262
−0.3071
MethodResume
text data
Vacancy
text data
Text
representation
VMFull textSummaryChar ng0.62870.5908
VMFull textSummarytfidf + char ng + word ng + sbert0.61790.5794
VMFull textkw + summaryChar ng0.58850.5455
VMFull textkw + summarytfidf + char ng + word ng + sbert0.50750.4555
VMkw + summaryFull texttfidf + char ng + word ng + sbert0.49180.4444
VMkw + summaryFull textChar ng0.49180.4444
VMne + summaryFull texttfidf + char ng + word ng + sbert0.49180.4444
VMne + summaryFull textChar ng0.49180.4444
VMSummaryFull texttfidf + char ng + word ng + sbert0.49180.4444
VMSummaryFull textChar ng0.49180.4444
VMne + summaryFull texttfidf0.48580.4377
VMFull textFull texttfidf0.47760.4273
VMFull textFull texttfidf + char ng + word ng + sbert0.47430.4210
VMkw + summaryFull texttfidf0.47160.4206
VMSummaryFull texttfidf0.47160.4206
VMFull textFull textChar ng0.46970.4192
VMkw + summarySummarysbert0.36950.3003
VMne + summaryne + summaryChar ng0.36450.2972
VMkw + summaryne + summarytfidf + char ng + word ng + sbert0.36230.2892
VMkw + summaryFull textWord ng0.35520.2900
VMne + summaryFull textWord ng0.35520.2900
VMSummaryFull textWord ng0.35520.2900
VMFull textFull textWord ng0.35520.2900
VMne + summaryFull textsbert0.34660.2773
VMSummaryne + summaryChar ng0.33830.2625
VMSummaryne + summarytfidf + char ng + word ng + sbert0.33120.2556
VMne + summarySummarysbert0.31550.2403
VMkw + summaryne + summaryChar ng0.31540.2377
VMFull textFull textsbert0.30970.2340
Table 9. Results for the VM method with cosine similarity on human-annotated data.
Table 9. Results for the VM method with cosine similarity on human-annotated data.
MethodResume
Text Data
Vacancy
Text Data
Text
Representation
Krippendorff’s
Alpha
Spearman’s
Correlation
VM cosineFull textFull textChar ng0.51860.4708
VM cosineFull textFull texttfidf + char ng + word ng + sbert0.50060.4508
VM cosineSummaryFull texttfidf + char ng + word ng + sbert0.44730.3894
VM cosineSummaryFull textChar ng0.39930.3361
VM cosinene + summaryFull texttfidf + char ng + word ng + sbert0.37410.3021
VM cosinener + summaryFull textChar ng0.36810.2955
VM cosinekw + summaryFull texttfidf + char ng + word ng + sbert0.32160.2516
VM cosinekw + summaryFull textChar ng0.32160.2516
VM cosinener + summaryner + summarysbert0.26120.1867
VM cosineFull textner + summarysbert0.18920.1067
VM cosineSummaryner + summarysbert0.17190.0861
Table 10. Relevance evaluation for the top VM method and OKAPI-BM25.
Table 10. Relevance evaluation for the top VM method and OKAPI-BM25.
MethodAverage Relevance
OKAPI-BM-252.0227
VM CV full text + vacancy summary + char ng2.1894
Table 11. Baselines and VM method results for automatically annotated data.
Table 11. Baselines and VM method results for automatically annotated data.
Baseline Krippendorff’s
Alpha
Spearman’s
Correlation
Acc
OKAPI-BM25
BERT-rank
0.7000
0.2700
0.6000
0.0300
0.8000
0.5200
MethodResume
text data
Vacancy
text data
Text
representation
VMFull textne + summaryChar ng0.90910.87880.9394
VMFull textne + summarytfidf + char ng + word ng + sbert0.86360.81820.9091
VMSummarySummaryWord ng0.84090.78790.8939
VMne + summarySummaryWord ng0.79550.72730.8636
VMkw + summarySummaryWord ng0.77270.69700.8485
VMFull textkw + SummaryChar ng0.65910.54550.7727
VMFull textSummaryChar ng0.54550.39390.6970
VMFull textkw + summarytfidf + char ng + word ng + sbert0.52270.36360.6818
VMFull textSummaryWord ng0.40910.21210.6061
VMFull textSummarytfidf + char ng + word ng + sbert0.38640.18180.5909
VMSummarykw + summaryWord ng0.36360.15150.5758
VMne + summarykw + summaryWord ng0.31820.09090.5455
VMne + summaryne + summaryWord ng0.25000.00000.5000
VMne + summaryne + summarysbert0.1818−0.09090.4545
VMSummaryne + summarysbert0.1818−0.09090.4545
VMSummaryne + summaryWord ng0.1591−0.12120.4394
VMkw + summarykw + summaryWord ng0.0909−0.21210.3939
VMFull textne + summarysbert0.0227−0.30300.3485
Table 12. Baselines and VM method results for extended human-annotated data.
Table 12. Baselines and VM method results for extended human-annotated data.
Baseline Krippendorff’s
Alpha
Spearman’s
Correlation
OKAPI-BM25
BERT-rank
0.5350
−0.2905
0.4908
−0.4070
MethodResume
text data
Vacancy
text data
Text
representation
VMFull textSummaryChar ng0.76300.7435
VMFull textSummarytfidf + char ng + word ng + sbert0.74110.7198
VMFull textkw + summaryChar ng0.73780.7158
VMkw + summaryFull texttfidf + char ng + word ng + sbert0.70450.6827
VMkw + summaryFull textChar ng0.70450.6827
VMne + summaryFull texttfidf + char ng + word ng + sbert0.70450.6827
VMne + summaryFull textChar ng0.70450.6827
VMSummaryFull texttfidf + char ng + word ng + sbert0.70450.6827
VMSummaryFull textChar ng0.70450.6827
VMFull textkw + summarytfidf + char ng + word ng + sbert0.68770.6605
VMkw + summaryFull textWord ng0.62520.5947
VMne + summaryFull textWord ng0.62520.5947
VMSummaryFull textWord ng0.62520.5947
VMFull textFull textWord ng0.62520.5947
VMFull textFull texttfidf + char ng + word ng + sbert0.59760.5643
VMFull textFull textChar ng0.56480.5298
VMkw + summarySummarysbert0.55870.5185
VMFull textSummarysbert0.53750.4958
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Vanetik, N.; Kogan, G. Job Vacancy Ranking with Sentence Embeddings, Keywords, and Named Entities. Information 2023, 14, 468. https://doi.org/10.3390/info14080468

AMA Style

Vanetik N, Kogan G. Job Vacancy Ranking with Sentence Embeddings, Keywords, and Named Entities. Information. 2023; 14(8):468. https://doi.org/10.3390/info14080468

Chicago/Turabian Style

Vanetik, Natalia, and Genady Kogan. 2023. "Job Vacancy Ranking with Sentence Embeddings, Keywords, and Named Entities" Information 14, no. 8: 468. https://doi.org/10.3390/info14080468

APA Style

Vanetik, N., & Kogan, G. (2023). Job Vacancy Ranking with Sentence Embeddings, Keywords, and Named Entities. Information, 14(8), 468. https://doi.org/10.3390/info14080468

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop