Next Article in Journal
Navigating the Digital Neurolandscape: Analyzing the Social Perception of and Sentiments Regarding Neurological Disorders through Topic Modeling and Unsupervised Research Using Twitter
Previous Article in Journal
An Agent-Based Model for Disease Epidemics in Greece
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Hierarchical Classification of Transversal Skills in Job Advertisements Based on Sentence Embeddings

by
Florin Leon
1,
Marius Gavrilescu
1,
Sabina-Adriana Floria
1 and
Alina Adriana Minea
2,*
1
Faculty of Automatic Control and Computer Engineering, “Gheorghe Asachi” Technical University of Iași, 700050 Iasi, Romania
2
Faculty of Materials Science and Engineering, “Gheorghe Asachi” Technical University of Iași, 700050 Iași, Romania
*
Author to whom correspondence should be addressed.
Information 2024, 15(3), 151; https://doi.org/10.3390/info15030151
Submission received: 15 February 2024 / Revised: 3 March 2024 / Accepted: 6 March 2024 / Published: 8 March 2024

Abstract

:
This paper proposes a classification methodology aimed at identifying correlations between job ad requirements and transversal skill sets, with a focus on predicting the necessary skills for individual job descriptions using a deep learning model. The approach involves data collection, preprocessing, and labeling using ESCO (European Skills, Competences, and Occupations) taxonomy. Hierarchical classification and multi-label strategies are used for skill identification, while augmentation techniques address data imbalance, enhancing model robustness. A comparison between results obtained with English-specific and multi-language sentence embedding models reveals close accuracy. The experimental case studies detail neural network configurations, hyperparameters, and cross-validation results, highlighting the efficacy of the hierarchical approach and the suitability of the multi-language model for the diverse European job market. Thus, a new approach is proposed for the hierarchical classification of transversal skills from job ads.

1. Introduction

The field of text classification, a fundamental subdomain within the natural language processing (NLP) field of machine learning (ML), has witnessed a remarkable evolution in recent years. With the exponential increase in the textual data generated across various domains, the need for effective text classification methods has become increasingly pressing. Text classification is the task of assigning predefined labels or categories to textual documents based on their content. This task holds immense importance across various industries and applications, including but not limited to sentiment analysis, spam detection, content recommendation, and news classification. The ability to automatically organize and categorize large volumes of text can streamline information retrieval, enhance decision-making processes, and enable efficient data management.
Traditional text classification methods rely on well-established techniques such as term frequency–inverse document frequency (TF-IDF) representations and traditional ML algorithms. TF-IDF measures the importance of each term within a document relative to a corpus of documents, providing a numerical representation of textual data. Classic ML algorithms, such as k-nearest neighbors, decision trees, naïve Bayes, or random forest, process the TF-IDF vectors to identify patterns and relationships among terms and have been successfully applied to text classification tasks.
While classic methods yielded commendable results, the emergence of deep learning (DL) has brought a new era of text classification. DL models, i.e., neural networks, have demonstrated unprecedented capabilities in handling the complexity and nuances of natural language. One of the key breakthroughs in DL for text classification lies in the use of word and sentence embeddings that represent words as vectors in high-dimensional spaces, capturing the semantic relationships between them. Sentence embeddings extend this idea to encode entire sentences or documents into vector-based representations.
DL models use these embeddings to learn complex patterns and contextual information within text data. Recurrent neural networks, e.g., long short-term memory (LSTM) models, and more advanced architectures such as those based on Transformers have achieved a state-of-the-art performance across a wide range of text classification tasks.
The motivation behind employing text classification for the identification of skills from job advertisements is rooted in the ever-evolving job market dynamics and the imperative need for efficient workforce matching. In today’s fast-paced and highly competitive job market, where the demand for specific skills is continually changing, exploiting the power of text classification offers several significant advantages.
First, it allows for precise talent matching. Job seekers possess a diverse range of skills and employers have specific skill requirements for their open positions. Text classification ensures that the right individuals with the exact skill sets are efficiently matched with job opportunities that necessitate those particular competences. This results in reduced skill mismatches, higher job satisfaction, and increased productivity.
Moreover, the job market is a dynamic entity characterized by rapid skill turnover due to technological advancements and evolving industry needs. Text classification offers the ability to perform real-time analysis of job descriptions and skill requirements. By staying up-to-date with the latest trends and demands, organizations can rapidly adapt, ensuring that their workforce remains competitive and aligned with market needs.
Furthermore, text classification optimizes resource allocation for human resources departments and job search platforms. It automates the labor-intensive process of scanning and categorizing skills from large numbers of job listings or CVs, thus saving valuable time and effort.
Lastly, another compelling motivation is the identification of skill gaps. For job seekers, text classification helps individuals pinpoint areas where they may lack necessary skills for a specific role. This knowledge empowers them to proactively seek out relevant training or education, fostering lifelong learning and career development.
Our contribution involves the development of classification models that accurately represent the relationships between job ads and transversal skill bases. The goal is to automate the identification of standardized transversal skills from the corresponding phrases in job ad texts. This is a challenging task, considering that the manner in which transversal skill requirements are expressed in job ads rarely matches standardized skill names or identifiers from the skill base. We address this problem by developing, training, and testing models based on neural networks, designed for the classification of job ads according to the standardized transversal skills from the ESCO platform. To accomplish this, we created our own dataset, where job ads are manually labeled based on the skills mentioned in the respective text. We assign skill labels to each phrase in the job ad, depending on whether or how transversal skill requirements are formulated in the phrase. Our classifiers consider the hierarchical structure outlined in the ESCO platform, where skills are organized on two levels, as follows: the higher level comprises more broadly phrased skills with a wider context, while the lower level encompasses sets of subskills for each top level skill, providing a more detailed specification.
The rest of the paper is structured as follows: In Section 2, we include a short survey of the related literature. Section 3 details the proposed methodology, including data preprocessing and our approach to hierarchical skill classification. In Section 4, we show the experimental results and in Section 5 we provide a brief discussion of our findings. In Section 6 we present the conclusions of this work.

2. Related Work

In general, the biggest obstacles in classifying job ads and identifying relevant skills can be summarized as follows: the availability of data, particularly pre-labeled sets of job ads and the corresponding skills that can be used for training supervised classifiers; the unstructured nature of many job ads, which feature a broad range of formats, phrasings, different manners of expressing similar requirements, and vagueness when formulating job requirements. These aspects cause the problem space to increase disproportionately compared to the available data, making it tedious and difficult to properly preprocess and standardize the training data, as well as to develop reliable classification models. Consequently, skill identification is a multifaceted topic addressed in a variety of ways in the related literature.
One of the more common and reliable methods for skill identification is skill counting. Manual skill counting relies on expert readers to identify relevant skills in job ads. This can be achieved without a knowledge base or using an already available skill set. The drawback of such an approach is that it is time-consuming and tedious; however, the availability of a prior list of skills to choose from makes the task easier for expert staff. The process can also be automated if a skill base is available. The competences can be identified either by Boolean indexing or by using a simple feature such as a word weight, which indicates the importance of the skill phrasing. A common weighing method makes use of metrics such as TF-IDF [1].
Early work in the direction of skill identification involved finding exact matches of skill labels from existing skill bases in job ad phrasing. Such methods are simple to implement and rely on searching for keywords or keyword combinations [2,3,4]. In the absence of a skill base, skill counting mainly relies on the assessment of expert annotators. The topic of manual or semi-automated content analysis has been a subject of thorough research, since it has the distinctive advantage of also performing a qualitative search alongside a quantitative one. Multiple studies employ content analysis for identifying job market needs, though the general consensus is that manual annotation, while often more reliable, is time consuming and ineffective for the systematic analysis of large bodies of text [5,6,7,8].
Multiple authors handle the skill search task by treating it as a topic modeling problem. In topic modeling, the main themes of a text are learned in an unsupervised manner, by determining and analyzing word distributions. The works that employ topic modeling algorithms identify the most required skills among the topics of the job ad. A common approach in this sense is to find the formulation of relevant skills in the most frequent keywords of the identified topics. In this context, the authors of [9] use latent semantic analysis (LSA) to carry out skill identification from job ads. LSA involves transforming the job ad set into a term matrix. This matrix is then subjected to dimensionality reduction via singular value decomposition, which results in sets of highly correlated keywords and documents. The phrases formed by these keywords are the topics identified in the texts, which are further subjected to data analysis techniques and expert evaluation. Similar work is reported by [10,11,12], where latent Dirichlet allocation (LDA) is used to identify popular keywords in job ads. Each document is transformed into a probability distribution of topics and each topic is treated as a probability distribution of words, all sharing a common Dirichlet prior [13].
Word embeddings have become increasingly popular in generating and classifying text. Similar to word embeddings, skill embedding methods generate vector representations of skill keywords, such that similar skills have a high similarity in the corresponding vector space. The aim of most works is to develop embedding spaces that work with simple similarity metrics, such as cosine or Jaccard similarities. In [14,15], the authors employ the Word2vec model [16] to derive vector representations for skill contexts. Subsequently, these vectors serve as inputs for a clustering algorithm, facilitating the grouping of aggregated contexts in clusters.
In ref. [17], the authors use Word2vec embeddings to assess the similarity between the skills mentioned in job ads and in professional standards. These standards encompass a set of principles, ethics, and behaviors obligatory for members of a specific profession. The training of Word2vec on a job ad corpus enables the model to learn contextual information from job ads, facilitating the clustering of skills based on their presence in the job ad texts.
The authors of [18] first identify skills explicitly, then infer implicit skills from job ads using Doc2vec [19]. The inference process involves identifying similar job descriptions that share common features such as location or company, assuming that they would also share similar skills. Implicit skills, in this context, are those not explicitly stated but considered important for a given position. Using document-level embeddings, the authors incorporate inferred skills into those extracted directly from the job ads based on their similarities.
The authors of [20] introduce Skill2vec, a technique aimed at optimizing candidate skill searches. Also inspired by Word2vec, Skill2vec maps skills into a vector space, revealing skill relationships. Training involves a neural network where skills are treated as words. This method creates a relationship graph among recruitment domain skills and can help candidates in identifying skill gaps relative to job requirements, guiding them towards suitable training opportunities.
In ref. [21], skill embeddings are determined using FastText [22], trained on job ads, to ensure the coherence of the extracted skills. This approach handles out-of-vocabulary instances, generating representations close to the original word, even when misspellings occur.
In ref. [23], the author addresses the evolving Norwegian job market’s needs by introducing a method to identify groups of words that represent skills in the text of job listings. The use cases, requirements, datasets, implementation, and design of such a skill extracting algorithm are described. The work also mentions some issues related to language ambiguity and semantic differences between datasets, which hinders precise skill extraction and underscores the need for the computation of semantic similarities to resolve ambiguity effectively. It also suggests leveraging other NLP techniques, such as named-entity recognition (NER) and part-of-speech (POS) tagging.
A popular class of methods involves the use of supervised ML algorithms. In particular, deep neural networks have proven particularly useful for NLP tasks. Various implementations and architectures have been developed in this direction, with promising results, often proving superior to unsupervised approaches. While methods based on deep neural networks have systematically demonstrated their reliability in capturing hidden word relationships and exhibit promising outcomes for various NLP tasks, they also have the downside of being data intensive. Achieving favorable results in the context of skill identification demands large, labeled datasets and meticulous fine-tuning.
In ref. [24], the authors use an LSTM architecture pre-trained to perform NER, i.e., the identification and classification of entities from unstructured text into predefined classes such as names, locations, codes, percentages, and organizations [25]. A common practice in this direction is to rely on manually labeled datasets comprising a large body of job ads, though the accuracy of manual annotation can greatly affect the reliability of the resulting models.
In ref. [26], the authors compare models based on convolutional neural networks (CNNs) and LSTM for a sentence classification problem. The CNN model incorporates word order criteria by applying a fixed-size window to the input array, consisting of words and their corresponding word embeddings [27]. The LSTM architecture takes advantage of the sequential nature of the text, addressing long-term dependencies and enabling predictions on variable-length inputs.
The authors of [28] rely on advanced language models such as Bidirectional Encoder Representations from Transformers (BERT) [29] for sentence classification in job ads. BERT is specifically designed to pre-train deep bidirectional representations from unlabeled text, considering both left and right contexts in all layers. Consequently, the pre-trained BERT model is fine-tuned with a single additional output layer, without significant modifications to the task-specific architecture. Other authors report the successful incorporation of BERT-based models into the classification pipeline. In [30], a BERT-based sentence transformer is used to perform initial feature extraction from job ad texts. Following a dimensionality reduction phase, a combination of NLP techniques and clustering methods is used to classify the job ads in the corresponding vector space. Another application of sentence transformers is by [31], who use the multilingual SBERT model [32] to determine vector representation of job ad phrases. They demonstrate that the embedding deduced by the transformer model is significantly more reliable at labeling skills in job ads than the alternative unsupervised approaches, while achieving accuracies close to manual annotation.
In ref. [33], the authors opt for a multi-label text classification approach to assign skills to each job description. Rather than classifying individual words in job descriptions, the authors treat the job descriptions as indicators for the binary relevance of multiple skills. To achieve this, they employ a BERT encoder and add an extra layer for multi-label classification. Additionally, a correlation aware bootstrapping process is introduced, encompassing structured semantic representations of skills and their co-occurrences to account for missing skills mentioned in job ads by augmenting the number of training examples.
Paper [34] addresses the scarcity of timely, comprehensive information on EU employers’ skill needs by proposing a system that analyzes online job vacancies. It aims to create a pan-European platform of these vacancies for insights into skill requirements, aiming toward real-time skill demand analysis. This system employs ontologies and ML models to process multilingual job postings across various European Union languages. ML algorithms match the job content to predefined terms, refining classification accuracy through expert validation and continuous ontology updates. In the proposed approach, each variable (e.g., occupation or region) and language require a separate ML model training, which is currently focused on occupation classification with plans to extend to other variables.
A few survey papers provide overviews and detailed analyses of the various methods employed in the related literature for finding relevant sources of job ads in academia [35], of knowledge extraction from job ads in the IT job market [36], or of skill identification techniques in terms of methods used, classification granularity, and existing implementations [37].

3. Proposed Methodology

This study aims to propose a methodology for identifying meaningful relationships between employers’ requirements in job ads and sets of transversal skills. To this extent, we generate an original dataset consisting of job ads annotated using labels drawn from the ESCO skill base. Additionally, our goal is to create a model that can effectively predict the necessary skills for individual job descriptions. To accomplish this task, our strategy involves the implementation of a deep learning model, considering that DL neural networks, with their ability to capture complex patterns and associations within data, are well suited for the nature of this problem. Our approach involves a comprehensive experimental study focusing on the application of neural network-based classification models for identifying transversal skills from job ads. The experimental pipeline includes several stages, as follows: initially, a preprocessing stage filters the job ads and divides them into sentences. Subsequently, a dataset generation stage creates training and test data using a sentence embedding model. Finally, the process involves generating and fine-tuning hierarchical classification models for the two ESCO skill levels.
Figure 1 summarizes the process described above.

3.1. Data Preprocessing

The first step is the data collection and preprocessing phase. Using 219 job ads downloaded from the EURES online platform [38], we devised an automated script that parses the text, eliminates personal information, such as web addresses and non-standard characters, and ultimately identifies the individual sentences in the ads.
Since the job ads are posted by companies from different European countries, they can be written in their respective national languages. As we will explain in the following sections, two approaches are used. The first one is to translate them into English using an automatic translation tool in order to use an English language sentence embedding model. The second one is to use a multi-language sentence embedding model applied to the original text.
These sentences then underwent manual labeling, for classifying them into skill classes and subclasses, using the skills and competences framework provided by the ESCO website [39]. ESCO, an initiative affiliated with the European Commission, serves as a classification system for European Skills, Competences, and Occupations. The framework includes six main classes of transversal skills, each with additional subclasses:
  • T1—Core skills and competences;
  • T2—Thinking skills and competences;
  • T3—Self-management skills and competences;
  • T4—Social and communication skills and competences;
  • T5—Physical and manual skills and competences;
  • T6—Life skills and competences.
The utilization of ESCO’s well-defined skill and competence framework ensures a structured and standardized approach to skill labeling. By classifying sentences from job ads into these skill classes, the ML models can enhance the job matching process and facilitate a better alignment between job seekers and employers in the European job market.
A sample of the results of the automatic parsing and manual classification of the dataset is presented in Figure 2. Each line corresponds to a sentence. The complete set of information has several components, delimited with semicolons. The first component is an identifier of the job ad and the second one is the identifier of the sentence in the job ad. The third component is the actual sentence in the form of a list of plain words, without punctuation marks or other identifiers. The last component includes the skills that the sentence refers to. One can notice that each sentence can belong to several classes (Tx) or subclasses (Tx.y). Also, there is a large number of sentences that do not represent transversal skills (marked with 0).

3.2. Hierarchical Skill Classification

We propose a hierarchical classification approach for the sentences extracted from the job ads. Our methodology involves several key steps aimed at enhancing the precision of skill identification and classification. In order to represent each sentence in a standardized manner, we employ a pre-trained sentence embedding model implemented in the “Sentence Transformers” library [32,40], based on the BERT architecture. This model transforms each sentence into a 768-element vector that captures its meaning. In this way, the text input is transformed into a numeric representation that can be used for further classification. The number of total sentences (i.e., instances in the dataset) is 5208.
For the first level of classification, we separately create individual classification models using neural networks for each type of competence (Tx). This results in the development of six distinct models, each dedicated to determining whether or not a sentence contains a specific type of competence. This addresses the fact that a single sentence can potentially belong to multiple skill classes (Tx).
One of the significant challenges encountered while working with the dataset was the great imbalance between sentences without skills and those containing skills. The majority of sentences in job ads typically do not contain explicit references to transversal skills. To mitigate this imbalance, we implement a data augmentation strategy; specifically, we employ sentence cloning to augment the dataset by replicating the sentences that contain skills to an extent where the positive and negative samples have approximately the same number. This augmentation technique aims to improve the robustness and performance of the classification models.
An alternative to simple cloning is paraphrasing, i.e., automatically creating other sentences with different words but the same meaning for a given input sentence. This can be performed, e.g., using the “Pegasus” library [41,42]. However, this procedure proved to be time-consuming and the final results did not show improvements compared to the simple cloning technique. Moreover, the paraphrasing model was created only for the English language and this posed additional limitations for working directly with multi-language models, as opposed to using English translations.
For the second level of the hierarchical classification, we employ a multi-label approach to further refine the classification of sentences into subclasses (Tx.y). While generating multiple single-label models is the more straightforward and easily interpretable approach, it has the downside of not accounting for instances which belong to multiple classes, therefore oversimplifying the problem. Multi-label models have been successfully employed for text classification tasks and, while more complex and difficult to train, provide a more comprehensive representation of the relationships between classes and instances, acknowledging that instances can belong to multiple categories. This level of classification aims to provide more granular information about the specific competences referred to in the sentences. However, this stage introduces several significant challenges. The very limited number of instances available for each subclass affects both the ability to create meaningful distinct models and the generalization capabilities of the models. Also, there is a lack of clear boundaries or distinctions between some of the subclasses, i.e., the same sentence can be classified into multiple subclasses simultaneously. More specifically, we had 133 instances (i.e., sentences) in T1, 238 instances in T2, 378 instances in T3, 395 instances in T4, 20 instances in T5, and 106 instances in T6. Therefore, we could not train separate models for each subclass, as we did in the “level 1” classification.
In order to address these issues, we adopted a multi-label classification strategy that allows a single sentence to be assigned to multiple subclasses simultaneously. One can see some examples in Figure 3. In this case, for each sentence there are a number of noi non-mutually exclusive outputs, where noi is the number of subclasses in class i. For the six main classes, (no1, …, no6) = (3, 4, 4, 5, 2, 6). For example, the second line in Figure 3 contains a sentence that belongs to class T1 in the first level, and to T1.1 and T1.3 in the second level. Since no1 = 3, a binary vector of three elements defines the desired output. Since only T1.1 and T1.3 are relevant, the corresponding output vector is (1, 0, 1).
In the previous stages of our research, we relied on a sentence embedding model tailored for the English language to process and analyze job ads. However, the next step was to employ a multi-language sentence embedding model directly, to be able to effectively handle job ads written in their original languages. This broader linguistic coverage better reflects the multilingual nature of today’s European job market, where job seekers and employers often interact in languages other than English.

4. Experimental Results

4.1. “Level 1” Classification

4.1.1. The English Language Model

In the experimental study, we tested multiple configurations of neural networks for the classification of the vectors representing sentences. In order to assess the performance of a model, cross-validation was used. This method can assess the generalization capabilities of a ML model, ensuring that it can make accurate predictions on unseen data and avoiding overfitting, which occurs when a model is too specific to the training data and performs poorly on new, real-world examples. k-fold cross-validation is a widely used technique for model evaluation and selection. In this method, the dataset is divided into k equally sized subsets or “folds”. The model is trained k times, each time using a different fold as the testing set and the remaining folds as the training set. This process helps to evaluate the model performance for different partitions of the data. By averaging the results from the k iterations, one can obtain a more reliable estimate of the model performance. k-fold cross-validation also provides a way to tune the hyperparameters and to assess generalization ability. In our experiments, we used k = 5.
The process of training neural networks involves addressing a considerable search space encompassing network architecture, e.g., the number of hidden layers, neurons, and activation functions, as well as hyperparameters such as the choice of the optimizer (the optimization algorithm), learning rate, and regularization details. The multitude of these variables results in an overwhelming number of potential configurations, rendering exhaustive exploration unfeasible. Consequently, employing a heuristic, empirical approach becomes imperative to efficiently identify the best available options within this vast parameter space.
We explored it by iterating through various configurations, initially favoring simpler ones and progressively trying more complex architectures and conducting multiple trials with parameter adjustments to estimate their impact.
For each configuration we repeated the training process five times and we computed the average accuracy values for the testing sets. Although differences exist among different runs because the data are randomly selected in the cross-validation folds, they were typically less than 1% between the minimum and maximum obtained values.
In Table 1, we provide a description of the network architecture and hyperparameters that were tested, the obtained accuracy values, and the primary motivation underlying the selection of each parameter combination.
In the column “Architecture and hyperparameters”, the network architecture is presented on the first line in the following format:
( number   of   inputs ) : ( number   of   neurons   in   hidden   layer   1 )   ( activation   function   of   the   neurons   in   hidden   layer   1 ) :   the   same   for   the   other   hidden   layers : ( number   of   outputs )   ( activation   function   of   the   neurons   in   the   output   layer )
The activation functions that we used are as follows: sigmoid (the unipolar sigmoid), tanh (the hyperbolic tangent), elu (exponential linear unit), and lrelu (leaky rectified linear unit).
Then, the main hyperparameters are mentioned on the second line, as follows: n is the number of training epochs, η is the learning rate, and opt represents the optimization algorithm.
For all configurations, the binary cross-entropy loss function was used, as it is better-suited to classification problems.
The best results were obtained with balanced configurations, i.e., those which are sufficiently complex to capture the underlying patterns of the data but without excessive complexity that may lead to overfitting. This equilibrium extends to both the architectural complexity of the network and the number of training epochs.
Furthermore, it is important to note that the best model varies between the different classes of skills (Tx) and, therefore, different neural models are used in each specific case.
Figure 4 shows the best cross-validation results for the six “level 1” models corresponding to the six main skill classes (Tx). One can see that the accuracy values for all models are quite high, above 94%, with the best one reaching 99%. We should emphasize that these values represent the averages obtained for the testing sets, not for the training sets. Therefore, we can conclude that the “level 1” classification models are well suited for the given task.

4.1.2. The Multi-Language Model

Our evaluation revealed that the performance results achieved using the multi-language model were remarkably close to those obtained when using the English-specific model. In this subsection we show the results for the “level 1” problem, but the hierarchical classification methodology designed here is independent of the specific sentence embedding model. The network architectures and hyperparameters that were tested for the multi-language sentence embeddings are presented in Table 2.
The results of the experiments, in terms of accuracy, are presented in Table 3. The “Trial” column corresponds to the configurations in Table 1 and Table 2. The rest of the columns show the accuracy values obtained for the testing sets in the cross-validation procedure. The best results for each skill class are highlighted in bold red. Still, one can see that several models can give comparably good results. Therefore, the results that are within a 0.2% range from the best one are also marked in bold italics.
As shown in Figure 4 and Figure 5, the multi-language model has very close accuracy values compared to the English language model. In Figure 5, we also include the relative difference between the two models computed with the following equation:
r d = a m a e 1
where am is the best accuracy obtained with the multi-language embeddings and ae is the best accuracy obtained with the English language embeddings.
One can see that the multi-language model is slightly better for the classes T1, T3, and T4 and slightly worse for the rest, but the accuracy values are all within a 0.55% range. Given the approximate nature of neural network-based classifications, we can conclude that the quality of the results obtained in the two situations is basically the same. However, the multi-language model has the advantage of flexibility and using it avoids the additional step of detecting the original language and performing an automatic translation that may even distort the original message to some extent. Therefore, it was selected as the default model to be used for the classification of skills and also for the results of the second level classification presented in the next subsection.

4.2. “Level 2” Classification

A similar approach was adopted for the “level 2” classification. However, “level 2” models are designed for multiple classes, as opposed to the binary classifiers used in the case of “level 1”. Thus, a variable, class-specific number of neurons was used in the output layer, denoted as no. Because the small number of training instances may lead to overfitting, we also tested the effect of regularization. The corresponding factor, denoted as λ, represents the regularization rate.
The evaluated neural network configurations are presented in Table 4.
Table 5 presents the results of the experiments for the second level of classification. The same conventions regarding colors and font styles as in Table 4 are used, as follows: bold red denotes the best results, while bold italics represent results within 0.2% proximity to the best. Unlike the results for the first level, the accuracy values for cross-validation are lower here. The results for the training sets are not included, but they were all above 90% and the general average across all trials and all skill subclasses was 95.5%.
The average training and testing results in terms of accuracy, for the k = 5 cross-validation folds are presented in Figure 6. Since this is a multi-label classification, the accuracy value for an instance represents the average of the accuracy values obtained for all outputs.
In this case, one can notice that the testing performance is not as good as for the “level 1” task. It is likely that the root causes are the low number of instances, presented in Table 6, and the ambiguity of the subclass membership.

5. Analysis and Discussion

The experimental results show that our models generally perform well in identifying transversal skills within job ads, considering the limited available data. In particular, the high accuracies of the “level 1” models allow the reliable screening of job ads that are lacking in phrases containing transversal skills, which we found to be in the vast majority. Identifying individual transversal skills requires data in an amount that can sufficiently cover the space of possible skill-related phrasings. Currently, our “level 2” models achieve high accuracies for the T1 category, which focuses on language skills. In this case, the phrasings are generally consistent (e.g., “Has good knowledge of [language]” or “A good grasp of [language] is beneficial”). High accuracy is also achieved for the T6 category, where, similarly, the phrasings found in job ads are more consistent than for other categories. In most cases, some of the best results were obtained using neural networks consisting of two hidden layers with descending sizes. We found that extending the architecture with additional layers of greater sizes did not improve the accuracy, instead resulting in overcomplicated models. However, especially in the case of “level 2” models, we surmise that a far greater improvement would be achieved with a broader dataset with better coverage of possible phrasings, than by tweaking our current models.

6. Conclusions

The present work was carried out in order to facilitate the identification of correlations between the phrasings from job advertisements and transversal skills, in order to predict the required skills from individual job descriptions. This is achieved by developing classification models according to the hierarchical skill base from the ESCO platform. The comparison between results obtained with English-specific and multi-language sentence embeddings reveals a comparable performance, validating the adaptability and efficiency of multi-language embeddings, subsequently adopted as the default model due to its flexibility. At the forefront of our approach lie the “level 1” and “level 2” hierarchical classifications, each showcasing high accuracy, yet “level 2” exhibits a lower cross-validation accuracy, attributed to subclass ambiguity and the limited number of instances.
The main contribution of this approach consists of developing a methodology based on neural networks for automating the identification of transversal skill requirements from job ad texts, considering the standardized skill set from the ESCO platform. The key innovations include an original dataset comprising manually labeled job ads where specific skills are referenced. The resulting data are used to train, validate, and test multiple neural network configurations, so that the resulting models can be reliably used for job ad classification. Furthermore, our classifiers account for the ESCO hierarchical structure, distinguishing between broader top-level skills and more specialized subskills.
We also acknowledge certain limitations of our work. The accuracy of our models and their capability for generalization are susceptible to improvement. Currently, only ESCO skills can be identified using our approach, since we train our models based on standard identifiers from the ESCO platform. In terms of classifier performance, so far, the main constraining factor has been the limited availability of data. Consequently, future work will take place on multiple fronts. We intend to improve subclass identification by developing higher-accuracy classifiers that can handle the ambiguities often found in job ad texts. We anticipate that employing advanced augmentation techniques may address the limitations of our current models and enhance their capability for generalization. Also, investigating multi-task learning approaches and enabling the simultaneous classification of multiple subclasses, could provide a better understanding and finer granularity in skill identification. Exploring domain adaptation techniques to address linguistic variations across job descriptions in different languages remains very important. Additionally, investigating interpretable models or explainable AI methodologies can aid in understanding model decisions, fostering trust and applicability in real-world scenarios. Collaborations with industry partners for larger datasets and validation in diverse job markets would further validate the efficacy of the proposed solution.

Author Contributions

F.L.: Conceptualization; Investigation; Methodology; Validation; Software; Writing—original draft; Writing—review and editing. M.G.: Investigation; Validation; Software; Writing—original draft; Writing—review and editing. S.-A.F.: Software. A.A.M.: Funding acquisition; Project administration; Writing—original draft; Writing—review and editing. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Horizon-Widera, grant number 101095292 and the APC was funded by authors.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data is unavailable due to privacy restrictions.

Acknowledgments

This work was supported by the DocTalent4EU project—Transforming Europe through Doctoral Talent and Skills Recognition, Horizon-Widera, GA no. 101095292, 2022–2024.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Zhou, W.; Zhu, Y.; Javed, F.; Rahman, M.; Balaji, J.; McNair, M. Quantifying skill relevance to job titles. In Proceedings of the 2016 IEEE International Conference on Big Data (Big Data), Washington, DC, USA, 5–8 December 2016; pp. 1532–1541. [Google Scholar]
  2. Papoutsoglou, M.; Mittas, N.; Angelis, L. Mining People Analytics from StackOverflow Job Advertisements. In Proceedings of the 2017 43rd Euromicro Conference on Software Engineering and Advanced Applications (SEAA), Vienna, Austria, 30 August–1 September 2017; pp. 108–115. [Google Scholar]
  3. Malherbe, E.; Aufaure, M.A. Bridge the terminology gap between recruiters and candidates: A multilingual skills base built from social media and linked data. In Proceedings of the 2016 IEEE/ACM International Conference on Advances in Social Networks Analysis and Mining (ASONAM), San Francisco, CA, USA, 18–21 August 2016; pp. 583–590. [Google Scholar]
  4. Sibarani, E.M.; Scerri, S.; Morales, C.; Auer, S.; Collarana, D. Ontology-guided Job Market Demand Analysis: A Cross-Sectional Study for the Data Science field. In Proceedings of the 13th International Conference on Semantic Systems, Ser. Semantics 2017, Amsterdam, The Netherlands, 11–14 September 2017; Association for Computing Machinery: New York, NY, USA, 2017; pp. 25–32. [Google Scholar]
  5. Gardiner, A.; Aasheim, C.; Rutner, P.; Williams, S. Skill Requirements in Big Data: A Content Analysis of Job Advertisements. J. Comput. Inf. Syst. 2018, 58, 374–384. [Google Scholar] [CrossRef]
  6. Chaibate, H.; Hadek, A.; Ajana, S.; Bakkali, S.; Faraj, K. Analyzing the engineering soft skills required by Moroccan job market. In Proceedings of the 2019 5th International Conference on Optimization and Applications (ICOA), Kenitra, Morocco, 25–26 April 2019; pp. 1–6. [Google Scholar]
  7. Niederman, F.; Sumner, M. Resolving the Skills Paradox: A Content Analysis of a Jobs Database. In Proceedings of the 2019 on Computers and People Research Conference, Nashville, TN, USA, 20–22 June 2019; pp. 164–167. [Google Scholar]
  8. Rios, J.A.; Ling, G.; Pugh, R.; Becker, D.; Bacall, A. Identifying Critical 21st-Century Skills for Workplace Success: A Content Analysis of Job Advertisements. Educ. Res. 2020, 49, 80–89. [Google Scholar] [CrossRef]
  9. Debortoli, S.; Maller, O.; Brocke, J.V. Comparing Business Intelligence and Big Data Skills. Bus. Inf. Syst. Eng. 2014, 6, 289–300. [Google Scholar] [CrossRef]
  10. De Mauro, A.; Greco, M.; Grimaldi, M.; Ritala, P. Human resources for Big Data professions: A systematic classification of job roles and required skill sets. Inf. Process. Manag. 2018, 54, 807–817. [Google Scholar] [CrossRef]
  11. Gurcan, F.; Sevik, S. Expertise Roles and Skills Required by the Software Development Industry. In Proceedings of the 2019 1st International Informatics and Software Engineering Conference (UBMYK), Ankara, Turkey, 6–7 November 2019; pp. 1–4. [Google Scholar]
  12. Gurcan, F.; Cagiltay, N.E. Big Data Software Engineering: Analysis of Knowledge Domains and Skill Sets Using LDA-Based Topic Modeling. IEEE Access 2019, 7, 82541–82552. [Google Scholar] [CrossRef]
  13. Jelodar, H.; Wang, Y.; Yuan, C.; Feng, X.; Jiang, X.; Li, Y.; Zhao, L. Latent Dirichlet allocation (LDA) and topic modeling: Models, applications, a survey. Multimed. Tools Appl. 2019, 78, 15169–15211. [Google Scholar] [CrossRef]
  14. Javed, F.; Hoang, P.; Mahoney, T.; McNair, M. Large-scale occupational skills normalization for online recruitment. In Proceedings of the Twenty-Ninth AAAI Conference, Austin, TX, USA, 25–30 January 2015. [Google Scholar]
  15. Zhao, M.; Javed, F.; Jacob, F.; McNair, M. SKILL: A System for Skill Identification and Normalization. In Proceedings of the Twenty-Ninth AAAI Conference on Artificial Intelligence, Austin, TX, USA, 25–30 January 2015; pp. 4012–4017. [Google Scholar]
  16. Mikolov, T.; Sutskever, I.; Chen, K.; Corrado, G.S.; Dean, J. Distributed representations of words and phrases and their compositionality. Adv. Neural Inf. Process. Syst. 2013, 26, 3111–3119. [Google Scholar]
  17. Botov, D.; Klenin, J.; Melnikov, A.; Dmitrin, Y.; Nikolaev, I.; Vinel, M. Mining Labor Market Requirements Using Distributional Semantic Models and Deep Learning. In Business Information Systems, Ser. Lecture Notes in Business Information Processing; Abramowicz, W., Corchuelo, R., Eds.; Springer International Publishing: Cham, Switzerland, 2019; pp. 177–190. [Google Scholar]
  18. Gugnani, A.; Misra, H. Implicit Skills Extraction Using Document Embedding and Its Use in Job Recommendation. AAAI 2020, 34, 13286–13293. [Google Scholar] [CrossRef]
  19. Le, Q.; Mikolov, T. Distributed representations of sentences and documents. In Proceedings of the International Conference on Machine Learning, PMLR, Baltimore, MD, USA, 17–23 July 2022; pp. 1188–1196. [Google Scholar]
  20. Van-Duyet, L.; Quan, V.M.; An, D.Q. Skill2vec: Machine Learning Approach for Determining the Relevant Skills from Job Description. arXiv 2019, arXiv:1707.09751. [Google Scholar]
  21. Li, S.; Shi, B.; Yang, J.; Yan, J.; Wang, S.; Chen, F.; He, Q. Deep Job Understanding at LinkedIn. In Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval, Ser. SIGIR ’20, Xi’an, China, 25–30 July 2020; Association for Computing Machinery: New York, NY, USA, 2020; pp. 2145–2148. [Google Scholar]
  22. Bojanowski, P.; Grave, E.; Joulin, A.; Mikolov, T. Enriching Word Vectors with Subword Information. Trans. Assoc. Comput. Linguist. 2017, 5, 135–146. [Google Scholar] [CrossRef]
  23. Fagerbakk, A.G. Keeping Up with the Market: Extracting Competencies from Norwegian Job Listings. Master’s Thesis, The Arctic University of Norway, Tromsø, Norway, 2021. [Google Scholar]
  24. Jia, S.; Liu, X.; Zhao, P.; Liu, C.; Sun, L.; Peng, T. Representation of Job-Skill in Artificial Intelligence with Knowledge Graph Analysis. In Proceedings of the 2018 IEEE Symposium on Product Compliance Engineering—Asia (ISPCE-CN), Shenzhen, China, 5–7 December 2018; pp. 1–6. [Google Scholar]
  25. Lample, G.; Ballesteros, M.; Subramanian, S.; Kawakami, K.; Dyer, C. Neural Architectures for Named Entity Recognition. In Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, San Diego, CA, USA, 12–17 June 2016; pp. 260–270. [Google Scholar]
  26. Sayfullina, L.; Malmi, E.; Kannala, J. Learning Representations for Soft Skill Matching. In Analysis of Images, Social Networks and Texts, Lecture Notes in Computer Science; Springer International Publishing: Berlin/Heidelberg, Germany, 2018; pp. 141–152. [Google Scholar]
  27. Kim, Y. Convolutional Neural Networks for Sentence Classification. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), Doha, Qatar, 25–29 October 2014; pp. 1746–1751. [Google Scholar]
  28. Tamburri, D.A.; Heuvel, W.-J.V.D.; Garriga, M. DataOps for Societal Intelligence: A Data Pipeline for Labor Market Skills Extraction and Matching. In Proceedings of the 2020 IEEE 21st International Conference on Information Reuse and Integration for Data Science (IRI), Las Vegas, NV, USA, 11–13 August 2020; pp. 391–394. [Google Scholar]
  29. Devlin, J.; Chang, M.-W.; Lee, K.; Toutanova, K. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. arXiv 2019, arXiv:1810.04805. [Google Scholar]
  30. Lukauskas, M.; Šarkauskaitė, V.; Pilinkienė, V.; Stundžienė, A.; Grybauskas, A.; Bruneckienė, J. Enhancing Skills Demand Understanding through Job Ad Segmentation Using NLP and Clustering Techniques. Appl. Sci. 2023, 13, 6119. [Google Scholar] [CrossRef]
  31. Mathiasen, M.; Nielsen, J.; Laub, S. A Transformer Based Semantic Analysis of (non-English) Danish Job ads. In Proceedings of the 15th International Conference on Computer Supported Education, Prague, Czech Republic, 21–23 April 2023. [Google Scholar]
  32. Reimers, N.; Gurevych, I. Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), Hong Kong, China, 3–7 November 2019. [Google Scholar]
  33. Bhola, A.; Halder, K.; Prasad, A.; Kan, M.-Y. Retrieving Skills from Job Descriptions: A Language Model Based Extreme Multi-label Classification Framework. In Proceedings of the 28th International Conference on Computational Linguistics, Online, 8–13 December 2020; pp. 5832–5842. [Google Scholar]
  34. Cedefop. Online Job Vacancies and Skills Analysis: A Cedefop Pan-European Approach; Publications Office: Luxembourg, 2019; Available online: http://data.europa.eu/doi/10.2801/097022 (accessed on 5 January 2024).
  35. Applegate, R. Job ads, jobs, and researchers: Searching for valid sources. Libr. Inf. Sci. Res. 2010, 32, 163–170. [Google Scholar] [CrossRef]
  36. Papoutsoglou, M.; Ampatzoglou, A.; Mittas, N.; Angelis, L. Extracting Knowledge from On-Line Sources for Software Engineering Labor Market: A Mapping Study. IEEE Access 2019, 7, 157595–157613. [Google Scholar] [CrossRef]
  37. Khaouja, I.; Kassou, I.; Ghogho, M. A Survey on Skill Identification from Online Job Ads. IEEE Access 2021, 9, 118134–118153. [Google Scholar] [CrossRef]
  38. European Commission, Directorate-General for Employment, Social Affairs and Inclusion, EURES. 2023. Available online: https://eures.ec.europa.eu/index_en (accessed on 5 January 2024).
  39. European Commission, Directorate-General for Employment, Social Affairs and Inclusion, ESCO Publications: Skills & Competences. 2022. Available online: https://esco.ec.europa.eu/en/classification/skill_main (accessed on 5 January 2024).
  40. Reimers, N.; Gurevych, I. Making Monolingual Sentence Embeddings Multilingual using Knowledge Distillation. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing, Online, 16–20 November 2020; Association for Computational Linguistics: Kerrville, TX, USA, 2020. [Google Scholar]
  41. Zhang, J.; Zhao, Y.; Saleh, M.; Liu, P.J. PEGASUS: Pre-Training with Extracted Gap-Sentences for Abstractive Summarization. 2020. Available online: https://github.com/google-research/pegasus (accessed on 5 January 2024).
  42. Adarsh, A. Pegasus-Paraphrase Library. 2022. Available online: https://github.com/adarshgowdaa/pegasus-paraphrase (accessed on 5 January 2024).
Figure 1. Overview of the proposed methodology.
Figure 1. Overview of the proposed methodology.
Information 15 00151 g001
Figure 2. A sample result of data preprocessing.
Figure 2. A sample result of data preprocessing.
Information 15 00151 g002
Figure 3. A sample result of data preprocessing for the “level 2” classification.
Figure 3. A sample result of data preprocessing for the “level 2” classification.
Information 15 00151 g003
Figure 4. Comparison between the performance of classification using English language and multi-language sentence embeddings.
Figure 4. Comparison between the performance of classification using English language and multi-language sentence embeddings.
Information 15 00151 g004
Figure 5. Relative difference obtained by the multi-language model compared to the English language model.
Figure 5. Relative difference obtained by the multi-language model compared to the English language model.
Information 15 00151 g005
Figure 6. The results obtained for the “level 2” classification.
Figure 6. The results obtained for the “level 2” classification.
Information 15 00151 g006
Table 1. Neural network configurations evaluated for the English dataset.
Table 1. Neural network configurations evaluated for the English dataset.
No. TrialArchitecture and HyperparametersMotivation
1768: 128 (elu): 1 (sigmoid)
n = 1000, η = 0.001, opt = Adam
Simple architecture (one hidden layer)
2768: 1536 (tanh): 512 (tanh): 128 (tanh):
32 (tanh): 8 (tanh): 1 (sigmoid)
n = 1000, η = 0.001, opt = Adam
Complex architecture (five hidden layers)
3768: 20 (lrelu): 4 (lrelu): 1 (sigmoid)
n = 1000, η = 0.001, opt = Adam
Medium size architecture (two hidden layers), strong information compression in the first hidden layer
4768: 81 (lrelu): 9 (lrelu): 1 (sigmoid)
n = 1000, η = 0.001, opt = Adam
Medium size architecture (two hidden layers), balanced information compression
5768: 81 (sigmoid): 9 (sigmoid): 1 (sigmoid)
n = 1000, η = 0.001, opt = Adam
Sigmoid activation function in the hidden layers
6768: 81 (tanh): 9 (tanh): 1 (sigmoid)
n = 1000, η = 0.001, opt = Adam
Hyperbolic tangent activation function in the hidden layers
7768: 81 (lrelu): 9 (lrelu): 1 (sigmoid)
n = 1000, η = 0.001, opt = RMSprop
Leaky ReLU activation function in the hidden layers and the RMSprop optimizer
8768: 81 (lrelu): 9 (lrelu): 1 (sigmoid)
n = 1000, η = 0.01, opt = Adam
Higher learning rate (0.01)
9768: 81 (lrelu): 9 (lrelu): 1 (sigmoid)
n = 10,000, η = 0.01, opt = Adam
Large number of training epochs (10,000)
10768: 81 (lrelu): 9 (lrelu): 1 (sigmoid)
n = 100, η = 0.001, opt = Adam
Small number of training epochs (100)
Table 2. Neural network configurations evaluated for the multi-language dataset.
Table 2. Neural network configurations evaluated for the multi-language dataset.
No. TrialArchitecture and HyperparametersMotivation
1768: 128 (elu): 1 (sigmoid)
n = 1000, η = 0.001, opt = Adam
Simple architecture (one hidden layer)
2768: 128 (elu): 1 (sigmoid)
n = 1000, η = 0.01, opt = Adam
Higher learning rate (0.01)
3768: 128 (elu): 1 (sigmoid)
n = 1000, η = 0.01, opt = RMSprop
RMSprop optimizer
4768: 128 (elu): 1 (sigmoid)
n = 10,000, η = 0.01, opt = Adam
Large number of training epochs (10,000)
5768: 128 (lrelu): 1 (sigmoid)
n = 1000, η = 0.01, opt = Adam
Leaky ReLU activation function in the hidden layer
6768: 81 (lrelu): 9 (lrelu): 1 (sigmoid)
n = 1000, η = 0.01, opt = Adam
Medium size architecture (two hidden layers), balanced information compression
7768: 81 (lrelu): 9 (lrelu): 1 (sigmoid)
n = 1000, η = 0.01, opt = RMSprop
RMSprop optimizer
Table 3. Accuracy values obtained for the evaluated neural network configurations for the first level classification.
Table 3. Accuracy values obtained for the evaluated neural network configurations for the first level classification.
English Language Model
TrialT1T2T3T4T5T6
196.9195.1692.7093.6899.3396.14
295.9393.5591.7493.2698.5295.14
397.2094.8793.0394.1899.3796.76
497.5095.8593.5994.6099.5297.25
597.1094.6891.9993.7699.1795.62
696.9194.6892.3493.2899.2995.80
797.4595.9193.5994.7899.5497.33
897.4896.0493.7494.8499.5097.14
997.4896.2494.0794.4399.5897.43
1095.3590.0089.5589.1997.8191.03
Multi-Language Model
TrialT1T2T3T4T5T6
197.1494.7192.9293.2698.9396.01
297.5489.1193.6694.0899.0396.95
397.7995.6193.8094.6999.2096.70
497.5695.8693.9794.6799.0496.97
597.8296.0193.8994.8399.2797.23
698.0295.8294.1294.9099.2497.08
797.9895.9794.0494.5699.3997.02
Table 4. Neural network configurations evaluated for the second level classification.
Table 4. Neural network configurations evaluated for the second level classification.
No. TrialArchitecture and HyperparametersMotivation
1768: 128 (lrelu): no (sigmoid)
n = 100, η = 0.001, opt = Adam
Simple architecture (one hidden layer)
2768: 128 (lrelu): no (sigmoid)
n = 100, η = 0.01, opt = Adam
Higher learning rate (0.01)
3768: 150 (lrelu): 30 (lrelu): no (sigmoid)
n = 100, η = 0.01, opt = Adam
Larger architecture (two hidden layers)
4768: 128 (lrelu): no (sigmoid)
n = 100, η = 0.01, λ = 10−5, opt = Adam
Regularization
5768: 128 (lrelu): no (sigmoid)
n = 1000, η = 0.01, λ = 10−5, opt = Adam
Larger number of training epochs (1000)
6768: 128 (tanh): no (sigmoid)
n = 100, η = 0.01, λ = 10−5, opt = Adam
Hyperbolic tangent activation function in the hidden layer
7768: 128 (lrelu): no (sigmoid)
n = 50, η = 0.01, λ = 10−5, opt = Adam
Smaller number of training epochs (50)
8768: 128 (elu): no (sigmoid)
n = 100, η = 0.01, λ = 10−5, opt = Adam
Exponential Linear Unit (ELU) activation function in the hidden layer
9768: 128 (elu): no (sigmoid)
n = 100, η = 0.01, λ = 10−5, opt = RMSprop
RMSprop optimizer
10768: 150 (sigmoid): 30 (sigmoid): no (sigmoid)
n = 100, η = 0.01, λ = 10−5, opt = Adam
Larger architecture (two hidden layers) with sigmoid activation function in the hidden layers
Table 5. Accuracy values obtained for the evaluated neural network configurations for the second level classification.
Table 5. Accuracy values obtained for the evaluated neural network configurations for the second level classification.
TrialT1T2T3T4T5T6
191.4666.9372.2374.4976.2081.55
292.9567.6973.0976.0577.4782.51
388.8165.4169.8372.9173.9178.61
495.8968.1773.0575.2375.6784.82
592.8265.4769.6173.3477.1381.07
696.2267.5372.4476.0580.0584.32
793.1965.6370.4073.3477.6881.92
891.7165.0969.3472.6776.8781.03
995.0266.9871.4674.7278.7983.02
1096.3567.8572.1676.2176.7583.10
Table 6. The total number of instances in each subclass.
Table 6. The total number of instances in each subclass.
T1T2T3T4T5T6
T1.1: 106T2.1: 99T3.1: 186T4.1: 169T5.1: 8T6.1: 8
T1.2: 7T2.2: 101T3.2: 176T4.2: 181T5.2: 13T6.2: 10
T1.3: 29T2.3: 137T3.3: 94T4.3: 209 T6.3: 15
T2.4: 107T3.4: 122T4.4: 142 T6.4: 31
T4.5: 54 T6.5: 32
T6.6: 33
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Leon, F.; Gavrilescu, M.; Floria, S.-A.; Minea, A.A. Hierarchical Classification of Transversal Skills in Job Advertisements Based on Sentence Embeddings. Information 2024, 15, 151. https://doi.org/10.3390/info15030151

AMA Style

Leon F, Gavrilescu M, Floria S-A, Minea AA. Hierarchical Classification of Transversal Skills in Job Advertisements Based on Sentence Embeddings. Information. 2024; 15(3):151. https://doi.org/10.3390/info15030151

Chicago/Turabian Style

Leon, Florin, Marius Gavrilescu, Sabina-Adriana Floria, and Alina Adriana Minea. 2024. "Hierarchical Classification of Transversal Skills in Job Advertisements Based on Sentence Embeddings" Information 15, no. 3: 151. https://doi.org/10.3390/info15030151

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop