Leveraging the Inherent Hierarchy of Vacancy Titles for Automated Job
Ontology Expansion
- URL: http://arxiv.org/abs/2004.02814v1
- Date: Mon, 6 Apr 2020 16:55:41 GMT
- Title: Leveraging the Inherent Hierarchy of Vacancy Titles for Automated Job
Ontology Expansion
- Authors: Jeroen Van Hautte, Vincent Schelstraete, Mika\"el Wornoo
- Abstract summary: We introduce a novel, purely data-driven approach towards the detection of new job titles.
Our method is conceptually simple, extremely efficient and competitive with traditional NER-based approaches.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Machine learning plays an ever-bigger part in online recruitment, powering
intelligent matchmaking and job recommendations across many of the world's
largest job platforms. However, the main text is rarely enough to fully
understand a job posting: more often than not, much of the required information
is condensed into the job title. Several organised efforts have been made to
map job titles onto a hand-made knowledge base as to provide this information,
but these only cover around 60\% of online vacancies. We introduce a novel,
purely data-driven approach towards the detection of new job titles. Our method
is conceptually simple, extremely efficient and competitive with traditional
NER-based approaches. Although the standalone application of our method does
not outperform a finetuned BERT model, it can be applied as a preprocessing
step as well, substantially boosting accuracy across several architectures.
Related papers
- Towards Efficient Resume Understanding: A Multi-Granularity Multi-Modal Pre-Training Approach [43.0592922448035]
We propose a novel model, namely ERU, to achieve efficient resume understanding.
We first introduce a layout-aware multi-modal fusion transformer for encoding the segments in the resume with integrated textual, visual, and layout information.
Next, we fine-tune the model with a multi-granularity sequence labeling task to extract structured information from resumes.
arXiv Detail & Related papers (2024-04-13T14:31:24Z) - Distribution Matching for Multi-Task Learning of Classification Tasks: a
Large-Scale Study on Faces & Beyond [62.406687088097605]
Multi-Task Learning (MTL) is a framework, where multiple related tasks are learned jointly and benefit from a shared representation space.
We show that MTL can be successful with classification tasks with little, or non-overlapping annotations.
We propose a novel approach, where knowledge exchange is enabled between the tasks via distribution matching.
arXiv Detail & Related papers (2024-01-02T14:18:11Z) - Extreme Multi-Label Skill Extraction Training using Large Language
Models [19.095612333241288]
We describe a cost-effective approach to generate an accurate, fully synthetic labeled dataset for skill extraction.
Our results show a consistent increase of between 15 to 25 percentage points in textitR-Precision@5 compared to previously published results.
arXiv Detail & Related papers (2023-07-20T11:29:15Z) - Pre-training Multi-task Contrastive Learning Models for Scientific
Literature Understanding [52.723297744257536]
Pre-trained language models (LMs) have shown effectiveness in scientific literature understanding tasks.
We propose a multi-task contrastive learning framework, SciMult, to facilitate common knowledge sharing across different literature understanding tasks.
arXiv Detail & Related papers (2023-05-23T16:47:22Z) - Learning Job Titles Similarity from Noisy Skill Labels [0.11498015270151059]
Measuring semantic similarity between job titles is an essential functionality for automatic job recommendations.
In this paper, we propose an unsupervised representation learning method for training a job title similarity model using noisy skill labels.
arXiv Detail & Related papers (2022-07-01T15:30:10Z) - Hierarchical Skills for Efficient Exploration [70.62309286348057]
In reinforcement learning, pre-trained low-level skills have the potential to greatly facilitate exploration.
Prior knowledge of the downstream task is required to strike the right balance between generality (fine-grained control) and specificity (faster learning) in skill design.
We propose a hierarchical skill learning framework that acquires skills of varying complexity in an unsupervised manner.
arXiv Detail & Related papers (2021-10-20T22:29:32Z) - JobBERT: Understanding Job Titles through Skills [12.569546741576515]
Job titles form a cornerstone of today's human resources (HR) processes.
Job titles are a compact, convenient, and readily available data source.
We propose a neural representation model for job titles, by augmenting a pre-trained language model with co-occurrence information from skill labels extracted from vacancies.
arXiv Detail & Related papers (2021-09-20T15:00:10Z) - DAGA: Data Augmentation with a Generation Approach for Low-resource
Tagging Tasks [88.62288327934499]
We propose a novel augmentation method with language models trained on the linearized labeled sentences.
Our method is applicable to both supervised and semi-supervised settings.
arXiv Detail & Related papers (2020-11-03T07:49:15Z) - COG: Connecting New Skills to Past Experience with Offline Reinforcement
Learning [78.13740204156858]
We show that we can reuse prior data to extend new skills simply through dynamic programming.
We demonstrate the effectiveness of our approach by chaining together several behaviors seen in prior datasets for solving a new task.
We train our policies in an end-to-end fashion, mapping high-dimensional image observations to low-level robot control commands.
arXiv Detail & Related papers (2020-10-27T17:57:29Z) - Job2Vec: Job Title Benchmarking with Collective Multi-View
Representation Learning [51.34011135329063]
Job Title Benchmarking (JTB) aims at matching job titles with similar expertise levels across various companies.
Traditional JTB approaches mainly rely on manual market surveys, which is expensive and labor-intensive.
We reformulate the JTB as the task of link prediction over the Job-Graph that matched job titles should have links.
arXiv Detail & Related papers (2020-09-16T02:33:32Z) - Transfer Learning for Information Extraction with Limited Data [2.201264358342234]
This paper presents a practical approach to fine-grained information extraction.
We first exploit BERT to deal with the limitation of training data in real scenarios.
We then stack BERT with Convolutional Neural Networks to learn hidden representation for classification.
arXiv Detail & Related papers (2020-03-06T08:08:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.