Learning Effective Representations for Person-Job Fit by Feature Fusion
- URL: http://arxiv.org/abs/2006.07017v1
- Date: Fri, 12 Jun 2020 09:02:41 GMT
- Title: Learning Effective Representations for Person-Job Fit by Feature Fusion
- Authors: Junshu Jiang and Songyun Ye and Wei Wang and Jingran Xu and Xiaosheng
Luo
- Abstract summary: Person-job fit is to match candidates and job posts on online recruitment platforms using machine learning algorithms.
In this paper, we propose to learn comprehensive and effective representations of the candidates and job posts via feature fusion.
Experiments over 10 months real data show that our solution outperforms existing methods with a large margin.
- Score: 4.884826427985207
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Person-job fit is to match candidates and job posts on online recruitment
platforms using machine learning algorithms. The effectiveness of matching
algorithms heavily depends on the learned representations for the candidates
and job posts. In this paper, we propose to learn comprehensive and effective
representations of the candidates and job posts via feature fusion. First, in
addition to applying deep learning models for processing the free text in
resumes and job posts, which is adopted by existing methods, we extract
semantic entities from the whole resume (and job post) and then learn features
for them. By fusing the features from the free text and the entities, we get a
comprehensive representation for the information explicitly stated in the
resume and job post. Second, however, some information of a candidate or a job
may not be explicitly captured in the resume or job post. Nonetheless, the
historical applications including accepted and rejected cases can reveal some
implicit intentions of the candidates or recruiters. Therefore, we propose to
learn the representations of implicit intentions by processing the historical
applications using LSTM. Last, by fusing the representations for the explicit
and implicit intentions, we get a more comprehensive and effective
representation for person-job fit. Experiments over 10 months real data show
that our solution outperforms existing methods with a large margin. Ablation
studies confirm the contribution of each component of the fused representation.
The extracted semantic entities help interpret the matching results during the
case study.
Related papers
- Likelihood as a Performance Gauge for Retrieval-Augmented Generation [78.28197013467157]
We show that likelihoods serve as an effective gauge for language model performance.
We propose two methods that use question likelihood as a gauge for selecting and constructing prompts that lead to better performance.
arXiv Detail & Related papers (2024-11-12T13:14:09Z) - Web-Scale Visual Entity Recognition: An LLM-Driven Data Approach [56.55633052479446]
Web-scale visual entity recognition presents significant challenges due to the lack of clean, large-scale training data.
We propose a novel methodology to curate such a dataset, leveraging a multimodal large language model (LLM) for label verification, metadata generation, and rationale explanation.
Experiments demonstrate that models trained on this automatically curated data achieve state-of-the-art performance on web-scale visual entity recognition tasks.
arXiv Detail & Related papers (2024-10-31T06:55:24Z) - Prompt-based Personality Profiling: Reinforcement Learning for Relevance Filtering [8.20929362102942]
Author profiling is the task of inferring characteristics about individuals by analyzing content they share.
We propose a new method for author profiling which aims at distinguishing relevant from irrelevant content first, followed by the actual user profiling only with relevant data.
We evaluate our method for Big Five personality trait prediction on two Twitter corpora.
arXiv Detail & Related papers (2024-09-06T08:43:10Z) - Combining Embeddings and Domain Knowledge for Job Posting Duplicate Detection [42.49221181099313]
Job descriptions are posted on many online channels, including company websites, job boards or social media platforms.
It is helpful to aggregate job postings across platforms and thus detect duplicate descriptions that refer to the same job.
We show that combining overlap-based character similarity with text embedding and keyword matching methods lead to convincing results.
arXiv Detail & Related papers (2024-06-10T13:38:15Z) - TAROT: A Hierarchical Framework with Multitask Co-Pretraining on
Semi-Structured Data towards Effective Person-Job Fit [60.31175803899285]
We propose TAROT, a hierarchical multitask co-pretraining framework, to better utilize structural and semantic information for informative text embeddings.
TAROT targets semi-structured text in profiles and jobs, and it is co-pretained with multi-grained pretraining tasks to constrain the acquired semantic information at each level.
arXiv Detail & Related papers (2024-01-15T07:57:58Z) - Distribution Matching for Multi-Task Learning of Classification Tasks: a
Large-Scale Study on Faces & Beyond [62.406687088097605]
Multi-Task Learning (MTL) is a framework, where multiple related tasks are learned jointly and benefit from a shared representation space.
We show that MTL can be successful with classification tasks with little, or non-overlapping annotations.
We propose a novel approach, where knowledge exchange is enabled between the tasks via distribution matching.
arXiv Detail & Related papers (2024-01-02T14:18:11Z) - Exploring Effective Factors for Improving Visual In-Context Learning [56.14208975380607]
In-Context Learning (ICL) is to understand a new task via a few demonstrations (aka. prompt) and predict new inputs without tuning the models.
This paper shows that prompt selection and prompt fusion are two major factors that have a direct impact on the inference performance of visual context learning.
We propose a simple framework prompt-SelF for visual in-context learning.
arXiv Detail & Related papers (2023-04-10T17:59:04Z) - On the Use of External Data for Spoken Named Entity Recognition [40.93448412171246]
Recent advances in self-supervised speech representations have made it feasible to consider learning models with limited labeled data.
We draw on a variety of approaches, including self-training, knowledge distillation, and transfer learning, and consider their applicability to both end-to-end models and pipeline approaches.
arXiv Detail & Related papers (2021-12-14T18:49:26Z) - DSSL: Deep Surroundings-person Separation Learning for Text-based Person
Retrieval [40.70100506088116]
We propose a novel Deep Surroundings-person Separation Learning (DSSL) model in this paper.
A surroundings-person separation and fusion mechanism plays the key role to realize an accurate and effective surroundings-person separation.
Extensive experiments are carried out to evaluate the proposed DSSL on CUHK-PEDES.
arXiv Detail & Related papers (2021-09-12T15:09:09Z) - Pretext Tasks selection for multitask self-supervised speech
representation learning [23.39079406674442]
This paper introduces a method to select a group of pretext tasks among a set of candidates.
Experiments conducted on speaker recognition and automatic speech recognition validate our approach.
arXiv Detail & Related papers (2021-07-01T16:36:29Z) - Learning to Match Jobs with Resumes from Sparse Interaction Data using
Multi-View Co-Teaching Network [83.64416937454801]
Job-resume interaction data is sparse and noisy, which affects the performance of job-resume match algorithms.
We propose a novel multi-view co-teaching network from sparse interaction data for job-resume matching.
Our model is able to outperform state-of-the-art methods for job-resume matching.
arXiv Detail & Related papers (2020-09-25T03:09:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.