Self-supervised fine-tuning for neural expert finding

Publication Type

Conference Proceeding Article

Publication Date

12-2024

Abstract

Expert finding systems allow ones to find individuals who have expertise in specific fields or domains. Traditional expert finding are mostly based on topic modeling or keyword search methods that are limited in their capability to encode contextual knowledge from natural language. To address the limitation, this paper presents Neural Expert Finder (NEF), a novel method that takes a transfer learning approach based on transformer encoder networks to leverage the rich semantic and syntactic patterns of language encoded in pre-trained language models (PLMs). We propose a self-supervised learning approach utilizing contrastive training using both positive and automatically generated negative samples to fine-tune the PLMs to realize NEF. In addition, we also contribute a new benchmark data set for expert finding named SGComp, curated from experts’ university and Google Scholar profiles. Our empirical evaluations demonstrate that the proposed method can effectively capture contextual representations and improve the retrieval of experts most relevant to their corresponding research areas. Both SGComp and three domain specific public data sets are utilized to compare NEF against ExpFinder nVSM, a state-of-the-art (SOTA) system in expert finding, and the results demonstrate consistent better performance of the proposed NEF.

Keywords

Expert finding systems, Transfer learning, Transformer encoder networks, Natural language processing

Discipline

Artificial Intelligence and Robotics | Computer Sciences

Research Areas

Data Science and Engineering; Intelligent Systems and Optimization

Publication

Proceedings of the 23rd IEEE/WIC International Conference on Web Intelligence and Intelligent Agent Technology (WI-IAT 2024) : Bangkok, Thailand, December 9-12

First Page

3142

Last Page

3146

Identifier

10.1145/3459637.3482179

Publisher

IEEE

City or Country

Bangkok, Thailand

Comments

PDF provided by faculty

Additional URL

https://doi.org/10.1145/3459637.3482179

This document is currently not available here.

Share

COinS