Papers
arxiv:2406.18125

ResumeAtlas: Revisiting Resume Classification with Large-Scale Datasets and Large Language Models

Published on Jun 26
· Submitted by ahmedheakl on Jun 28
Authors:
,
,

Abstract

The increasing reliance on online recruitment platforms coupled with the adoption of AI technologies has highlighted the critical need for efficient resume classification methods. However, challenges such as small datasets, lack of standardized resume templates, and privacy concerns hinder the accuracy and effectiveness of existing classification models. In this work, we address these challenges by presenting a comprehensive approach to resume classification. We curated a large-scale dataset of 13,389 resumes from diverse sources and employed Large Language Models (LLMs) such as BERT and Gemma1.1 2B for classification. Our results demonstrate significant improvements over traditional machine learning approaches, with our best model achieving a top-1 accuracy of 92\% and a top-5 accuracy of 97.5\%. These findings underscore the importance of dataset quality and advanced model architectures in enhancing the accuracy and robustness of resume classification systems, thus advancing the field of online recruitment practices.

Community

Paper author Paper submitter
edited 2 days ago

This work addresses challenges in resume classification for online recruitment by curating a large-scale dataset of 13,389 resumes and employing Large Language Models (LLMs) such as BERT and Gemma1.1 2B. Our approach significantly improves accuracy, achieving a top-1 accuracy of 92% and a top-5 accuracy of 97.5%, highlighting the importance of high-quality datasets and advanced model architectures. These results advance the effectiveness and robustness of resume classification systems in online recruitment. We open-source our codes and dataset.

Sign up or log in to comment

Models citing this paper 0

No model linking this paper

Cite arxiv.org/abs/2406.18125 in a model README.md to link it from this page.

Datasets citing this paper 0

No dataset linking this paper

Cite arxiv.org/abs/2406.18125 in a dataset README.md to link it from this page.

Spaces citing this paper 0

No Space linking this paper

Cite arxiv.org/abs/2406.18125 in a Space README.md to link it from this page.

Collections including this paper 0

No Collection including this paper

Add this paper to a collection to link it from this page.