Edit model card

PubMLP: Automatic Publication Classifier

Hugging Face | GitHub | Documentation

This model is a fine-tuned version of bert-base-uncased on bibliometric data exported from the Web of Science (WOS). The tabular data include text, categorical data, and numeric features. The utilized functions include the BERT (Bidirectional Encoder Representations from Transformers) embeddings and an MLP for classifying whether the publication meets the researcher-designated inclusion criteria.

Model description

PubMLP is a PyTorch module for a Multilayer Perceptron (MLP) model with BERT embeddings.

Pipeline

The project consists of several components executed sequentially for the classification of publications:

  • Prepare Data (prepare.py): Extract country entities and split data.
  • Preprocess Data (preprocess.py): Tokenize combined text columns and normalize numeric data.
  • Train and Evaluate the Model (model.py): Train and evaluate the MLP model using the preprocessed data.
  • Plot Results (plot.py): Visualize results.

Training and evaluation data

The dataset includes a subset of WOS bibliometric data. The data are based on publications between 1970 and 2023 and focus on single-case research.

Framework versions

  • pandas==2.1.2
  • numpy>=1.21,<1.25
  • spacy==3.7.2
  • torch==2.1.0
  • scikit-learn==1.3.2
  • transformers>=4.36.0
  • matplotlib==3.8.1
Downloads last month
0
Unable to determine this model’s pipeline type. Check the docs .