fin-perceiver / README.md
Tomás Fernandes
Update model card metadata and W&B
1ee8539
---
language: "en"
license: apache-2.0
tags:
- financial-sentiment-analysis
- sentiment-analysis
- language-perceiver
datasets:
- financial_phrasebank
widget:
- text: "INDEX100 fell sharply today."
- text: "ImaginaryJetCo bookings hit by Omicron variant as losses total £1bn."
- text: "Q1 ImaginaryGame's earnings beat expectations."
- text: "Should we buy IMAGINARYSTOCK today?"
metrics:
- recall
- f1
- accuracy
- precision
model-index:
- name: fin-perceiver
results:
- task:
name: Text Classification
type: text-classification
dataset:
name: financial_phrasebank
type: financial_phrasebank
args: sentences_50agree
metrics:
- name: Accuracy
type: accuracy
value: 0.8624
- name: F1
type: f1
value: 0.8416
args: macro
- name: Precision
type: precision
value: 0.8438
args: macro
- name: Recall
type: recall
value: 0.8415
args: macro
---
# FINPerceiver
FINPerceiver is a fine-tuned Perceiver IO language model for financial sentiment analysis.
More details on the training process of this model are available on the [GitHub repository](https://github.com/warwickai/fin-perceiver).
Weights & Biases was used to track experiments.
We achieved the following results with 10-fold cross validation.
```
eval/accuracy 0.8624 (stdev 0.01922)
eval/f1 0.8416 (stdev 0.03738)
eval/loss 0.4314 (stdev 0.05295)
eval/precision 0.8438 (stdev 0.02938)
eval/recall 0.8415 (stdev 0.04458)
```
The hyperparameters used are as follows.
```
per_device_train_batch_size 16
per_device_eval_batch_size 16
num_train_epochs 4
learning_rate 2e-5
```
## Datasets
This model was trained on the Financial PhraseBank (>= 50% agreement)