WinterJet2021 commited on
Commit
060249c
·
verified ·
1 Parent(s): cf9ee25

Upload FallacyHunter RoBERTa classifier

Browse files
Files changed (7) hide show
  1. .DS_Store +0 -0
  2. README.md +67 -1
  3. config.json +61 -0
  4. label_map.json +34 -0
  5. model.safetensors +3 -0
  6. tokenizer.json +0 -0
  7. tokenizer_config.json +16 -0
.DS_Store ADDED
Binary file (6.15 kB). View file
 
README.md CHANGED
@@ -1,3 +1,69 @@
1
  ---
2
- license: mit
 
 
 
 
 
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ license: other
3
+ base_model: roberta-base
4
+ tags:
5
+ - text-classification
6
+ - fallacy-classification
7
+ - fairness
8
+ - FallacyHunter
9
+ pipeline_tag: text-classification
10
  ---
11
+
12
+ # FallacyHunter RoBERTa Fallacy Classifier
13
+
14
+ This model is a RoBERTa-based fallacy classifier fine-tuned for the FallacyHunter project.
15
+ It predicts one of 14 fallacy labels for a given argument or statement.
16
+
17
+ ## Model Details
18
+
19
+ - Base model: RoBERTa checkpoint
20
+ - Task: fallacy classification
21
+ - Output labels: ad hominem, ad populum, appeal to emotion, circular reasoning, equivocation, fallacy of credibility, fallacy of extension, fallacy of logic, fallacy of relevance, false causality, false dilemma, faulty generalization, intentional, no_fallacy
22
+ - Repository artifact: local checkpoint directory used for upload
23
+
24
+ ## Intended Use
25
+
26
+ Use this model to label argumentative text for FallacyHunter experiments and related analysis.
27
+ It is suited for offline evaluation, fairness testing, and research workflows.
28
+
29
+ ## Limitations
30
+
31
+ - The model is trained on the FallacyHunter label set and should not be treated as a general-purpose reasoning system.
32
+ - Predictions are only as reliable as the text distribution seen during fine-tuning.
33
+ - Multi-label style outputs should be interpreted according to the checkpoint configuration and downstream decoding logic.
34
+
35
+ ## Labels
36
+
37
+ 0. ad hominem
38
+ 1. ad populum
39
+ 2. appeal to emotion
40
+ 3. circular reasoning
41
+ 4. equivocation
42
+ 5. fallacy of credibility
43
+ 6. fallacy of extension
44
+ 7. fallacy of logic
45
+ 8. fallacy of relevance
46
+ 9. false causality
47
+ 10. false dilemma
48
+ 11. faulty generalization
49
+ 12. intentional
50
+ 13. no_fallacy
51
+
52
+ ## Files
53
+
54
+ This repository folder contains the full local checkpoint used for upload:
55
+
56
+ - `config.json`
57
+ - `label_map.json`
58
+ - `model.safetensors`
59
+ - `tokenizer.json`
60
+ - `tokenizer_config.json`
61
+
62
+ ## Example
63
+
64
+ ```python
65
+ from transformers import pipeline
66
+
67
+ classifier = pipeline("text-classification", model="<username>/<repo_name>")
68
+ classifier("That argument ignores the evidence and attacks the person instead.")
69
+ ```
config.json ADDED
@@ -0,0 +1,61 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_cross_attention": false,
3
+ "architectures": [
4
+ "RobertaForSequenceClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "bos_token_id": 0,
8
+ "classifier_dropout": null,
9
+ "dtype": "float32",
10
+ "eos_token_id": 2,
11
+ "hidden_act": "gelu",
12
+ "hidden_dropout_prob": 0.1,
13
+ "hidden_size": 768,
14
+ "id2label": {
15
+ "0": "ad hominem",
16
+ "1": "ad populum",
17
+ "2": "appeal to emotion",
18
+ "3": "circular reasoning",
19
+ "4": "equivocation",
20
+ "5": "fallacy of credibility",
21
+ "6": "fallacy of extension",
22
+ "7": "fallacy of logic",
23
+ "8": "fallacy of relevance",
24
+ "9": "false causality",
25
+ "10": "false dilemma",
26
+ "11": "faulty generalization",
27
+ "12": "intentional",
28
+ "13": "no_fallacy"
29
+ },
30
+ "initializer_range": 0.02,
31
+ "intermediate_size": 3072,
32
+ "is_decoder": false,
33
+ "label2id": {
34
+ "ad hominem": 0,
35
+ "ad populum": 1,
36
+ "appeal to emotion": 2,
37
+ "circular reasoning": 3,
38
+ "equivocation": 4,
39
+ "fallacy of credibility": 5,
40
+ "fallacy of extension": 6,
41
+ "fallacy of logic": 7,
42
+ "fallacy of relevance": 8,
43
+ "false causality": 9,
44
+ "false dilemma": 10,
45
+ "faulty generalization": 11,
46
+ "intentional": 12,
47
+ "no_fallacy": 13
48
+ },
49
+ "layer_norm_eps": 1e-05,
50
+ "max_position_embeddings": 514,
51
+ "model_type": "roberta",
52
+ "num_attention_heads": 12,
53
+ "num_hidden_layers": 12,
54
+ "pad_token_id": 1,
55
+ "problem_type": "single_label_classification",
56
+ "tie_word_embeddings": true,
57
+ "transformers_version": "5.0.0",
58
+ "type_vocab_size": 1,
59
+ "use_cache": false,
60
+ "vocab_size": 50265
61
+ }
label_map.json ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "id2label": {
3
+ "0": "ad hominem",
4
+ "1": "ad populum",
5
+ "2": "appeal to emotion",
6
+ "3": "circular reasoning",
7
+ "4": "equivocation",
8
+ "5": "fallacy of credibility",
9
+ "6": "fallacy of extension",
10
+ "7": "fallacy of logic",
11
+ "8": "fallacy of relevance",
12
+ "9": "false causality",
13
+ "10": "false dilemma",
14
+ "11": "faulty generalization",
15
+ "12": "intentional",
16
+ "13": "no_fallacy"
17
+ },
18
+ "label2id": {
19
+ "ad hominem": 0,
20
+ "ad populum": 1,
21
+ "appeal to emotion": 2,
22
+ "circular reasoning": 3,
23
+ "equivocation": 4,
24
+ "fallacy of credibility": 5,
25
+ "fallacy of extension": 6,
26
+ "fallacy of logic": 7,
27
+ "fallacy of relevance": 8,
28
+ "false causality": 9,
29
+ "false dilemma": 10,
30
+ "faulty generalization": 11,
31
+ "intentional": 12,
32
+ "no_fallacy": 13
33
+ }
34
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:19b430594f67ad05442ac0f7597ffcd887dbb3448f225bc7723c1577ed3436ce
3
+ size 498649712
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "backend": "tokenizers",
4
+ "bos_token": "<s>",
5
+ "cls_token": "<s>",
6
+ "eos_token": "</s>",
7
+ "errors": "replace",
8
+ "is_local": false,
9
+ "mask_token": "<mask>",
10
+ "model_max_length": 512,
11
+ "pad_token": "<pad>",
12
+ "sep_token": "</s>",
13
+ "tokenizer_class": "RobertaTokenizer",
14
+ "trim_offsets": true,
15
+ "unk_token": "<unk>"
16
+ }