Update README.md
Browse files
README.md
CHANGED
@@ -1,76 +1,76 @@
|
|
1 |
-
---
|
2 |
-
annotations_creators:
|
3 |
-
- inoid
|
4 |
-
- MajorIsaiah
|
5 |
-
- Ximyer
|
6 |
-
- clavel
|
7 |
-
tags:
|
8 |
-
- "transformers"
|
9 |
-
- "text-classification"
|
10 |
-
languages: "es"
|
11 |
-
license: "apache-2.0"
|
12 |
-
datasets: "unam_tesis"
|
13 |
-
metrics: "accuracy"
|
14 |
-
widget:
|
15 |
-
- text: "Introducción al análisis de riesgos competitivos bajo el enfoque de la función de incidencia acumulada (FIA) y su aplicación con R"
|
16 |
-
- text: "Asociación del polimorfismo rs1256031 del receptor beta de estrógenos en pacientes con diabetes tipo 2"
|
17 |
-
---
|
18 |
-
|
19 |
-
# Unam_tesis_beto_finnetuning: Unam's thesis classification with BETO
|
20 |
-
|
21 |
-
This model is created from the finetuning of the pre-model
|
22 |
-
for Spanish [BETO] (https://huggingface.co/dccuchile/bert-base-spanish-wwm-uncased), using PyTorch framework,
|
23 |
-
and trained with a set of theses of the National Autonomous University of Mexico (UNAM) (https://tesiunam.dgb.unam.mx/F?func=find-b-0&local_base=TES01).
|
24 |
-
The model classifies a text into for five (Psicología, Derecho, Química Farmacéutico Biológica, Actuaría, Economía)
|
25 |
-
possible careers at the UNAM.
|
26 |
-
|
27 |
-
## Training Dataset
|
28 |
-
|
29 |
-
1000 documents (Thesis introduction, Author´s first name, Author´s last name, Thesis title, Year, Career)
|
30 |
-
|
31 |
-
| Careers | Size |
|
32 |
-
|--------------|----------------------|
|
33 |
-
| Actuaría | 200 |
|
34 |
-
| Derecho| 200 |
|
35 |
-
| Economía| 200 |
|
36 |
-
| Psicología| 200 |
|
37 |
-
| Química Farmacéutico Biológica| 200 |
|
38 |
-
|
39 |
-
## Example of use
|
40 |
-
|
41 |
-
For further details on how to use unam_tesis_BETO_finnetuning you can visit the Huggingface Transformers library, starting with the Quickstart section. Unam_tesis models can be accessed simply as 'hackathon-pln-e/unam_tesis_BETO_finnetuning' by using the Transformers library. An example of how to download and use the models on this page can be found in this colab notebook.
|
42 |
-
|
43 |
-
```python
|
44 |
-
|
45 |
-
tokenizer = AutoTokenizer.from_pretrained('hiiamsid/BETO_es_binary_classification', use_fast=False)
|
46 |
-
model = AutoModelForSequenceClassification.from_pretrained(
|
47 |
-
'hackathon-pln-e/unam_tesis_BETO_finnetuning', num_labels=5, output_attentions=False,
|
48 |
-
output_hidden_states=False)
|
49 |
-
pipe = TextClassificationPipeline(model=model, tokenizer=tokenizer, return_all_scores=True)
|
50 |
-
|
51 |
-
classificationResult = pipe("Análisis de las condiciones del aprendizaje desde casa en los alumnos de preescolar y primaria del municipio de Nicolás Romero")
|
52 |
-
|
53 |
-
```
|
54 |
-
|
55 |
-
To cite this resource in a publication please use the following:
|
56 |
-
|
57 |
-
## Citation
|
58 |
-
|
59 |
-
[UNAM's Tesis with BETO finetuning classify] (https://huggingface.co/hackathon-pln-es/unam_tesis_BETO_finnetuning)
|
60 |
-
|
61 |
-
To cite this resource in a publication please use the following:
|
62 |
-
|
63 |
-
```
|
64 |
-
@inproceedings{SpanishNLPHackaton2022,
|
65 |
-
title={UNAM's Theses with BETO fine-tuning classify
|
66 |
-
author={López López, Isaac Isaías
|
67 |
-
booktitle={Somos NLP Hackaton 2022},
|
68 |
-
year={2022}
|
69 |
-
}
|
70 |
-
```
|
71 |
-
|
72 |
-
## Team members
|
73 |
-
- Isaac Isaías López López ([MajorIsaiah](https://huggingface.co/MajorIsaiah))
|
74 |
-
- Dionis López Ramos ([inoid](https://huggingface.co/inoid))
|
75 |
-
- Yisel Clavel Quintero ([clavel](https://huggingface.co/clavel))
|
76 |
- Ximena Yeraldin López López ([Ximyer](https://huggingface.co/Ximyer))
|
|
|
1 |
+
---
|
2 |
+
annotations_creators:
|
3 |
+
- inoid
|
4 |
+
- MajorIsaiah
|
5 |
+
- Ximyer
|
6 |
+
- clavel
|
7 |
+
tags:
|
8 |
+
- "transformers"
|
9 |
+
- "text-classification"
|
10 |
+
languages: "es"
|
11 |
+
license: "apache-2.0"
|
12 |
+
datasets: "unam_tesis"
|
13 |
+
metrics: "accuracy"
|
14 |
+
widget:
|
15 |
+
- text: "Introducción al análisis de riesgos competitivos bajo el enfoque de la función de incidencia acumulada (FIA) y su aplicación con R"
|
16 |
+
- text: "Asociación del polimorfismo rs1256031 del receptor beta de estrógenos en pacientes con diabetes tipo 2"
|
17 |
+
---
|
18 |
+
|
19 |
+
# Unam_tesis_beto_finnetuning: Unam's thesis classification with BETO
|
20 |
+
|
21 |
+
This model is created from the finetuning of the pre-model
|
22 |
+
for Spanish [BETO] (https://huggingface.co/dccuchile/bert-base-spanish-wwm-uncased), using PyTorch framework,
|
23 |
+
and trained with a set of theses of the National Autonomous University of Mexico (UNAM) (https://tesiunam.dgb.unam.mx/F?func=find-b-0&local_base=TES01).
|
24 |
+
The model classifies a text into for five (Psicología, Derecho, Química Farmacéutico Biológica, Actuaría, Economía)
|
25 |
+
possible careers at the UNAM.
|
26 |
+
|
27 |
+
## Training Dataset
|
28 |
+
|
29 |
+
1000 documents (Thesis introduction, Author´s first name, Author´s last name, Thesis title, Year, Career)
|
30 |
+
|
31 |
+
| Careers | Size |
|
32 |
+
|--------------|----------------------|
|
33 |
+
| Actuaría | 200 |
|
34 |
+
| Derecho| 200 |
|
35 |
+
| Economía| 200 |
|
36 |
+
| Psicología| 200 |
|
37 |
+
| Química Farmacéutico Biológica| 200 |
|
38 |
+
|
39 |
+
## Example of use
|
40 |
+
|
41 |
+
For further details on how to use unam_tesis_BETO_finnetuning you can visit the Huggingface Transformers library, starting with the Quickstart section. Unam_tesis models can be accessed simply as 'hackathon-pln-e/unam_tesis_BETO_finnetuning' by using the Transformers library. An example of how to download and use the models on this page can be found in this colab notebook.
|
42 |
+
|
43 |
+
```python
|
44 |
+
|
45 |
+
tokenizer = AutoTokenizer.from_pretrained('hiiamsid/BETO_es_binary_classification', use_fast=False)
|
46 |
+
model = AutoModelForSequenceClassification.from_pretrained(
|
47 |
+
'hackathon-pln-e/unam_tesis_BETO_finnetuning', num_labels=5, output_attentions=False,
|
48 |
+
output_hidden_states=False)
|
49 |
+
pipe = TextClassificationPipeline(model=model, tokenizer=tokenizer, return_all_scores=True)
|
50 |
+
|
51 |
+
classificationResult = pipe("Análisis de las condiciones del aprendizaje desde casa en los alumnos de preescolar y primaria del municipio de Nicolás Romero")
|
52 |
+
|
53 |
+
```
|
54 |
+
|
55 |
+
To cite this resource in a publication please use the following:
|
56 |
+
|
57 |
+
## Citation
|
58 |
+
|
59 |
+
[UNAM's Tesis with BETO finetuning classify] (https://huggingface.co/hackathon-pln-es/unam_tesis_BETO_finnetuning)
|
60 |
+
|
61 |
+
To cite this resource in a publication please use the following:
|
62 |
+
|
63 |
+
```
|
64 |
+
@inproceedings{SpanishNLPHackaton2022,
|
65 |
+
title={UNAM's Theses with BETO fine-tuning classify},
|
66 |
+
author={López López, Isaac Isaías; Clavel Quintero, Yisel; López Ramos, Dionis & López López, Ximena Yeraldin},
|
67 |
+
booktitle={Somos NLP Hackaton 2022},
|
68 |
+
year={2022}
|
69 |
+
}
|
70 |
+
```
|
71 |
+
|
72 |
+
## Team members
|
73 |
+
- Isaac Isaías López López ([MajorIsaiah](https://huggingface.co/MajorIsaiah))
|
74 |
+
- Dionis López Ramos ([inoid](https://huggingface.co/inoid))
|
75 |
+
- Yisel Clavel Quintero ([clavel](https://huggingface.co/clavel))
|
76 |
- Ximena Yeraldin López López ([Ximyer](https://huggingface.co/Ximyer))
|