clavel commited on
Commit
75532ac
1 Parent(s): 88341a3

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +75 -75
README.md CHANGED
@@ -1,76 +1,76 @@
1
- ---
2
- annotations_creators:
3
- - inoid
4
- - MajorIsaiah
5
- - Ximyer
6
- - clavel
7
- tags:
8
- - "transformers"
9
- - "text-classification"
10
- languages: "es"
11
- license: "apache-2.0"
12
- datasets: "unam_tesis"
13
- metrics: "accuracy"
14
- widget:
15
- - text: "Introducción al análisis de riesgos competitivos bajo el enfoque de la función de incidencia acumulada (FIA) y su aplicación con R"
16
- - text: "Asociación del polimorfismo rs1256031 del receptor beta de estrógenos en pacientes con diabetes tipo 2"
17
- ---
18
-
19
- # Unam_tesis_beto_finnetuning: Unam's thesis classification with BETO
20
-
21
- This model is created from the finetuning of the pre-model
22
- for Spanish [BETO] (https://huggingface.co/dccuchile/bert-base-spanish-wwm-uncased), using PyTorch framework,
23
- and trained with a set of theses of the National Autonomous University of Mexico (UNAM) (https://tesiunam.dgb.unam.mx/F?func=find-b-0&local_base=TES01).
24
- The model classifies a text into for five (Psicología, Derecho, Química Farmacéutico Biológica, Actuaría, Economía)
25
- possible careers at the UNAM.
26
-
27
- ## Training Dataset
28
-
29
- 1000 documents (Thesis introduction, Author´s first name, Author´s last name, Thesis title, Year, Career)
30
-
31
- | Careers | Size |
32
- |--------------|----------------------|
33
- | Actuaría | 200 |
34
- | Derecho| 200 |
35
- | Economía| 200 |
36
- | Psicología| 200 |
37
- | Química Farmacéutico Biológica| 200 |
38
-
39
- ## Example of use
40
-
41
- For further details on how to use unam_tesis_BETO_finnetuning you can visit the Huggingface Transformers library, starting with the Quickstart section. Unam_tesis models can be accessed simply as 'hackathon-pln-e/unam_tesis_BETO_finnetuning' by using the Transformers library. An example of how to download and use the models on this page can be found in this colab notebook.
42
-
43
- ```python
44
-
45
- tokenizer = AutoTokenizer.from_pretrained('hiiamsid/BETO_es_binary_classification', use_fast=False)
46
- model = AutoModelForSequenceClassification.from_pretrained(
47
- 'hackathon-pln-e/unam_tesis_BETO_finnetuning', num_labels=5, output_attentions=False,
48
- output_hidden_states=False)
49
- pipe = TextClassificationPipeline(model=model, tokenizer=tokenizer, return_all_scores=True)
50
-
51
- classificationResult = pipe("Análisis de las condiciones del aprendizaje desde casa en los alumnos de preescolar y primaria del municipio de Nicolás Romero")
52
-
53
- ```
54
-
55
- To cite this resource in a publication please use the following:
56
-
57
- ## Citation
58
-
59
- [UNAM's Tesis with BETO finetuning classify] (https://huggingface.co/hackathon-pln-es/unam_tesis_BETO_finnetuning)
60
-
61
- To cite this resource in a publication please use the following:
62
-
63
- ```
64
- @inproceedings{SpanishNLPHackaton2022,
65
- title={UNAM's Theses with BETO fine-tuning classify },
66
- author={López López, Isaac Isaías and López Ramos, Dionis and Clavel Quintero, Yisel and López López, Ximena Yeraldin },
67
- booktitle={Somos NLP Hackaton 2022},
68
- year={2022}
69
- }
70
- ```
71
-
72
- ## Team members
73
- - Isaac Isaías López López ([MajorIsaiah](https://huggingface.co/MajorIsaiah))
74
- - Dionis López Ramos ([inoid](https://huggingface.co/inoid))
75
- - Yisel Clavel Quintero ([clavel](https://huggingface.co/clavel))
76
  - Ximena Yeraldin López López ([Ximyer](https://huggingface.co/Ximyer))
 
1
+ ---
2
+ annotations_creators:
3
+ - inoid
4
+ - MajorIsaiah
5
+ - Ximyer
6
+ - clavel
7
+ tags:
8
+ - "transformers"
9
+ - "text-classification"
10
+ languages: "es"
11
+ license: "apache-2.0"
12
+ datasets: "unam_tesis"
13
+ metrics: "accuracy"
14
+ widget:
15
+ - text: "Introducción al análisis de riesgos competitivos bajo el enfoque de la función de incidencia acumulada (FIA) y su aplicación con R"
16
+ - text: "Asociación del polimorfismo rs1256031 del receptor beta de estrógenos en pacientes con diabetes tipo 2"
17
+ ---
18
+
19
+ # Unam_tesis_beto_finnetuning: Unam's thesis classification with BETO
20
+
21
+ This model is created from the finetuning of the pre-model
22
+ for Spanish [BETO] (https://huggingface.co/dccuchile/bert-base-spanish-wwm-uncased), using PyTorch framework,
23
+ and trained with a set of theses of the National Autonomous University of Mexico (UNAM) (https://tesiunam.dgb.unam.mx/F?func=find-b-0&local_base=TES01).
24
+ The model classifies a text into for five (Psicología, Derecho, Química Farmacéutico Biológica, Actuaría, Economía)
25
+ possible careers at the UNAM.
26
+
27
+ ## Training Dataset
28
+
29
+ 1000 documents (Thesis introduction, Author´s first name, Author´s last name, Thesis title, Year, Career)
30
+
31
+ | Careers | Size |
32
+ |--------------|----------------------|
33
+ | Actuaría | 200 |
34
+ | Derecho| 200 |
35
+ | Economía| 200 |
36
+ | Psicología| 200 |
37
+ | Química Farmacéutico Biológica| 200 |
38
+
39
+ ## Example of use
40
+
41
+ For further details on how to use unam_tesis_BETO_finnetuning you can visit the Huggingface Transformers library, starting with the Quickstart section. Unam_tesis models can be accessed simply as 'hackathon-pln-e/unam_tesis_BETO_finnetuning' by using the Transformers library. An example of how to download and use the models on this page can be found in this colab notebook.
42
+
43
+ ```python
44
+
45
+ tokenizer = AutoTokenizer.from_pretrained('hiiamsid/BETO_es_binary_classification', use_fast=False)
46
+ model = AutoModelForSequenceClassification.from_pretrained(
47
+ 'hackathon-pln-e/unam_tesis_BETO_finnetuning', num_labels=5, output_attentions=False,
48
+ output_hidden_states=False)
49
+ pipe = TextClassificationPipeline(model=model, tokenizer=tokenizer, return_all_scores=True)
50
+
51
+ classificationResult = pipe("Análisis de las condiciones del aprendizaje desde casa en los alumnos de preescolar y primaria del municipio de Nicolás Romero")
52
+
53
+ ```
54
+
55
+ To cite this resource in a publication please use the following:
56
+
57
+ ## Citation
58
+
59
+ [UNAM's Tesis with BETO finetuning classify] (https://huggingface.co/hackathon-pln-es/unam_tesis_BETO_finnetuning)
60
+
61
+ To cite this resource in a publication please use the following:
62
+
63
+ ```
64
+ @inproceedings{SpanishNLPHackaton2022,
65
+ title={UNAM's Theses with BETO fine-tuning classify},
66
+ author={López López, Isaac Isaías; Clavel Quintero, Yisel; López Ramos, Dionis & López López, Ximena Yeraldin},
67
+ booktitle={Somos NLP Hackaton 2022},
68
+ year={2022}
69
+ }
70
+ ```
71
+
72
+ ## Team members
73
+ - Isaac Isaías López López ([MajorIsaiah](https://huggingface.co/MajorIsaiah))
74
+ - Dionis López Ramos ([inoid](https://huggingface.co/inoid))
75
+ - Yisel Clavel Quintero ([clavel](https://huggingface.co/clavel))
76
  - Ximena Yeraldin López López ([Ximyer](https://huggingface.co/Ximyer))