inoid's picture
Update README.md
bd0f9f1
---
license: apache-2.0
---
# Unam_tesis_beto_finnetuning: Unam's thesis classification with BETO
This model is created from the finetuning of the pre-model
for Spanish [BETO] (https://huggingface.co/dccuchile/bert-base-spanish-wwm-uncased), using PyTorch framework,
and trained with a set of theses at the Autonomous University of Mexico [UNAM](https://tesiunam.dgb.unam.mx/F?func=find-b-0&local_base=TES01).
The model classifies for five (Psychology, Law, Química Farmaceutico Biológica, Actuaría, Economy)
possible careers at the University of Mexico.
List of races from a text.
## Example of use
For further details on how to use unam_tesis_beto_finnetuning you can visit the Huggingface Transformers library, starting with the Quickstart section. Unam_tesis models can be accessed simply as 'inoid/unam_tesis_beto_finnetuning' by using the Transformers library. An example of how to download and use the models on this page can be found in this colab notebook.
```python
tokenizer = AutoTokenizer.from_pretrained('inoid/unam_tesis_beto_finnetuning', use_fast=False)
model = AutoModelForSequenceClassification.from_pretrained(
'inoid/unam_tesis_beto_finnetuning', num_labels=5, output_attentions=False,
output_hidden_states=False)
pipe = TextClassificationPipeline(model=model, tokenizer=tokenizer, return_all_scores=True)
classificationResult = pipe("El objetivo de esta tesis es elaborar un estudio de las condiciones asociadas al aprendizaje desde casa")
```
To cite this resource in a publication please use the following:
## Citation
[UNAM's Tesis with BETO finetuning classify ](https://huggingface.co/hackathon-pln-es/unam_tesis_BETO_finnetuning)
To cite this resource in a publication please use the following:
```
@inproceedings{SpanishNLPHackaton2022,
title={UNAM's Tesis with BETO finetunning classify },
author={Cañete, Isahías and López, Dionis and Clavel, Yisell and López López, Ximena Yeraldin },
booktitle={Somos NLP Hackaton 2022},
year={2022}
}
```