--- language: - en tags: - retrieval - math-retrieval datasets: - MathematicalStackExchange - ARQMath --- # ALBERT for ARQMath 3 This repository contains our best model for ARQMath 3, the math_10 model. It was initialised from ALBERT-base-v2 and further pre-trained on Math StackExchange in three different stages. We also added more LaTeX tokens to the tokenizer to enable a better tokenization of mathematical formulas. math_10 was fine-tuned on a classification task to determine whether a given question (sequence 1) matches a given answer (sequence 2). The classification output can be used for ranking the best answers. For further details, please read our paper: http://ceur-ws.org/Vol-3180/paper-07.pdf. ## Other Models for ARQMath 3 We plan on also publishing the other fine-tuned models as well as the base models. Links to these repositories will be added here soon. | Model | Initialised from | Pre-training | Fine-Tuned | Link | |-------------|------------------|----------------------------|-------------------------------------|------| | roberta_10 | RoBERTa | MathSE (1) | yes, N=10 MathSE | | | base_10 | ALBERT | MathSE (1) | yes, N=10 MathSE | | | math_10_add | ALBERT | MathSE (1)-(3) | yes, N=10 MathSE and annotated data | | | Khan_SE_10 | ALBERT | MathSE (1) | yes, N=10 MathSE | | | roberta | RoBERTa | MathSE (1) | no | [AnReu/math_pretrained_roberta](https://huggingface.co/AnReu/math_pretrained_roberta) | | math albert | ALBERT | MathSE (1)-(3) | no | [AnReu/math_albert](https://huggingface.co/AnReu/math_albert) | | base | ALBERT | MathSE (1) | no | | | Khan_SE | ALBERT | MathSE (1) mixed with Khan | no | | ### Update We have also further pre-trained a BERT-base-cased model in the same way as our ALBERT model. You can find it here: [AnReu/math_pretrained_bert](https://huggingface.co/AnReu/math_pretrained_bert). # Usage ```python # based on https://huggingface.co/docs/transformers/main/en/task_summary#sequence-classification from transformers import AutoTokenizer, AutoModelForSequenceClassification tokenizer = AutoTokenizer.from_pretrained("AnReu/albert-for-arqmath-3") model = AutoModelForSequenceClassification.from_pretrained("AnReu/albert-for-arqmath-3") classes = ["non relevant", "relevant"] sequence_0 = "How can I calculate x in $3x = 5$" sequence_1 = "Just divide by 3: $x = \\frac{5}{3}$" sequence_2 = "The general rule for squaring a sum is $(a+b)^2=a^2+2ab+b^2$" # The tokenizer will automatically add any model specific separators (i.e. and ) and tokens to # the sequence, as well as compute the attention masks. irrelevant = tokenizer(sequence_0, sequence_2, return_tensors="pt") relevant = tokenizer(sequence_0, sequence_1, return_tensors="pt") irrelevant_classification_logits = model(**irrelevant).logits relevant_classification_logits = model(**relevant).logits irrelevant_results = torch.softmax(irrelevant_classification_logits, dim=1).tolist()[0] relevant_results = torch.softmax(relevant_classification_logits, dim=1).tolist()[0] # Should be irrelevant for i in range(len(classes)): print(f"{classes[i]}: {int(round(irrelevant_results[i] * 100))}%") # Should be relevant for i in range(len(classes)): print(f"{classes[i]}: {int(round(relevant_results[i] * 100))}%") ``` # Citation If you find this model useful, consider citing our paper: ``` @article{reusch2022transformer, title={Transformer-Encoder and Decoder Models for Questions on Math}, author={Reusch, Anja and Thiele, Maik and Lehner, Wolfgang}, year={2022}, organization={CLEF} } ```