--- tags: - pytorch_model_hub_mixin - model_hub_mixin license: gpl-3.0 --- This model has been pushed to the Hub using the [PytorchModelHubMixin](https://huggingface.co/docs/huggingface_hub/package_reference/mixins#huggingface_hub.PyTorchModelHubMixin) integration: - Library: [More Information Needed] - Docs: [More Information Needed] ## Steps to run model - First install [transforna](https://github.com/gitHBDX/TransfoRNA/tree/master) - Example code: ``` from transforna import GeneEmbeddModel,RnaTokenizer import torch model_name = 'Seq-Struct' model_path = f"HBDX/{model_name}-TransfoRNA" #load model and tokenizer model = GeneEmbeddModel.from_pretrained(model_path) model.eval() #init tokenizer. Tokenizer will automatically get secondary structure of sequence using Vienna RNA package tokenizer = RnaTokenizer.from_pretrained(model_path,model_name=model_name) output = tokenizer(['AAAGTCGGAGGTTCGAAGACGATCAGATAC','TTTTCGGAACTGAGGCCATGATTAAGAGGG']) #inference #gene_embedds and second input embedds are the latent space representation of the input sequence and the second input respectively. #In this case, the second input would be the secondary structure of the sequence gene_embedd, second_input_embedd, activations,attn_scores_first,attn_scores_second = \ model(output['input_ids']) #get sub class labels sub_class_labels = model.convert_ids_to_labels(activations) #get major class labels major_class_labels = model.convert_subclass_to_majorclass(sub_class_labels) ```