AjayP13 commited on
Commit
4d0857f
·
verified ·
1 Parent(s): 9fc7967

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +11 -2
README.md CHANGED
@@ -34,8 +34,17 @@ tokenizer = AutoTokenizer.from_pretrained('datadreamer-dev/abstracts_to_tweet_mo
34
  model = AutoModelForSeq2SeqLM.from_pretrained('datadreamer-dev/abstracts_to_tweet_model', revision=None) # Load model
35
  pipe = pipeline('text2text-generation', model=model, tokenizer=tokenizer, pad_token_id=tokenizer.pad_token_id)
36
 
37
- inputs = ["In this paper, we delve into advanced techniques and methods in Natural Language Processing (NLP), innovatively incorporating Transformer architectures and self-supervised learning methods. We aim to reiterate the current understanding of Transformer-based models in executing various language tasks by dissecting their versatility and expandability on broad language systems.\n\nMoreover, stabilization measures, tokenization assortment, and interpreting latent spaces provide an in-depth novelty to our pipeline, overcoming long-known obstacles. We explore meta-architectural modifications focusing on enhancing prompt language models' efficiency, allowing flexible adaptations to the core Transformer technique's abundance in BERT, GPT-like systems.\n\nTo implement these adaptations, several experiments were conducted on varied benchmark datasets to evaluate core metrics such as Bleu, Rouge, and Warp-CTC metrics in translation and transcription tasks. We carried out significant analysis focusing on module interpretability, additional error inspection, task-specific regulatory mechanisms, execution speed, and computational considerations.\n\nOur experimental results bring in distraction from widespread but sub-optimal benchmarks and offer evidence underpinning the contrary yet potent issues yet to be addressed methodically. We invite the community to reflect on these novel insights, develop and refine our proposed techniques, speeding technical progress, avoiding prototypical retrodiction in the Natural Language Understanding ecosystem to respect inclusive, diverse, and correctly perceived expressive content."]
38
- print(pipe(inputs, max_length=512, do_sample=False))
 
 
 
 
 
 
 
 
 
39
  ```
40
 
41
  ---
 
34
  model = AutoModelForSeq2SeqLM.from_pretrained('datadreamer-dev/abstracts_to_tweet_model', revision=None) # Load model
35
  pipe = pipeline('text2text-generation', model=model, tokenizer=tokenizer, pad_token_id=tokenizer.pad_token_id)
36
 
37
+ # For example, run the model on the abstract of the LoRA paper (https://arxiv.org/abs/2106.09685)
38
+ abstract = "An important paradigm of natural language processing consists of large-scale pre-training on general domain data and adaptation to particular tasks or domains. As we pre-train larger models, full fine-tuning, which retrains all model parameters, becomes less feasible. Using GPT-3 175B as an example -- deploying independent instances of fine-tuned models, each with 175B parameters, is prohibitively expensive. We propose Low-Rank Adaptation, or LoRA, which freezes the pre-trained model weights and injects trainable rank decomposition matrices into each layer of the Transformer architecture, greatly reducing the number of trainable parameters for downstream tasks. Compared to GPT-3 175B fine-tuned with Adam, LoRA can reduce the number of trainable parameters by 10,000 times and the GPU memory requirement by 3 times. LoRA performs on-par or better than fine-tuning in model quality on RoBERTa, DeBERTa, GPT-2, and GPT-3, despite having fewer trainable parameters, a higher training throughput, and, unlike adapters, no additional inference latency. We also provide an empirical investigation into rank-deficiency in language model adaptation, which sheds light on the efficacy of LoRA. We release a package that facilitates the integration of LoRA with PyTorch models and provide our implementations and model checkpoints for RoBERTa, DeBERTa, and GPT-2 at this https URL."
39
+ generated_tweet = pipe(abstract, max_length=512)['generated_text']
40
+
41
+ # Print the generated tweet
42
+ print(generated_tweet)
43
+
44
+ # This will print:
45
+ # "Exciting news in #NLP! We've developed Low-Rank Adaptation, or LoRA, to reduce the number of
46
+ # trainable parameters for downstream tasks. It reduces model weights by 10,000 times and GPU
47
+ # memory by 3 times. #AI #MachineLearning"
48
  ```
49
 
50
  ---