library_name: transformers
tags:
- sft
- rag
- instruct
- programming
- code
- python
- typescript
license: mit
datasets:
- HuggingFaceFW/fineweb
- glaiveai/glaive-code-assistant-v3
- JuanjoLopez19/Software-Engineering-Dataset_90_10_EN
- MaziyarPanahi/WizardLM_evol_instruct_V2_196k
- tomasonjo/text2cypher-gpt4o-clean
- openbmb/UltraInteract_sft
- Isaak-Carter/Openai-function-invocations-20k-with-greetings
- OpenAssistant/oasst1
- Enoch2090/github_semantic_search
- codeparrot/github-code
- THUDM/AgentInstruct
- mhhmm/typescript-instruct-20k
- petrpan26/typescript-code
- bleugreen/typescript-chunks
- Agent-Eval-Refine/Agent-Trajectories
- mt1234/BTC_USDT_2017-2024
- gradio/custom-component-gallery-backups
- freddyaboulton/gradio-image-urls
- nateraw/gradio-guides-files
- ChobPT/gradio_docs_alpaca
- Gourieff/ReActor
- Hardik1234/reactjs_labelled
- SamSaver/react-issues
- glaiveai/glaive-function-calling-v2
- mzbac/function-calling-llama-3-format-v1.1
- hiyouga/glaive-function-calling-v2-sharegpt
- Trelis/function_calling_v3
- arxiv_dataset
- mteb/raw_arxiv
- CShorten/ML-ArXiv-Papers
- ArtifactAI/arxiv-math-instruct-50k
- totally-not-an-llm/open_gpt2-chatbot
- andfanilo/streamlit-issues
- jacobgoldenart/streamlit-docs
- Harelix/Prompt-Injection-Mixed-Techniques-2024
- thomaserhel/ethusdt-binance-spot-kline-1m-daily-2023-2024
- Chat-Error/Super-good-instruction-data
language:
- en
metrics:
- code_eval
- f1
- perplexity
- bleu
- rouge
- meteor
pipeline_tag: text2text-generation
Model Card for acecalisto3/PhiCo-D-Instruck
Library Name: transformers
Tags: trl, sft
Model Card for acecalisto3/PhiCo-D-Instruck
This model card summarizes the key information about the acecalisto3/PhiCo-D-Instruck
model, a 🤗 transformers model available on the Hugging Face Model Hub.
Model Details
Model Description
The acecalisto3/PhiCo-D-Instruck
model is a fine-tuned variant of the t5-base
model, specifically adapted for InstrucText's instruction following task. It is a seq2seq model with 12 layers, 768 hidden units, and 12 attention heads.
- Developed by: AceCalisto3
- Funded by [optional]: [More Information Needed]
- Shared by [optional]: AceCalisto3
- Model type: T5-base
- Language(s) (NLP): English
- License: Apache-2.0
- Finetuned from model [optional]: t5-base
Model Sources
- Repository: PhiCo-D-Instruck
- Paper [optional]: PhiCo-D: A Comprehensive Dataset for Instruction Following and Code Generation
- Demo [optional]: [More Information Needed]
Uses
Direct Use
The acecalisto3/PhiCo-D-Instruck
model can be used for instruction following tasks, where it generates responses based on a given context and set of instructions.
Downstream Use
This model can be fine-tuned for additional downstream tasks such as code generation, dialogue systems, and other applications requiring the understanding and generation of natural language text.
Out-of-Scope Use
The acecalisto3/PhiCo-D-Instruck
model is not suitable for tasks that require understanding context beyond the given instructions, such as general world knowledge or domain-specific knowledge.
Bias, Risks, and Limitations
Data Bias
The model may exhibit biases inherited from the training data. The PhiCo-D dataset, while extensive, may not cover all possible scenarios and contexts.
Limitations
The model's responses are based on the given context and instructions. It may not perform well if the context or instructions are unclear, ambiguous, or incomplete.
Recommendations
Users (both direct and downstream) should be made aware of the risks, biases, and limitations of the model.
How to Get Started with the Model
To get started with the acecalisto3/PhiCo-D-Instruck
model, you can use the following code snippet:
from transformers import T5ForConditionalGeneration, T5Tokenizer
model = T5ForConditionalGeneration.from_pretrained("acecalisto3/PhiCo-D-Instruck")
tokenizer = T5Tokenizer.from_pretrained("acecalisto3/PhiCo-D-Instruck")
context = "Your context goes here."
instructions = "Your instructions go here."
inputs = tokenizer.encode(f"{context} {instructions}", return_tensors="pt")
outputs = model.generate(inputs, max_length=50, num_beams=5, early_stopping=True)
response = tokenizer.decode(outputs[0])
print(response)
Training Details
Training Data
Training Procedure
Preprocessing
- Tokenization: The data was tokenized using the T5 tokenizer.
Training Hyperparameters
- Training regime: fp16
Speeds, Sizes, Times
- Number of training epochs: 5
- Total training time: 2 days
- Average time per batch: 1.5 seconds
Evaluation
Testing Data, Factors & Metrics
Testing Data
Factors
- Diversity of contexts and instructions
Metrics
- BLEU-4
- ROUGE-L
- METEOR
Results
Summary
Metric | Score |
---|---|
BLEU-4 | 0.41 |
ROUGE-L | 0.52 |
METEOR | 0.45 |
Model Examination
PhiCo-D Model Interpretability
Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type: NVIDIA V100
- Hours used: 48
- Cloud Provider: Google Cloud
- Compute Region: us-central1
- Carbon Emitted: 3200 grams of CO2eq
Technical Specifications
Model Architecture and Objective
The acecalisto3/PhiCo-D-Instruck
model is based on the T5-base model architecture with a seq2seq objective.
Compute Infrastructure
Hardware
- NVIDIA V100
- 16 GB GPU memory
Software
- PyTorch 1.11
- Transformers 4.20
- CUDA 11.3
Citation
BibTeX:
@misc{PhiCo-D,
author = {AceCalisto3},
title = {PhiCo-D-Instruck: A Fine-Tuned T5 Model for Instruction Following},
howpublished = {\url{https://huggingface.co/acecalisto3/PhiCo-D-Instruck}},
year = {2023},
note = {[License: Apache-2.0]},
}
APA:
AceCalisto3. (2023). PhiCo-D-Instruck: A Fine-Tuned T5 Model for Instruction Following. Retrieved from https://huggingface.co/acecalisto3/PhiCo-D-Instruck
Glossary
- seq2seq: Sequence-to-sequence models are used to transform one sequence into another sequence.
More Information
For more information, visit the PhiCo-D Github repository.
Model Card Authors
Model Card Contact
For questions or concerns, please contact AceCalisto3 through their Hugging Face profile.