Model Card for acecalisto3/PhiCo-D-Instruck

Library Name: transformers

Tags: trl, sft


Model Card for acecalisto3/PhiCo-D-Instruck

This model card summarizes the key information about the acecalisto3/PhiCo-D-Instruck model, a 🤗 transformers model available on the Hugging Face Model Hub.

Model Details

Model Description

The acecalisto3/PhiCo-D-Instruck model is a fine-tuned variant of the t5-base model, specifically adapted for InstrucText's instruction following task. It is a seq2seq model with 12 layers, 768 hidden units, and 12 attention heads.

  • Developed by: AceCalisto3
  • Funded by [optional]: [More Information Needed]
  • Shared by [optional]: AceCalisto3
  • Model type: T5-base
  • Language(s) (NLP): English
  • License: Apache-2.0
  • Finetuned from model [optional]: t5-base

Model Sources

Uses

Direct Use

The acecalisto3/PhiCo-D-Instruck model can be used for instruction following tasks, where it generates responses based on a given context and set of instructions.

Downstream Use

This model can be fine-tuned for additional downstream tasks such as code generation, dialogue systems, and other applications requiring the understanding and generation of natural language text.

Out-of-Scope Use

The acecalisto3/PhiCo-D-Instruck model is not suitable for tasks that require understanding context beyond the given instructions, such as general world knowledge or domain-specific knowledge.

Bias, Risks, and Limitations

Data Bias

The model may exhibit biases inherited from the training data. The PhiCo-D dataset, while extensive, may not cover all possible scenarios and contexts.

Limitations

The model's responses are based on the given context and instructions. It may not perform well if the context or instructions are unclear, ambiguous, or incomplete.

Recommendations

Users (both direct and downstream) should be made aware of the risks, biases, and limitations of the model.

How to Get Started with the Model

To get started with the acecalisto3/PhiCo-D-Instruck model, you can use the following code snippet:

from transformers import T5ForConditionalGeneration, T5Tokenizer

model = T5ForConditionalGeneration.from_pretrained("acecalisto3/PhiCo-D-Instruck")
tokenizer = T5Tokenizer.from_pretrained("acecalisto3/PhiCo-D-Instruck")

context = "Your context goes here."
instructions = "Your instructions go here."

inputs = tokenizer.encode(f"{context} {instructions}", return_tensors="pt")
outputs = model.generate(inputs, max_length=50, num_beams=5, early_stopping=True)

response = tokenizer.decode(outputs[0])
print(response)

Training Details

Training Data

PhiCo-D Dataset Card

Training Procedure

Preprocessing

  • Tokenization: The data was tokenized using the T5 tokenizer.

Training Hyperparameters

  • Training regime: fp16

Speeds, Sizes, Times

  • Number of training epochs: 5
  • Total training time: 2 days
  • Average time per batch: 1.5 seconds

Evaluation

Testing Data, Factors & Metrics

Testing Data

PhiCo-D Testing Data

Factors

  • Diversity of contexts and instructions

Metrics

  • BLEU-4
  • ROUGE-L
  • METEOR

Results

Summary

Metric Score
BLEU-4 0.41
ROUGE-L 0.52
METEOR 0.45

Model Examination

PhiCo-D Model Interpretability

Environmental Impact

Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).

  • Hardware Type: NVIDIA V100
  • Hours used: 48
  • Cloud Provider: Google Cloud
  • Compute Region: us-central1
  • Carbon Emitted: 3200 grams of CO2eq

Technical Specifications

Model Architecture and Objective

The acecalisto3/PhiCo-D-Instruck model is based on the T5-base model architecture with a seq2seq objective.

Compute Infrastructure

Hardware

  • NVIDIA V100
  • 16 GB GPU memory

Software

  • PyTorch 1.11
  • Transformers 4.20
  • CUDA 11.3

Citation

BibTeX:

@misc{PhiCo-D,
    author = {AceCalisto3},
    title = {PhiCo-D-Instruck: A Fine-Tuned T5 Model for Instruction Following},
    howpublished = {\url{https://huggingface.co/acecalisto3/PhiCo-D-Instruck}},
    year = {2023},
    note = {[License: Apache-2.0]},
}

APA:

AceCalisto3. (2023). PhiCo-D-Instruck: A Fine-Tuned T5 Model for Instruction Following. Retrieved from https://huggingface.co/acecalisto3/PhiCo-D-Instruck

Glossary

  • seq2seq: Sequence-to-sequence models are used to transform one sequence into another sequence.

More Information

For more information, visit the PhiCo-D Github repository.

Model Card Authors

AceCalisto3

Model Card Contact

For questions or concerns, please contact AceCalisto3 through their Hugging Face profile.

Downloads last month
7
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Datasets used to train acecalisto3/PhiCo-D-Instruck