OGAI-3.1-Engineer / README.md
tommytracx's picture
Update README.md
fe347d3 verified
|
raw
history blame
7.14 kB
---
license: apache-2.0
language:
- en
metrics:
- accuracy
- precision
base_model:
- nvidia/Llama-3.1-Nemotron-70B-Instruct-HF
new_version: nvidia/Llama-3.1-Nemotron-70B-Instruct-HF
pipeline_tag: text-classification
library_name: transformers
tags:
- llm
- oil-and-gas
- engineering
- custom-llm
- ogai-3.1-engineer
- nvidia
- llama
- Nemotron
- drilling-engineering
---
# OGAI 3.1 Engineer
**Model Author:** Gain.Energy
**Lead Developers:** Dr. Vlad Karén Payrazyan, CEO and Founder at Gain.Energy; Tommy Xaypanya, Lead AI Scientist and Developer at Gain.Energy
**Date Created:** November 12, 2024
## Overview
**OGAI 3.1 Engineer** is a large language model built on NVIDIA’s **Llama-3.1-Nemotron-70B-Instruct-HF** and customized specifically for the oil and gas industry, with a focus on drilling engineering. This model has been fine-tuned to understand and process technical calculations, interpret engineering documents, and generate domain-specific insights, making it a valuable asset for engineers and analysts.
**Applications:**
- Complex engineering calculations
- Document interpretation and summarization
- Drilling optimization and safety compliance
- Collaborative, real-time engineering workspaces
---
## Model Details
- **Base Model:** nvidia/Llama-3.1-Nemotron-70B-Instruct-HF
- **Parameter Count:** 70 billion
- **Architecture:** Transformer-based
- **Input Format:** Text prompts up to 128k tokens
- **Output Format:** Text responses up to 4k tokens
## Revision History
### Revision 1.0 - Initial Release (November 12, 2024)
- **Base Model:** nvidia/Llama-3.1-Nemotron-70B-Instruct-HF
- **Custom Training:** Focused on oil and gas drilling engineering documents, industry standards, technical calculations, and safety protocols.
- **Training Data:**
- Industry-specific manuals, textbooks, and historical operational data.
- Preprocessed datasets to ensure consistency and confidentiality.
- **Fine-Tuning Techniques:**
- **Low-Rank Adaptation (LoRA):** Applied LoRA for efficient parameter fine-tuning.
- **Retrieval-Augmented Generation (RAG):** Integrated for real-time knowledge base retrieval.
- **Prompt Engineering:** Crafted domain-specific prompts for enhanced accuracy.
---
## Installation
To install and run **OGAI 3.1 Engineer**, you’ll need:
- Python 3.9 or higher
- PyTorch 1.12 or higher
- CUDA 11.8 for GPU support
### Clone the Repository
```bash
git clone https://huggingface.co/gain-energy/OGAI-3.1-Engineer
cd OGAI-3.1-Engineer
pip install -r requirements.txt
```
---
### Usage Example
Here is an example code to load and interact with OGAI 3.1 Engineer:
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
model_name = "gain-energy/OGAI-3.1-Engineer"
model = AutoModelForCausalLM.from_pretrained(model_name, torch_dtype=torch.bfloat16, device_map="auto")
tokenizer = AutoTokenizer.from_pretrained(model_name)
prompt = "Calculate the mud weight required for a well with a true vertical depth of 15,000 feet and formation pressure of 10,000 psi."
inputs = tokenizer(prompt, return_tensors="pt").to("cuda")
outputs = model.generate(**inputs, max_length=200)
generated_text = tokenizer.decode(outputs[0], skip_special_tokens=True)
print(generated_text)
```
---
## Model Performance and Evaluation
The model was benchmarked on several evaluation metrics relevant to oil and gas applications:
- Domain-Specific Accuracy: 88% accuracy in answering technical questions.
- Calculation Precision: Improved calculation accuracy by 90% over baseline.
- Benchmark Scores:
- Arena Hard: 86.5%
- AlpacaEval 2.0 LC: 60%
- GPT-4-Turbo MT-Bench: Score of 9.1
---
## Training and Fine-Tuning
- Training Hardware: NVIDIA DGX systems with A100 GPUs (80 GB VRAM per GPU).
- Training Parameters: Batch size of 8 per GPU, learning rate of 1e-4 with a cosine decay, 3 epochs.
- Optimization Algorithm: AdamW with weight decay.
---
## Intended Use and Limitations
### Intended Use
OGAI 3.1 Engineer is intended for professionals in the oil and gas industry, particularly those focused on drilling operations, safety compliance, and technical calculations. Its specialized training enables it to handle domain-specific terminology, calculations, and documentation with a high degree of accuracy.
### Limitations
- Numerical Computation: While enhanced for complex calculations, the model may require external computational tools for highly intricate numerical tasks.
- Generalization: The model may not perform optimally on general knowledge topics outside its fine-tuned oil and gas domain.
---
## License
This model is released under the Apache License 2.0. Please see the LICENSE file for more details.
---
## Acknowledgments
Special thanks to NVIDIA AI Research for the development of the base model and to the Gain.Energy team for domain expertise and support in model fine-tuning and evaluation.
---
### Contact Information
For support, inquiries, or collaboration opportunities, please contact:
- Tommy Xaypanya
Lead AI Scientist and Developer at Gain.Energy
Email: tommy@gain.energy
- Dr. Vlad Karén Payrazyan
CEO and Founder at Gain.Energy
Email: karen@gain.energy
---
model-index:
- name: OGAI 3.1 Engineer
results:
- task:
type: text-generation
dataset:
name: oil_gas_docs
type: GainEnergy-OilGasDocs
metrics:
- name: Domain-Specific Accuracy
type: accuracy
value: 88.0
source:
name: Gain Energy Internal Evaluation
url: https://gain.energy/evaluations/ogai-3-1-engineer
- task:
type: text-generation
dataset:
name: technical_calculations
type: TechnicalCalculations-OilGas
metrics:
- name: Calculation Precision
type: precision
value: 90.0
source:
name: Gain Energy Internal Evaluation
url: https://gain.energy/evaluations/ogai-3-1-engineer
- task:
type: text-generation
dataset:
name: arena_hard
type: arena_hard
metrics:
- name: Arena Hard
type: helpfulness and alignment
value: 86.5
source:
name: Gain Energy Internal Evaluation
url: https://gain.energy/evaluations/ogai-3-1-engineer
- task:
type: text-generation
dataset:
name: alpaca_eval_2_lc
type: AlpacaEval 2.0 Length Controlled
metrics:
- name: AlpacaEval 2.0 Length Controlled (LC)
type: length-controlled
value: 60.0
source:
name: Gain Energy Internal Evaluation
url: https://gain.energy/evaluations/ogai-3-1-engineer
- task:
type: text-generation
dataset:
name: gpt_4_turbo_mt_bench
type: gpt_4_turbo_mt_bench
metrics:
- name: GPT-4-Turbo MT-Bench
type: reasoning and problem-solving
value: 9.1
source:
name: Gain Energy Internal Evaluation
url: https://gain.energy/evaluations/ogai-3-1-engineer