A newer version of this model is available: nvidia/Llama-3.1-Nemotron-70B-Instruct-HF

You need to agree to share your contact information to access this model

This repository is publicly accessible, but you have to accept the conditions to access its files and content.

Log in or Sign Up to review the conditions and access this model content.

OGAI 3.1 Engineer

Model Author: Gain.Energy
Lead Developers: Dr. Vlad Karén Payrazyan, CEO and Founder at Gain.Energy; Tommy Xaypanya, Lead AI Scientist and Developer at Gain.Energy
Date Created: November 12, 2024

Overview

OGAI 3.1 Engineer is a large language model built on NVIDIA’s Llama-3.1-Nemotron-70B-Instruct-HF and customized specifically for the oil and gas industry, with a focus on drilling engineering. This model has been fine-tuned to understand and process technical calculations, interpret engineering documents, and generate domain-specific insights, making it a valuable asset for engineers and analysts.

Applications:

  • Complex engineering calculations
  • Document interpretation and summarization
  • Drilling optimization and safety compliance
  • Collaborative, real-time engineering workspaces

Model Details

  • Base Model: nvidia/Llama-3.1-Nemotron-70B-Instruct-HF
  • Parameter Count: 70 billion
  • Architecture: Transformer-based
  • Input Format: Text prompts up to 128k tokens
  • Output Format: Text responses up to 4k tokens

Revision History

Revision 1.0 - Initial Release (November 12, 2024)

  • Base Model: nvidia/Llama-3.1-Nemotron-70B-Instruct-HF
  • Custom Training: Focused on oil and gas drilling engineering documents, industry standards, technical calculations, and safety protocols.
  • Training Data:
    • Industry-specific manuals, textbooks, and historical operational data.
    • Preprocessed datasets to ensure consistency and confidentiality.
  • Fine-Tuning Techniques:
    • Low-Rank Adaptation (LoRA): Applied LoRA for efficient parameter fine-tuning.
    • Retrieval-Augmented Generation (RAG): Integrated for real-time knowledge base retrieval.
    • Prompt Engineering: Crafted domain-specific prompts for enhanced accuracy.

Installation

To install and run OGAI 3.1 Engineer, you’ll need:

  • Python 3.9 or higher
  • PyTorch 1.12 or higher
  • CUDA 11.8 for GPU support

Clone the Repository

git clone https://huggingface.co/gain-energy/OGAI-3.1-Engineer
cd OGAI-3.1-Engineer
pip install -r requirements.txt

Usage Example

Here is an example code to load and interact with OGAI 3.1 Engineer:

from transformers import AutoModelForCausalLM, AutoTokenizer

model_name = "gain-energy/OGAI-3.1-Engineer"
model = AutoModelForCausalLM.from_pretrained(model_name, torch_dtype=torch.bfloat16, device_map="auto")
tokenizer = AutoTokenizer.from_pretrained(model_name)

prompt = "Calculate the mud weight required for a well with a true vertical depth of 15,000 feet and formation pressure of 10,000 psi."
inputs = tokenizer(prompt, return_tensors="pt").to("cuda")
outputs = model.generate(**inputs, max_length=200)
generated_text = tokenizer.decode(outputs[0], skip_special_tokens=True)
print(generated_text)

Model Performance and Evaluation

The model was benchmarked on several evaluation metrics relevant to oil and gas applications:

  • Domain-Specific Accuracy: 88% accuracy in answering technical questions.
  • Calculation Precision: Improved calculation accuracy by 90% over baseline.
  • Benchmark Scores:
  • Arena Hard: 86.5%
  • AlpacaEval 2.0 LC: 60%
  • GPT-4-Turbo MT-Bench: Score of 9.1

Training and Fine-Tuning

  • Training Hardware: NVIDIA DGX systems with A100 GPUs (80 GB VRAM per GPU).
  • Training Parameters: Batch size of 8 per GPU, learning rate of 1e-4 with a cosine decay, 3 epochs.
  • Optimization Algorithm: AdamW with weight decay.

Intended Use and Limitations

Intended Use

OGAI 3.1 Engineer is intended for professionals in the oil and gas industry, particularly those focused on drilling operations, safety compliance, and technical calculations. Its specialized training enables it to handle domain-specific terminology, calculations, and documentation with a high degree of accuracy.

Limitations

  • Numerical Computation: While enhanced for complex calculations, the model may require external computational tools for highly intricate numerical tasks.
  • Generalization: The model may not perform optimally on general knowledge topics outside its fine-tuned oil and gas domain.

License

This model is released under the Apache License 2.0. Please see the LICENSE file for more details.


Acknowledgments

Special thanks to NVIDIA AI Research for the development of the base model and to the Gain.Energy team for domain expertise and support in model fine-tuning and evaluation.


Contact Information

For support, inquiries, or collaboration opportunities, please contact:


model-index:

  • name: OGAI 3.1 Engineer results:
    • task: type: text-generation dataset: name: oil_gas_docs type: GainEnergy-OilGasDocs metrics:

    • task: type: text-generation dataset: name: technical_calculations type: TechnicalCalculations-OilGas metrics:

    • task: type: text-generation dataset: name: arena_hard type: arena_hard metrics:

    • task: type: text-generation dataset: name: alpaca_eval_2_lc type: AlpacaEval 2.0 Length Controlled metrics:

    • task: type: text-generation dataset: name: gpt_4_turbo_mt_bench type: gpt_4_turbo_mt_bench metrics:

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for GainEnergy/OGAI-3.1-Engineer