Model Card for Llama 3.1

Model Details

Model Developer: Shashank
Model Architecture: Llama 3.1, an auto-regressive language model utilizing an optimized transformer architecture. The instruction-tuned versions leverage supervised fine-tuning (SFT) and reinforcement learning with human feedback (RLHF) for improved alignment with human preferences.

Supported Languages: English, German, French, Italian, Portuguese, Hindi, Spanish, and Thai.

Model Release Date: August 10, 2024
Model Sizes: 8B, 70B, 405B parameters

Model Description

Llama 3.1 is a collection of multilingual large language models designed for a variety of natural language generation tasks. These models are optimized for multilingual dialogue use cases and excel in various industry benchmarks. The Llama 3.1 collection includes both pretrained and instruction-tuned versions.

How to Get Started with the Model

Use with Transformers

To use the model with the Transformers library:

import transformers

model_id = "shashankyadav03/Llama-3-8B-Instruct-Finance-Asset-Strategy"

pipeline = transformers.pipeline(
    "text-generation",
    model=model_id,
    model_kwargs={"torch_dtype": torch.bfloat16},
    device_map="auto",
)

messages = [
    {"role": "system", "content": "You are an expert finance asset advice expert"},
    {"role": "user", "content": "Should i invest in crypto in 2024"},
]

outputs = pipeline(
    messages,
    max_new_tokens=256,
)
print(outputs[0]["generated_text"][-1])

Tool Use with Transformers

Llama 3.1 supports tool use formats through chat templates in Transformers. You can define custom tools and integrate them into chat sessions.

Use with Llama Codebase

For usage with the original Llama codebase, follow the instructions provided in the repository.

Training Details

  • Training Data: Trained on ~15 trillion tokens from publicly available data, with fine-tuning on a mixture of human-generated and synthetic data.
  • Training Infrastructure: Utilized custom GPU clusters. The total computation involved 39.3M GPU hours on H100-80GB hardware.

Benchmark Scores

Llama 3.1 models excel across various benchmarks. For instance:

  • MMLU (5-shot): 69.4 (8B Instruct), 83.6 (70B Instruct), 87.3 (405B Instruct)
  • HumanEval (pass@1): 72.6 (8B Instruct), 80.5 (70B Instruct), 89.0 (405B Instruct)
  • ARC-Challenge: 83.4 (8B Instruct), 94.8 (70B Instruct), 96.9 (405B Instruct)

How to Use

Llama 3.1 is intended for commercial and research use in supported languages. It can be adapted for various NLP tasks, including text generation and assistant-like chat interfaces. The models also support the generation of synthetic data for improving other models.

Training and Evaluation Data

The pretraining data was sourced from publicly available datasets, with a cutoff in December 2023. Fine-tuning involved over 25 million synthetic examples.

Hardware and Software Requirements

The model was trained using high-performance GPUs, specifically H100-80GB units, and is compatible with the Transformers library starting from version 4.43.0.

Contact

For any questions or issues related to the model, contact shashankyadav03

Downloads last month
-
Safetensors
Model size
8B params
Tensor type
F16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Dataset used to train shashankyadav03/Llama-3-8B-Instruct-Finance-Asset-Strategy