Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

RL training of Llama 3.1 8b on multi-gpus #78

Open
wants to merge 77 commits into
base: main
Choose a base branch
from
Open

Conversation

AlexPiche
Copy link
Collaborator

@AlexPiche AlexPiche commented Oct 31, 2024

Training Llama 3.1 8b on multi-gpus

The goal of this PR is to train Llama 3.1 70b on GSM8k in an efficient manner.

@AlexPiche AlexPiche changed the base branch from main to grpo_wild_chat October 31, 2024 02:42
@rizar
Copy link
Collaborator

rizar commented Oct 31, 2024

will address #77 and #76

Base automatically changed from grpo_wild_chat to main October 31, 2024 13:28
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants