Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

feat: New Gen AI pattern - Llama2 Distributed Pre-training on Trn1 with RayTrain and KubeRay Operator #536

Merged
merged 13 commits into from
May 26, 2024

Conversation

vara-bonthu
Copy link
Contributor

@vara-bonthu vara-bonthu commented May 19, 2024

…rain and KubeRay

What does this PR do?

🛑 Please open an issue first to discuss any significant work and flesh out details/direction - we would hate for your time to be wasted.
Consult the CONTRIBUTING guide for submitting pull-requests.

- Adds a new pattern for Llama2 Distributed Pre-training on Trn1 with RayTrain and KubeRay Operator.

Motivation

- To provide a robust solution for distributed pre-training of Llama2 using AWS Trainium instances, leveraging the capabilities of RayTrain and KubeRay Operator for efficient and scalable training workflows.

More

  • Yes, I have tested the PR using my local account setup (Provide any test evidence report under Additional Notes)
  • Mandatory for new blueprints. Yes, I have added a example to support my blueprint PR
  • Mandatory for new blueprints. Yes, I have updated the website/docs or website/blog section for this feature
  • Yes, I ran pre-commit run -a with this PR. Link for installing pre-commit locally

For Moderators

  • E2E Test successfully complete before merge?

Additional Notes

@vara-bonthu
Copy link
Contributor Author

@5cp please review the PR. Thanks

@vara-bonthu vara-bonthu merged commit 5a2d1df into main May 26, 2024
34 of 36 checks passed
@vara-bonthu vara-bonthu deleted the reatrain-llama2-trn1 branch May 26, 2024 04:19
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

None yet

2 participants