Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Update docs #588

Merged
merged 15 commits into from
Jul 17, 2024
Merged

Update docs #588

merged 15 commits into from
Jul 17, 2024

Conversation

michaelbenayoun
Copy link
Member

@michaelbenayoun michaelbenayoun commented May 3, 2024

What does this PR do?

  • Updates the sections: instead of having one big section for tutorials, we now have "Training tutorials" and "Inference tutorials".
  • It removes the overview pages, which seem a bit redundant.
  • Updates the guide for finetuning LLMs as it was too much of a copy from the notebook before

@michaelbenayoun michaelbenayoun changed the base branch from sync_transformers_and_accelerate to main May 3, 2024 09:51
@HuggingFaceDocBuilderDev

The docs for this PR live here. All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.

@HuggingFaceDocBuilderDev

This issue has been automatically marked as stale because it has not had recent activity. If you think this still needs to be addressed please comment on this thread. Thank you!

1 similar comment
@HuggingFaceDocBuilderDev

This issue has been automatically marked as stale because it has not had recent activity. If you think this still needs to be addressed please comment on this thread. Thank you!

@michaelbenayoun michaelbenayoun changed the title Separate sections for tutorials Update docs Jul 15, 2024
@pagezyhf pagezyhf self-requested a review July 16, 2024 08:27
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

We should add 2 sections to this guide:

  • How to enable Sequence Parallelism?
  • How to enable Pipeline Parallelism?

Copy link
Member Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Let's do it in another PR to keep it small. The PR is here: #657

Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

TBD if we keep recommending running notebooks on EC2 instances!

Copy link
Member Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Yes, same, let's make handle that in another PR.

- Mistral models, such as [Mistral 7b (`mistralai/Mistral-7B-Instruct-v0.3`)](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2)
- Llama-2 models, such as [Llama-2 7b (`meta-llama/Llama-2-7b-hf`)](https://huggingface.co/meta-llama/Llama-2-7b-hf)

And many others!
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

WDYT of adding a section in https://huggingface.co/docs/optimum-neuron/en/package_reference/supported_models for the ones we support for training? and reference it here

Copy link
Member Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Why not but it is a bit "unclear". Because:

  1. We do not know which architecture is supported for regular training specifically.
  2. The supported architectures for distributed training is known, we can add that.

Copy link
Collaborator

@pagezyhf pagezyhf left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Overall LGTM!

@michaelbenayoun michaelbenayoun marked this pull request as ready for review July 17, 2024 14:33
Copy link
Collaborator

@dacorvo dacorvo left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM except a few nits, thanks for the cleanup !

docs/source/training_tutorials/finetune_llm.py Outdated Show resolved Hide resolved
docs/source/training_tutorials/finetune_llm.mdx Outdated Show resolved Hide resolved
@michaelbenayoun michaelbenayoun merged commit 2430f9f into main Jul 17, 2024
2 checks passed
@michaelbenayoun michaelbenayoun deleted the update_docs branch July 17, 2024 15:51
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

4 participants