Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add basic documentation pages #988

Open
wants to merge 22 commits into
base: main
Choose a base branch
from

Conversation

melissawm
Copy link

Adds an initial set up of documentation pages for MaxText, including a .readthedocs.yaml file. The project will need to be set up on the ReadTheDocs admin interface separately by the maintainers.

Copy link

@shivajid shivajid left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Some early reviews.

docs/getting_started/end-to-end.md Outdated Show resolved Hide resolved

## In-depth documentation

You can find in-depth documentation at [the MaxText GitHub repository](https://github.com/AI-Hypercomputer/maxtext/blob/main/docs/advanced_docs/).

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

This link is broken.

Copy link
Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Yes, these links (all that refer to docs/advanced_docs) will become true once this reorganization of the documentation is merged.

docs/index.md Outdated Show resolved Hide resolved
docs/index.md Outdated Show resolved Hide resolved
docs/index.md Outdated Show resolved Hide resolved
docs/index.md Outdated Show resolved Hide resolved
@@ -0,0 +1,84 @@
# Full Finetuninhg LLama2/LLama3 Optimized configuration

In the pre-training section you saw the steps on how to do pre-training with

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Need to rewrite this as we do not have the pre-training section missing.

docs/getting_started/full_finetuning.md Outdated Show resolved Hide resolved
docs/concepts/steps_model.md Outdated Show resolved Hide resolved

[^1]: Tf.data is the most performant way of loading large scale datasets.

You can read more about the pipelines in [Data Input Pipeline](https://github.com/AI-Hypercompute/maxtext/blob/main/docs/advanced_docs/Data_Input_Pipeline.md).

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Copy link
Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Same as above, this will be fixed once the PR is merged.

docs/index.md Outdated Show resolved Hide resolved
Comment on lines +3 to +5
In the pre-training section you saw the steps on how to do pre-training with
MaxText. To perform full fine tuning, you need to pass the checkpoint to the
training script.
Copy link
Author

@melissawm melissawm Oct 31, 2024

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

@shivajid how about:

Suggested change
In the pre-training section you saw the steps on how to do pre-training with
MaxText. To perform full fine tuning, you need to pass the checkpoint to the
training script.
MaxText can perform pre-training and full finetuning. To perform full fine tuning, you need to pass the checkpoint to the training script.

@melissawm melissawm marked this pull request as ready for review November 6, 2024 14:57
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants