Skip to content

๐Ÿ“– Full Stack Practice of the Large Language Model Training @ RLChina 2024

License

Notifications You must be signed in to change notification settings

davendw49/llm_training_full_stack

Folders and files

NameName
Last commit message
Last commit date

Latest commit

ย 

History

7 Commits
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 

Repository files navigation

๐Ÿ“– Hands-on LLM: A Full-Stack LLM Practise Training

A full stack practice to train a large language model @ RLChina 2024. The overview is shown as follows:

Overview of the tutorial

This is technical material suitable for LLM training engineers and researchers interested in LLM. That is the content here contains pieces of completed scripts and .ipynb-format-files to enable you to quickly training and using LLM.

note: The list of topics will be improved over time, the first version is used for the courses in RLChina 2024, Guangzhou.

No. Section Description Code Last Update Date
1 Data Curation This section covers the process of collecting, cleaning, and preparing datasets for LLM training. It ensures that the data is suitable and ready for model training. data_curation 2024-10-04
2 LLM Model Setup Here we explain how to configure and initialize the LLM architecture. This includes defining model parameters and preparing the environment for training. llm_model_setup 2024-10-04
3 LLM Pre-Training This section guides you through pre-training the LLM on large-scale datasets. It focuses on the initial phase where the model learns general language patterns. llm_pretraining 2024-10-04
4 LLM Post-Training Post-training involves fine-tuning the model for specific tasks or domains. This section walks through adjusting the pretrained model for enhanced performance. llm_posttraining 2024-10-04
5 LLM Deployment Learn how to deploy the trained model into production environments. This includes integrating the model with applications and optimizing performance. llm_deployment 2024-10-04
6 Resources and References This section provides additional resources, including papers, tutorials, and tools for LLM training and deployment. It's a helpful reference for further learning and exploration. resource_and_references 2024-10-04

License

Unless specified otherwise the code in this repo is licensed under Apache License, Version 2.0.

About

๐Ÿ“– Full Stack Practice of the Large Language Model Training @ RLChina 2024

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published