forked from axolotl-ai-cloud/axolotl
-
Notifications
You must be signed in to change notification settings - Fork 0
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Feat/sharegpt multirole #1
Open
teknium1
wants to merge
60
commits into
NousResearch:main
Choose a base branch
from
axolotl-ai-cloud:feat/sharegpt_multirole
base: main
Could not load branches
Branch not found: {{ refName }}
Loading
Could not load tags
Nothing to show
Loading
Are you sure you want to change the base?
Some commits from the old base branch may be removed from the timeline,
and old review comments may become outdated.
Open
Feat/sharegpt multirole #1
teknium1
wants to merge
60
commits into
NousResearch:main
from
axolotl-ai-cloud:feat/sharegpt_multirole
Conversation
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
NanoCode012
force-pushed
the
feat/sharegpt_multirole
branch
from
February 22, 2024 13:13
13f31d8
to
33bcf57
Compare
* make mlflow optional * fix xformers don't patch swiglu if xformers not working fix the check for xformers swiglu * fix install of xformers with extra index url for docker builds * fix docker build arg quoting
* WIP conversion to use pydantic for config validation * wip, more fields, add capabilities * wip * update pydantic validation to match existing tests * tweak requirements * setup deprecated paams pydantic model * more validations * wrap up rest of the validations * flesh out the rest of the options from the readme into pydantic * fix model validators as class methods remember to return in validator missing return add missing relora attributes fix test for DictDefault change fix sys template for mistral from fastchat change in PR 2872 fix test for batch size warning * more missing attributes for cfg * updates from PR feedback * fix validation for datasets and pretrain datasets * fix test for lora check
* Add checkpoint logging to mlflow artifact registry * clean up * Update README.md Co-authored-by: NanoCode012 <[email protected]> * update pydantic config from rebase --------- Co-authored-by: NanoCode012 <[email protected]> Co-authored-by: Wing Lian <[email protected]>
* Add StableLM examples and configurations * Add FFT and LORA configuration files and modify readme with usage
* add lion-pytorch optimizer * update pydantic to support lion optimizer --------- Co-authored-by: Wing Lian <[email protected]>
* support user-defined prompt processing strategies for dpo * interpret dict dataset types as user-defined * fix lint errors * setup pydantic config for validation of User defined DPO --------- Co-authored-by: Wing Lian <[email protected]>
* Lora example for Mistral on MPS backend * Add some MPS documentation * Update examples/mistral/lora-mps.yml Co-authored-by: NanoCode012 <[email protected]> * Update examples/mistral/lora-mps.yml Co-authored-by: NanoCode012 <[email protected]> * Update README.md --------- Co-authored-by: NanoCode012 <[email protected]> Co-authored-by: Wing Lian <[email protected]>
seperated -> separated
* add gemma instruct chat template * support for chat tempalte strategy too
* add missing evals_per_epoch setting * more pydantic fixes * more fixes * move test from normalization to validation * increase eval size for sample packing tests
* run tests again on Modal * make sure to run the full suite of tests on modal * run cicd steps via shell script * run tests in different runs * increase timeout * split tests into steps on modal * increase workflow timeout * retry doing this with only a single script * fix yml launch for modal ci * reorder tests to run on modal * skip dpo tests on modal * run on L4s, A10G takes too long * increase CPU and RAM for modal test * run modal tests on A100s * skip phi test on modal * env not arg in modal dockerfile * upgrade pydantic and fastapi for modal tests * cleanup stray character * use A10s instead of A100 for modal
* plain input/output prompt strategy w/o chat templates * disable duplicate code check * make sure to add an eos/eot token to the end of the output so it will stop * multi turn segement support and test
* lora+ support * optimizer should default to None * include mit license
…r openai finetuning (#1361) * allow the sharegpt handler to also better handle datasets destined for openai finetuning * make sure to support system role
* add starcoder2 * Apply suggestions from code review Co-authored-by: NanoCode012 <[email protected]> * chore: lint * Apply suggestions from code review Co-authored-by: NanoCode012 <[email protected]> --------- Co-authored-by: Wing Lian <[email protected]> Co-authored-by: NanoCode012 <[email protected]>
* add docs * add docs * run linter
* add Jarvis cloud gpu and sponsorship * whitespace
* wip qlora + fsdp fixes * more fixes * make sure to load the lora 🤦 * only setup quantized meta on non-zero rank: * only run setup_quantized_peft_meta_for_training for qlora+fsdp * more fixes for qlora+fsdp * chore: lint * add example yml * support mistral too * fix for model_type and add mixtral support too * set cpu_offload: false to reduce vram, constrain new accleerator logic to qlora + fsdp * refactor for duplicate code
* validation for fsdp and deepspeed * make sure to return data
* Fix pydantic configuration for the max_memory input * chore: lint --------- Co-authored-by: Wing Lian <[email protected]>
* Add Glaive conversation format support * fix black formatting errors * Fix black and pylint formatting errors * only set role_key_tool if provided in the dataset constructor * Update src/axolotl/prompt_strategies/sharegpt.py Co-authored-by: Wing Lian <[email protected]> * sharegpt test * tokenizer test * fix formatting --------- Co-authored-by: Wing Lian <[email protected]>
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
Description
Motivation and Context
How has this been tested?
Screenshots (if appropriate)
Types of changes
Social Handles (Optional)