Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Any guidance for further fine-tuning of LLaMA-Mesh? #16

Open
jackswl opened this issue Nov 30, 2024 · 2 comments
Open

Any guidance for further fine-tuning of LLaMA-Mesh? #16

jackswl opened this issue Nov 30, 2024 · 2 comments

Comments

@jackswl
Copy link

jackswl commented Nov 30, 2024

Great work!

Just wondering, if you can provide some form of guidance / fine-tuning template for further fine-tuning of LLaMA-Mesh for domain-specific meshes? (e.g., what exact chat template did you use for the SFT of your instruct model?) I am aware that the paper did mention about this, but having it spelt out explicitly will help tremendously especially in those possible pesky tokenization issues.

That would be really helpful.

Thanks!

@thuwzy
Copy link
Collaborator

thuwzy commented Dec 3, 2024

Thank you for your interest in our work! As for the training code, we use LLaMA-Factory, which is a very easy-to-use codebase for LLM traninig.

@jackswl
Copy link
Author

jackswl commented Dec 3, 2024

@thuwzy hi! thanks for your response. I mean, what kind of fine-tuning template did you use? Are you able to share the exact fine-tuning template? This is so I can conduct further fine-tuning by following your template.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants