Tentative clarification around supported LLM models #147
francoishernandez
started this conversation in
Show and tell
Replies: 1 comment 1 reply
-
Hi François, Iʻm keen to test fine-tuning Llama 2 or 3.1 with my bilingual datasets. In the recipe provided for wmt22_with_TowerInstruct-llama2 I donʻt see a yaml config file to train from the Llama model. Can you explain briefly how I can go about fine-tuning an LLM with my bilingual datasets? |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Context
LLMs have been moving fast. While we’ve been striving to stay ahead with various implementations along the way, it’s not always clear what models are fully supported.
To address this, I’m kickstarting a small framework to clarify and track model support. This initiative is a starting point for better transparency and collaboration within the eole community.
This is also a good place to start for potential contributors, outlining what kind of adaptations would be nice additions to support new models.
Also, keeping track of some relatively simple benchmark(s) allows for a replicable setup, facilitating the prevention of breaking changes. This is relatively critical in the context of eole, because of its unified architecture, sharing main components across various models.
So, what’s the plan?
Disclaimer
more_tokenizers
branch (Supporting HF tokenizers #122) with a few tweaks. I’m currently running on themain
branch to identify potential discrepancies.Observations
Most results are within a reasonable margin of the reference results. Slight differences are probably not alarming, considering the low absolute reliability of the MMLU benchmark.
That being said, there are a few very noticeable gaps between reference MMLU (grabbed from HF model cards mostly), and our numbers.
Tracking
Experiments will be tracked here, as well as in the specific Github Project.
This project is a draft, and might evolve. For instance, we could create some
model-support
repo to create clean issues and keep track of various model related topics. Also, it might move out of Github Projects for better extendability. We’ll see how this evolves.There are a few structural fields, notably:
Next steps
Beta Was this translation helpful? Give feedback.
All reactions