Transformer Model Translator is a Python project designed to translate text between languages using transformer-based models. The project provides flexibility to integrate with popular transformer architectures (e.g., BERT, GPT, T5) for language translation tasks.
- Transformer-based language translation
- Configurable language pairs and hyperparameters
- Easy dependency management with Poetry
- Python 3.6+
- Poetry (as the dependency manager)
-
Clone the repository:
git clone https://github.com/your-username/transformer-model-translater.git cd transformer-model-translater
-
Install dependencies using Poetry:
poetry install
-
Activate the environment:
poetry shell
-
Run the main to start training
The project uses a configuration file, config.yaml
, to define key parameters for the translation task. This includes:
- Source and target languages: Defines the language pair for translation (e.g., English to French).
- Batch size: The number of samples processed in each training/testing step.
- Learning rate: The step size for model optimization.
- Number of epochs: The number of times the model will iterate over the training dataset.