Skip to content

Latest commit

 

History

History
53 lines (27 loc) · 3.83 KB

README.md

File metadata and controls

53 lines (27 loc) · 3.83 KB

Nvidia-Hackathon

Drive : https://drive.google.com/drive/folders/1Fynid6c0kF7EHdXoO69GGbrSxJEQXWz3?usp=drive_link

  • This repository is the codebase of the Nvidia-Hackathon that me and my other team mate participated in during Oct 25 to Oct 27 2024.

Here is the simple description of the work we did:

Our goal was to enhance the accuracy of the Gemma 2B model for generating tags on the questions from the Law domain, for improved relevance search. We have used the Law Stack Exchange dataset from huggingface for this.

To achieve this, we utilized NVIDIA's NeMo Curator for faster data processing and NeMo microservices for fine tuning and evaluating the Large Language Model.

The data has a question, title, score and tags associated with it.

The key features that we have used for data processing from NeMo curator include:

  • DocumentDataset: This allowed us to handle our dataset efficiently with Dask for scalable processing.

  • UnicodeReformatter: We used this to encode the questions and titles into Unicode format.

  • DocumentModifier: We have wrapped our custom text cleaning and section number formatting into the DocumentModifier functionality.

  • PIIModifier: We have used this to remove personally identifiable information like names, phone numbers, addresses, locations and email addresses from questions and titles.

  • We have manually investigated questions with 10, 20, 40, 50 words and found that questions with 15 or fewer words are not contextually complete and meaningful enough. So we have discarded all the questions with 15 or fewer words using WordCountFilter.

  • ScoreFilter: We have used a score filter to exclude questions with a net upvote score below zero.

  • We also utilized the RepeatingTopNGrams filter to eliminate questions that contained repetitive phrases, as these can introduce bias by overemphasizing certain terms and phrases. We have the n value of 5 and the repeating percentage as 15%.

  • Further We have leveraged NeMo microservices to convert the downloaded Gemma 2B model into .nemo format so that the model is compatible with the NeMo framework and is also optimized to take advantage of Nvidia GPUs’ capabilities.

  • Our approach involved designing a custom system prompt that instructs the LLM what its task is, provided the question, title and the list of tags. We applied this to every question and title.

  • We used a 95% to 5% training-validation split for our dataset.

  • In fine-tuning our model, we carefully selected our hyperparameters to balance performance, efficiency, and resource utilization.

  • Our choice of bf16 precision strikes a balance between training speed and memory usage, crucial for handling our large language model.

  • The micro batch size of 1 and global batch size of 16 were chosen to optimize memory usage and gradient accumulation, allowing for stable training.

  • We implemented a cosine annealing learning rate schedule with a base rate of 1e-6. This approach helps in finding an optimal convergence point by gradually reducing the learning rate, preventing overshooting in later stages of training.

  • Infact, during our first trial of fine tuning the model, the training and validation loss are too far apart and also the validation is not converging while the training loss is, which is a clear sight of overfitting.

  • So, we introduced weight decay of 0.01 was selected to prevent overfitting and ensure smooth gradient updates.

  • We set our max steps to 2500 with evaluations every 200 steps. This frequent evaluation allowed us to monitor the model's performance closely and make necessary adjustments.

  • We have used optimizer beta values of 0.9 and 0.95 to help achieve faster convergence and improved stability during the fine-tuning of large language models by adjusting the decay rates of the gradient and variance estimates in the Adam optimizer.

We have used virtual instances on Brev for this hackathon.