- Notifications
You must be signed in to change notification settings - Fork0
PrincySinghal/Html-code-generation-from-LLM
Folders and files
Name | Name | Last commit message | Last commit date | |
---|---|---|---|---|
Repository files navigation
Fine-tuning the Falcon 7B for the task of HTML code generation. The current model was selected based on its performance on complex reasoning benchmarks such as ARC and GSM8K and its compatibility with the available computational resources.
Usedhttps://huggingface.co/datasets/ttbui/html_alpaca dataset which contains:Size of data: 636 rows
- Instructions- user prompts (textual)
- Input-Further information needed as per the instruction could be html code or data points (textual+code)
- Response- empty
- Output- expected HTML code
- Model selection
- Dataset Preparation and Preprocessing
- Model Fine tuning script(setting hyperparameters and choosing fine tuning techniques and regularization)
- Model Evaluation
- API development to serve the model
- Understanding and implementing Parameter-Efficient Tuning (PeFT).
- Managing the computational complexity and memory limitations of large models.
- Ensuring reproducibility and consistency across training runs.
- Dealing with long training times and optimizing model runtime.
- Found ways to complete the training and evaluation without buying colab pro. Out of RAM error encountered during model training was solved by trying a different way to load my fine-tuned model rather than loading base model from scratch.
- Adopting PeFT techniques like LoRA.
- Utilizing quantization and model sharding to manage memory usage.
- Setting a random seed for train-test splitting to ensure reproducibility.
- Implementing precision training, early stopping and learning rate scheduling to improve convergence speed and solve GPU memory limitations.
- Regularization techniques such as dropout and scaling factor were applied.
- Training arguments were carefully set up to balance performance and resource usage.
1.learning_rate: 0.00022. train_batch_size:3. eval_batch_size: 84. seed: 425.gradient_accumulation_steps: 26.total_train_batch_size: 47.optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-088.lr_scheduler_type: cosine9.lr_scheduler_warmup_ratio: 0.0310.training_steps: 320
Model link-https://huggingface.co/PrincySinghal991/falcon-7b-sharded-bf16-finetuned-html-code-generation
- BLEU score: 0.01782
- LLMS more suited for code generation
- Hyperparamter tuning to improve low evalution score
About
Fine-Tuning and Evaluating a Falcon 7B Model for generating HTML code from input prompts.
Topics
Resources
Uh oh!
There was an error while loading.Please reload this page.