You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
⚠️ Please check that this feature request hasn't been suggested before.
I searched previous Ideas in Discussions didn't find any similar feature requests.
I searched previous Issues didn't find any similar feature requests.
🔖 Feature description
I would like to propose the addition of a new method called NEFTune to the Axolotl repository. NEFTune is a recent and highly effective augmentation technique for language model fine-tuning that has shown remarkable improvements in performance across various tasks without increasing training time significantly.
Motivation:
Currently, Axolotl relies on qlora, lora, and full fine-tuning methods. While these methods are effective, integrating NEFTune into the repository can further enhance the model's capabilities. NEFTune operates by adding noise to the embedding vectors during training, leading to substantial performance gains.
Benefits:
Performance Boost: Incorporating NEFTune can significantly improve the performance of the models in various evaluations, such as AlpacaEval.
Compatibility: NEFTune is designed to be compatible with existing fine-tuning methods, including qlora, lora, and full fine-tuning, ensuring ease of implementation.
Generalization: NEFTune has demonstrated the ability to improve model performance on a range of modern instruction datasets, including Evol-Instruct, ShareGPT, and OpenPlatypus.
Versatility: Even powerful models that undergo further refinement, such as those using RLHF like LLaMA-2-Chat, can benefit from the addition of NEFTune during training.
Evaluate the feasibility of integrating NEFTune into the Axolotl repository.
Implement NEFTune as an optional augmentation method in the training pipeline.
Provide clear documentation and guidelines for users on how to utilize NEFTune within the Axolotl framework.
Conduct thorough testing and validation to ensure that the integration does not negatively impact existing functionalities.
Monitor and maintain the NEFTune feature to keep it up-to-date with any changes in the Axolotl repository.
Additional Information:
I believe that incorporating NEFTune into the Axolotl repository can elevate the capabilities of the models and make them more competitive in various natural language understanding tasks. It aligns with the goal of continuous improvement and innovation in language model training.
I appreciate your consideration of this feature request and look forward to the potential enhancement of the Axolotl repository's capabilities.
❓ Alternatives
No response
📝 Additional Context
No response
Acknowledgements
My issue title is concise, descriptive, and in title casing.
I have searched the existing issues to make sure this feature has not been requested yet.
I have provided enough information for the maintainers to understand and evaluate this request.
The text was updated successfully, but these errors were encountered:
🔖 Feature description
I would like to propose the addition of a new method called NEFTune to the Axolotl repository. NEFTune is a recent and highly effective augmentation technique for language model fine-tuning that has shown remarkable improvements in performance across various tasks without increasing training time significantly.
Motivation:
Currently, Axolotl relies on qlora, lora, and full fine-tuning methods. While these methods are effective, integrating NEFTune into the repository can further enhance the model's capabilities. NEFTune operates by adding noise to the embedding vectors during training, leading to substantial performance gains.
Benefits:
Performance Boost: Incorporating NEFTune can significantly improve the performance of the models in various evaluations, such as AlpacaEval.
Compatibility: NEFTune is designed to be compatible with existing fine-tuning methods, including qlora, lora, and full fine-tuning, ensuring ease of implementation.
Generalization: NEFTune has demonstrated the ability to improve model performance on a range of modern instruction datasets, including Evol-Instruct, ShareGPT, and OpenPlatypus.
Versatility: Even powerful models that undergo further refinement, such as those using RLHF like LLaMA-2-Chat, can benefit from the addition of NEFTune during training.
Paper Reference:
NEFTune: Improved Language Model Fine-Tuning with Noisy Embeddings
✔️ Solution
Official Repo of the Method:
https://github.com/neelsjain/neftune
Request Implementation Steps:
Evaluate the feasibility of integrating NEFTune into the Axolotl repository.
Implement NEFTune as an optional augmentation method in the training pipeline.
Provide clear documentation and guidelines for users on how to utilize NEFTune within the Axolotl framework.
Conduct thorough testing and validation to ensure that the integration does not negatively impact existing functionalities.
Monitor and maintain the NEFTune feature to keep it up-to-date with any changes in the Axolotl repository.
Additional Information:
I believe that incorporating NEFTune into the Axolotl repository can elevate the capabilities of the models and make them more competitive in various natural language understanding tasks. It aligns with the goal of continuous improvement and innovation in language model training.
I appreciate your consideration of this feature request and look forward to the potential enhancement of the Axolotl repository's capabilities.
❓ Alternatives
No response
📝 Additional Context
No response
Acknowledgements
The text was updated successfully, but these errors were encountered: