Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Comprehensive Cookbook for Fine-tuning Gemma Model on Mental Health Assistant Dataset #43

Open
sitamgithub-MSIT opened this issue Feb 22, 2024 · 4 comments

Comments

@sitamgithub-MSIT
Copy link

The cookbook aims to provide a comprehensive guide for researchers and practitioners interested in fine-tuning the Gemma model from Google on a mental health assistant dataset.

Key components of the cookbook include an introduction to the Gemma model, a description of the dataset, preprocessing steps, fine-tuning Gemma using the Hugging Face Transformers library, training procedures, and usage examples. The cookbook will also provide guidance on training parameters, etc. This will help expand the available cookbook resources for leveraging Hugging Face models in various domains.

Usage Examples: Illustrative examples demonstrate how to use the fine-tuned Gemma model for various mental health support tasks, such as response generation.

@MKhalusova
Copy link
Contributor

Thank you for suggesting a new topic for the cookbook! A fine-tuning example of a large model can be an interesting technical notebook, however, the Mental Health use case can be a problematic example due to inherent ethical challenges of AI in sensitive and safety-critical domains such as medicine. I would suggest considering a different type of data to fine-tune the model on.

@sitamgithub-MSIT
Copy link
Author

Thank you for suggesting a new topic for the cookbook! A fine-tuning example of a large model can be an interesting technical notebook, however, the Mental Health use case can be a problematic example due to inherent ethical challenges of AI in sensitive and safety-critical domains such as medicine. I would suggest considering a different type of data to fine-tune the model on.

Thanks for your suggestion! Sure, I will change the dataset then. I am then thinking about fine-tuning the base 2B model with the DataBricks Dolly dataset. Same as what the Google team showed here. But of course, this is done by using Keras NLP, so I will implement the same using huggingface offerings. Is that okay, then?

@MKhalusova
Copy link
Contributor

@sitamgithub-MSIT sounds good to me! You can do this with transformers and peft. if you'd like, you can also add quantization into the mix.

@sitamgithub-MSIT
Copy link
Author

@sitamgithub-MSIT sounds good to me! You can do this with transformers and peft. if you'd like, you can also add quantization into the mix.

Yeah. Thanks for the suggestions! I will soon be starting to work on this.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants