Zephyr 7B Gemma
					Collection
				
Models, dataset, and Demo for Zephyr 7B Gemma. For code to train the models, see: https://github.com/huggingface/alignment-handbook
					• 
				5 items
				• 
				Updated
					
				•
					
					15
This model is a fine-tuned version of google/gemma-7b on the HuggingFaceH4/deita-10k-v0-sft dataset. It achieves the following results on the evaluation set:
More information needed
More information needed
More information needed
The following hyperparameters were used during training:
| Training Loss | Epoch | Step | Validation Loss | 
|---|---|---|---|
| 0.9482 | 1.0 | 299 | 0.9848 | 
| 0.8139 | 2.0 | 599 | 0.9610 | 
| 0.722 | 2.99 | 897 | 0.9732 |