Shimamura 70B

Model banner

Model Information

Shimamura-70B

70B parameters Creative / Funny Chat Model

This is a Finetune of zerofata/L3.3-GeneticLemonade-Unleashed-v3-70B to be a good Chat Model at a larger parameter size

This model has been trained on 100M tokens of Human chat logs from Bsky, 4chan & Most of all ShoujoAI.

Support me on Ko-Fi: https://ko-fi.com/deltavector

Quantized Versions

Available Downloads

Prompting

Model has been tuned with the Llama-3-Instruct formatting.

Samplers

For testing of this model, I used Temp=1, 0.1 Min-P.

See Axolotl Config

            https://wandb.ai/new-eden/austral/artifacts/axolotl-config/config-c61un0ze/v0/files/axolotl_config_cu4t7u4q.yml
            

Training

The training was done for 4 epoch using 8 x H200s GPUs for the fine-tuning of the model.

Credits

Thank you to Lucy Knada, Zerofata, Auri, Intervitens, Cgato, Kubernetes Bad and the rest of Anthracite.

Downloads last month
25
Safetensors
Model size
70.6B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for Delta-Vector/Shimamura-70B

Datasets used to train Delta-Vector/Shimamura-70B