|
--- |
|
language: |
|
- ko |
|
base_model: |
|
- EleutherAI/polyglot-ko-1.3b |
|
pipeline_tag: text-generation |
|
tags: |
|
- ingredient |
|
- chatbot |
|
- review |
|
- usage |
|
license: apache-2.0 |
|
--- |
|
|
|
## 🐥 Base Model |
|
|
|
**EleutherAI/polyglot-ko-1.3b** |
|
|
|
This model is based on the Korean version of Polyglot-1.3B, an open-source language model released by EleutherAI. |
|
It is pre-trained on a large-scale Korean corpus and designed for general-purpose Korean language understanding and generation tasks. |
|
|
|
--- |
|
|
|
## Training Procedure |
|
|
|
### Training Hyperparameters |
|
|
|
The following hyperparameters were used during training: |
|
|
|
- `output_dir`: `./qlora_model_eleutherai` |
|
- `per_device_train_batch_size`: `2` |
|
- `gradient_accumulation_steps`: `4` |
|
- `total_batch_size`: `8 (2 x 4)` |
|
- `learning_rate`: `2e-5` |
|
- `num_train_epochs`: `2` |
|
- `fp16`: `True` |
|
- `logging_dir`: `./logs` |
|
- `logging_steps`: `5` |
|
- `save_steps`: `100` |
|
- `save_total_limit`: `1` |
|
- `load_best_model_at_end`: `True` |
|
- `metric_for_best_model`: `loss` |