MedQuAD LoRA r=32

Configuraci贸n

  • Base: mistralai/Mistral-7B-Instruct-v0.3
  • LoRA r: 32
  • M贸dulos: q_proj, k_proj, v_proj
  • 4-bit NF4
  • Early Stopping: patience=3

Entrenamiento

Training logs (manual, Epoch estimado, 150 max steps):

Step Epoch Training Loss Validation Loss
50 0.023 0.882800 0.845063
100 0.046 0.805700 0.784524
150 0.070 0.754400 0.767385

Evaluaci贸n

  • BertScore F1 (sample test 50): 0.8452

Uso

from peft import PeftModel
from transformers import AutoModelForCausalLM
base = AutoModelForCausalLM.from_pretrained('mistralai/Mistral-7B-Instruct-v0.3', load_in_4bit=True)
model = PeftModel.from_pretrained(base, 'CHF0101/medquad-lora-r32-best-v2')
Downloads last month
1
Inference Providers NEW
This model isn't deployed by any Inference Provider. 馃檵 Ask for provider support