deepseeksft202507281606

This is a LoRA fine-tuned model based on deepseek-ai/deepseek-vl-7b-chat.

Model Details

  • Base Model: deepseek-ai/deepseek-vl-7b-chat
  • Fine-tuning Method: LoRA (Low-Rank Adaptation)
  • Training Data: Custom vision-language dataset
  • Model Size: large
  • Parameters: ~37,478,400 trainable parameters

Usage

from peft import PeftModel
from deepseek_vl.models import VLChatProcessor, MultiModalityCausalLM

# Load base model
base_model = MultiModalityCausalLM.from_pretrained("deepseek-ai/deepseek-vl-7b-chat")

# Load LoRA adapter
model = PeftModel.from_pretrained(base_model, "takashiyamamoto/deepseeksft202507281606")

# Load processor
processor = VLChatProcessor.from_pretrained("deepseek-ai/deepseek-vl-7b-chat")

Training Configuration

  • LoRA Rank (r): 16
  • LoRA Alpha: 32
  • Learning Rate: 0.0002
  • Epochs: 1
Downloads last month
1
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for takashiyamamoto/deepseeksft202507281606

Adapter
(4)
this model