HASS Llama 3.1 8B with Layernorms
This is a converted HASS (Eagle variant with layernorms) speculator checkpoint for Llama 3.1 8B, compatible with the speculators library.
Model Details
- Original checkpoint: nm-testing/Eagle_Speculator_Llama_3_1_8B_TTT
- Base model: meta-llama/Meta-Llama-3.1-8B-Instruct
- Architecture: HASS (Eagle with layernorms enabled)
- Size: 961MB (larger due to additional layernorm weights)
Usage
from speculators.models.eagle import EagleSpeculator
from transformers import AutoModelForCausalLM
# Load the HASS speculator
hass_model = EagleSpeculator.from_pretrained("nm-testing/hass-llama3.1-8b-layernorms")
# Attach verifier model
verifier = AutoModelForCausalLM.from_pretrained("meta-llama/Meta-Llama-3.1-8B-Instruct")
hass_model.attach_verifier(verifier)
# Use for speculative decoding
# ... your speculative decoding code here ...
Conversion Details
This checkpoint was converted using:
speculators convert --eagle --layernorms \
nm-testing/Eagle_Speculator_Llama_3_1_8B_TTT \
./hass-layernorms-converted \
meta-llama/Meta-Llama-3.1-8B-Instruct
License
Apache 2.0
- Downloads last month
- 178
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
๐
Ask for provider support