NanoChat Alpaca (Partial Fine-tuning)

Partially fine-tuned version of nanochat-speedrun-001 on the Alpaca instruction dataset.

Model Details

  • Base Model: royam0820/nanochat-speedrun-001
  • Fine-tuning Dataset: tatsu-lab/alpaca (52,002 examples)
  • Parameters: 561M
  • Vocab Size: 65,536

Training

  • Epochs: 3
  • Learning Rate: 1e-5
  • Loss: 3.83 โ†’ 3.24 โ†’ 2.95

Status

โš ๏ธ Partially trained - produces repetitive outputs. Needs more training.

Files

  • checkpoint_1.pt, checkpoint_2.pt, checkpoint_3.pt
  • tokenizer.pkl, token_bytes.pt
Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support