- Training code is https://github.com/tatsu-lab/stanford_alpaca
- Params were mostly default (as in stanford_alpaca README), except for: per_device_train_batch_size=1, per_device_eval_batch_size=1
- Fine-tuning dataset was based on https://github.com/tloen/alpaca-lora/raw/81eb72f707b0505a03b... with minor improvements; I'm going to publish my version soon
- The training itself took about 3 hours on 8x Nvidia A100 80GB