First steps, practice, learning fixes
Hugging Face defaults to
report_to="wandb"for metrics. I setreport_to=Noneto keep it local — no external deps, faster prototyping. Result: Still calling for a W&B API key.
Colab auto-started W&B despite
report_to=None.
I usedos.environ["WANDB_DISABLED"] = "true"+ runtime restart.
Result: Local training. No external deps.
GPT models don't have a padding token by default.
I settokenizer.pad_token = tokenizer.eos_tokento enable batch training.
Result: Garbage in = Gibberish out.
DialoGPT is autoregressive — needs
labels = input_idsto compute loss.
Without it:logits only.
I fixed tokenization + bumped epochs to 5 for 10-example dataset.
Result: Bot memorized my examples, but forgot how to respond to anything except exact training.
5 epochs on 10 examples → bot went silent.
Fix: Dropped to 1 epoch + usedchat_with_tuned_botrather thanchat_with_bot
Result:chat_with_tuned_botnot defined.
Lesson: — avoid overfittingchat_with_tuned_botwithtemperature=0.9. Result: Training on a small dataset of full conversation examples rather than next-token prediction doesn't work for this task.