Distilling Step-by-Step! Outperforming Larger Language Models with Less Training Data and Smaller Model Sizes

rw-book-cover

Metadata

Highlights

  • Distilling Step-by-Step! Outperforming Larger Language Models with Less Training Data and Smaller Model Sizes (View Highlight)