Distilling Wisdom: A Review on Optimizing Learning From Massive Language Models
<p dir="ltr">In the era of Large Language Models (LLMs), Knowledge Distillation (KD) enables the transfer of capabilities from proprietary LLMs to open-source models. This survey provides a detailed discussion of the basic principles, algorithms, and implementation methods of knowled...
Saved in:
| Main Author: | |
|---|---|
| Other Authors: | , , |
| Published: |
2025
|
| Subjects: | |
| Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Be the first to leave a comment!