Model distillation transfers knowledge from large language models to smaller ones for efficiency. However, excessive distillation can lead to model homogenization and reduced capability in handling ...
Some results have been hidden because they may be inaccessible to you
Show inaccessible results