UNDERLINE DOI: https://doi.org/10.48448/e1kc-1605
technical paper
{M}erge{D}istill: {M}erging Language Models using Pre-trained Distillation
Would you like to see your presentation here, made available to a global audience of researchers?
Add your own presentation or have us affordably record your next conference.
