Projected Language Models: A Large Model Pre-Segmented Into Smaller Ones

Projected Language Models: A Large Model Pre-Segmented Into Smaller Ones

This paper has been accepted at the Foundation Models in the Wild workshop at ICML 2024.
Large language models are versatile tools but are not suitable for small inference budgets. Small models have more efficient inference but their lower capacity means that their performance can be good only if one limits their scope to a specialized domain. This paper explores how to get a small language model with good specialized accuracy, even when specialization data is unknown during pretraining. We propose a novel architecture, projected networks (PN). PN is a high capacity network whose parameters can be linearly projected into a small network for fine tuning. We assess the empirical effectiveness of our solution compared to small model training, distillation and hard mixture of experts.

More

Choose the Perfect QA Role

Hey there! Software development is a complex field, and...

Upcoming Events, October + November 2024

Hello, cartoon fans, I'm back with a whole lot...

Art’s Purpose – Animated Spirit

“Art is not, as the metaphysicians say, the manifestation...

Create Walking Talking Mario for Live Performance

https://www.youtube.com/watch?v=4-wWqYY8L3g Hello, I’m Scott Evans, founder of Digital Puppets, an...

Tyler Perry Pauses $800 Million Studio Expansion Due to Open AI’s Sora

March 18th, 2024: Renowned filmmaker and producer Tyler...