"Shake" LLMs to make them better...?

This title could be clearer and more informative.Try out Clickbait Shieldfor free (5 uses left this month).

New AI research called 'neural thickets' reveals that after pre-training, large language model weights don't settle at a single optimal solution but reside in a dense region surrounded by alternative specialist models. By slightly perturbing weights with Gaussian noise ('jiggling'), the model can be nudged toward task-specific

1m watch time

Sort: