The race in AI should be toward models with fewer and fewer parameters for a given performance, not the opposite.
-
-
Vastauksena käyttäjälle @pmddomingos
While there are a few scientists training impractically large models, there are infinitely more engineers using distillation, pruning, quantization, and architecture optimization to get the same result with smaller and deployable models.
11 vastausta 25 uudelleentwiittausta 328 tykkäystä -
Vastauksena käyttäjälle @ylecun
It’s more than a few scientists. It’s a major trend in AI, spearheaded by large groups at several top AI labs.
2 vastausta 0 uudelleentwiittausta 52 tykkäystä -
Vastauksena käyttäjälle @pmddomingos
It's "a few" because (1) only "a few" have easy access to the required computing resources. (2) distributing large models on multiple GPUs and training them on many multi-GPU nodes requires a lot of infrastructure engineering. That's only available in a few industry labs.
2 vastausta 1 uudelleentwiittaus 30 tykkäystä
You’re confusing the head of the power law with the whole distribution. Everyone and their cat is trying to learn larger and larger models to the extent they can. Also, by Moore’s law etc. the requirements keep getting easier to meet.
Lataaminen näyttää kestävän hetken.
Twitter saattaa olla ruuhkautunut tai ongelma on muuten hetkellinen. Yritä uudelleen tai käy Twitterin tilasivulla saadaksesi lisätietoja.