(2/4) Most ML people deal with silent errors and slow feedback loops via the "ratchet" approach: 1) Start with known working model 2) Record learning curves on small task (~1min to train) 3) Make a tiny code change 4) Inspect curves 5) Run full training after ~5 tiny changes
-
-
Know your baselines! Not beating a linear model or even the average of your labels is a cheap way to figure out you have a bug
-
My heart goes out to all my poor language models with a 50K vocab that train to worse than 50K perplexity.
Kraj razgovora
Novi razgovor -
-
-
seems like a clever trick suggested by
@woj_zarembahttps://twitter.com/woj_zaremba/status/1102291771332018176?s=20 … -
That's filthy. I love it.
- Još 1 odgovor
Novi razgovor -
-
-
@threadreaderapp compile -
Hi! there is your unroll: Thread by
@nottombrown: "(1/4) Learning ML engineering is a long slog even for legendary hackers like@gdb. IMO, the two hardest parts of ML eng […]" https://threadreaderapp.com/thread/1156350013552807936.html … Talk to you soon.
Kraj razgovora
Novi razgovor -
-
-
Hallo there is your unroll: Thread by
@nottombrown: "(1/4) Learning ML engineering is a long slog even for legendary hackers like@gdb. IMO, the two hardest parts of ML eng […]" https://threadreaderapp.com/thread/1156350013552807936.html … Enjoy :)
Kraj razgovora
Novi razgovor
Čini se da učitavanje traje već neko vrijeme.
Twitter je možda preopterećen ili ima kratkotrajnih poteškoća u radu. Pokušajte ponovno ili potražite dodatne informacije u odjeljku Status Twittera.