- cross-posted to:
- [email protected]
- cross-posted to:
- [email protected]
You must log in or # to comment.
Redefines synthetic data generation? From “making fake data” to what? (I get it, it’s hyperbole.)
Training on synthetic data is the most garbage-sounding thing I’ve ever heard. Isn’t Reddit and Wattpad fake enough? At least those posts were generated by actual humans.
I guess I’ll wait and see, but I’m really skeptical that this will be a good thing.
Its a promising idea. Limitless training data in any domain you want. Whether or not it actually pans out is a whole different story.
340B is fucking huge, holy shit. How big is GPT-4?
The rumor is 1.76 trillion, or 8x220B (mixture of experts) to be specific: https://wandb.ai/byyoung3/ml-news/reports/AI-Expert-Speculates-on-GPT-4-Architecture---Vmlldzo0NzA0Nzg4