Collaborate with Jaime Sevilla on datasets for various values related to size, performance, training expense, etc. of large machine learning models.
Having high quality data on this which one knows is going to be maintained makes it much easier to elicit forecasts about these topics, and eventually resolve those forecasts and keep track of track-records, and I know that Jaime has been working on this.
Collaborate with Jaime Sevilla on datasets for various values related to size, performance, training expense, etc. of large machine learning models.
Having high quality data on this which one knows is going to be maintained makes it much easier to elicit forecasts about these topics, and eventually resolve those forecasts and keep track of track-records, and I know that Jaime has been working on this.
We now have a first chart based on their pre-print here: Estimated computation used in large training runs of AI systems
Wohoo, nice!
I would really like to see this!