Is there already a handy way to compare computation costs that went into training? E.g. compared to GPT3, AlphaZero, etc.?
I would love to know! If anyone finds out how many PF-DAYs or operations or whatever were used to train this stuff, I’d love to hear it. (Alternatively: How much money was spent on the compute, or the hardware.)
Is there already a handy way to compare computation costs that went into training? E.g. compared to GPT3, AlphaZero, etc.?
I would love to know! If anyone finds out how many PF-DAYs or operations or whatever were used to train this stuff, I’d love to hear it. (Alternatively: How much money was spent on the compute, or the hardware.)