If AGI is imminent, why can’t I hail a robotaxi?

Photo by gibblesmash asdf

My intuition is that driving is a domain narrow enough not to require AGI and, moreover, to require a system of far less sophistication and reasoning capabilities than an AGI. SAE Level 5 autonomy — which requires a vehicle to be able to drive autonomously wherever and whenever a typical human driver could — has not been achieved by any company. All autonomous driving projects currently require a human in the loop, either in the driver’s seat or available to provide remote assistance.

In a world where AGI is achieved by, say, 2030 or 2035, what are the odds Level 5 autonomy hasn’t been solved by 2023? My intuition is that we would expect autonomous vehicles to be a relatively low-hanging fruit that is plucked relatively early in the trajectory from AI solving video games to AI solving ~everything.

There are a few reasons why this intuition could be wrong:

  1. Maybe self-driving is actually an AGI-level problem or much closer to AGI-level than my intuition tells me. (I would rate this as highly plausible.)

  2. Maybe AI progress is such a steep exponential that the lag time between Level 5 autonomy and AGI is much shorter than my intuition tells me. (I would rate this as moderately plausible.)

  3. Perhaps Internet-scale data simply isn’t available to train self-driving AIs. (I would rate this as fairly implausible; it would be more much plausible if Tesla weren’t such a clear counterexample.)

  4. Robotics in general could prove to be either too hard or unimportant for an otherwise transformative or general AI. (I would rate this as highly implausible; it strikes me as special pleading.)

  5. Onboard compute for Teslas, which is a constraint on model size, is tightly limited, whereas LLMs that live in the cloud don’t have to worry nearly as much about the physical space they take up, the cost of the hardware, or their power consumption. (I would rate this as the most plausible objection, but I wonder why Tesla wouldn’t put a ton of GPUs in the trunk of a car and see if that works.)

  6. Self-driving cars don’t get to learn through trial-and-error and become gradually more reliable, whereas LLMs do. (I would rate this as somewhat plausible; the counterargument is that Tesla’s Autopilot is allowed to make mistakes, which humans can correct.)[1]

Please enumerate any additional reasons you can think of in the comments. Also, please present any arguments or evidence you can think of as to why I should accept any of the reasons given above.

  1. ^

    I owe both points (5) and (6) to a post by Daniel Kokotajlo.