How AI in Context approaches thumbnails

I used an LLM to help draft this post, but I’ve edited/​rewritten it extensively and endorse it.

AI in Context is a channel about transformative AI and its risks, published by 80,000 Hours.

Writing up our current approach to thumbnails, which is nowhere near perfect, for easy shareability and cross-pollination of lessons. Would love to hear what other people are trying!

Making thumbnails

We’re lucky enough to have folks at 80k with great design instincts. We work with them as well as with some external folks, but finding great people is harder than we expected. Let us know if this is something you or someone you know would be great at!

We iterate way more than people expect

Every video gets ~dozens of thumbnail variations, most of which are made after launch. You can see the full set of data on our IABIED thumbnails here.

I believe 23 of our winning thumbnails (maybe all 3) were made after launch. It’s pretty hard to predict ahead of time which thumbnails will do well.

  1. We launch with a few thumbnails we’re excited about, a/​b/​c testing them

    1. We tried pre-launch testing via paid ads once, didn’t correlate well, but we haven’t tried super intensely

  2. We iterate from there. If one is doing well, or the video is doing well, we do new ones similar to what we’ve tried. If the views are lower or nothing is breaking out, we try more variance. This could be from thumbnails we had ready, or new ones we make with the new information. We also swap out titles, usually not at the same time so we get the full information.

  3. We have someone checking ~continuously through the first few days.

  4. Tests run for about six hours. For our view rate, tests stabilize around six hours. If you check at three hours and again at six, the numbers often shift meaningfully. After six, they tend to settle.

    1. YouTube’s AB testing data updates roughly every 30 minutes, not in real time. You’ll see the same numbers for a while and then a sudden jump.

    2. We have a Slack thread for every video where we dump every thumbnail iteration and every test result. It’s kind of unhinged and there’s probably a better way to do it.

  5. If one thumbnail is clearly tanking, we’ll cut the test early rather than wait for statistical significance

  6. When views are low and the test is running but not informative, sometimes the right move is to stop testing entirely for a while and just let the winning thumbnail run since the algorithm is sometimes looking for your audience.

What we’ve learned

  1. Big text is good

  2. Graphs are surprisingly strong

  3. Host face helps, but not always

    1. e.g. this is our winning MechaHitler thumbnail

  4. Having a “glowing” comment in the thumbnail (that is, grabbing a really positive comment on the video and putting it in the thumbnail) sometimes works but not always. Veritasium and 3B1B do this. We have a theory that it needs to be really specific.

    1. My recollection is this one didn’t do great:

No comments.