Listen to more EA content with The Nonlinear Library
Listen here: Spotify, Google Podcasts, Pocket Casts, Apple. Or, just search for “Nonlinear Library” in your preferred podcasting app.
We are excited to announce the launch of The Nonlinear Library, which allows you to easily listen to top EA content on your podcast player. We use text-to-speech software to create an automatically updating repository of audio content from the EA Forum, Alignment Forum, LessWrong, and other EA blogs.
In the rest of this post, we’ll explain our reasoning for the audio library, why it’s useful, why it’s potentially high impact, its limitations, and our plans. You can read it here or listen to the post in podcast form here.
Goal: increase the number of people who read EA research
An EA koan: if your research is high quality, but nobody reads it, does it have an impact?
Generally speaking, the theory of change of research is that you investigate an area, come to better conclusions, people read those conclusions, they make better decisions, all ultimately leading to a better world. So the answer is no. Barring some edge cases (1), if nobody reads your research, you usually won’t have any impact.
Research → Better conclusion → People learn about conclusion → People make better decisions → The world is better
Nonlinear is working on the third step of this pipeline: increasing the number of people engaging with the research. By increasing the total number of EA articles read, we’re increasing the impact of all of that content.
This is often relatively neglected because researchers typically prefer doing more research instead of promoting their existing output. Some EAs seem to think that if their article was promoted one time, in one location, such as the EA Forum, then surely most of the community saw it and read it. In reality, it is rare that more than a small percentage of the community will read even the top posts. This is an expected-value tragedy, when a researcher puts hundreds of hours of work into an important report which only a handful of people read, dramatically reducing its potential impact.
Here are some purely hypothetical numbers just to illustrate this way of thinking:
Imagine that you, a researcher, have spent 100 hours producing outstanding research that is relevant to 1,000 out of a total of 10,000 EAs.
Each relevant EA who reads your research will generate $1,000 of positive impact. So, if all 1,000 relevant EAs read your research, you will generate $1 million of impact.
You post it to the EA Forum, where posts receive 500 views on average. Let’s say, because your report is long, only 20% read the whole thing—that’s 100 readers. So you’ve created 100*1,000 = $100,000 of impact. Since you spent 100 hours and created $100,000 of impact, that’s $1,000 per hour—pretty good!
But if you were to spend, say 1 hour, promoting your report - for example, by posting links on EA-related Facebook groups—to generate another 100 readers, that would produce another $100,000 of impact. That’s $100,000 per marginal hour or ~$2,000 per hour taking into account the fixed cost of doing the original research.
Likewise, if another 100 EAs were to listen to your report while commuting, that would generate an incremental $100,000 of impact—at virtually no cost, since it’s fully automated.
In this illustrative example, you’ve nearly tripled your cost-effectiveness and impact with one extra hour spent sharing your findings and having a public system that turns it into audio for you.
Another way the audio library is high expected value is that instead of acting as a multiplier on just one researcher or one organization, it acts as a multiplier on nearly the entire output of the EA research community. This allows for two benefits: long-tail capture and the power of large numbers and multipliers.
Long-tail capture. The value of research is extremely long tailed, with a small fraction of the research having far more impact than others. Unfortunately, it’s not easy to do highly impactful research or predict in advance which topics will lead to the most traction. If you as a researcher want to do research that dramatically changes the landscape, your odds are low. However, if you increase the impact of most of the EA community’s research output, you also “capture” the impact of the long tails when they occur. Your probability of applying a multiplier to very impactful research is actually quite high.
Power of large numbers and multipliers. If you apply a multiplier to a bigger number, you have a proportionately larger impact. This means that even a small increase in the multiplier leads to outsized improvements in output. For example, if a single researcher toiled away to increase their readership by 50%, that would likely have a smaller impact than the Nonlinear Library increasing the readership of the EA Forum by even 1%. This is because 50% times a small number is still very small, whereas 1% times a large number is actually quite large. And there’s reason to believe that the library could have much larger effects on readership, which brings us to our next section.
Why it’s useful
EA needs more audio content
EA has a vibrant online community, and there is an amazing amount of well researched, insightful, and high impact content. Unfortunately, it’s almost entirely in writing and very little is in audio format.
There are a handful of great podcasts, such as the 80,000 Hours and FLI podcasts, and some books are available on Audible. However, these episodes come out relatively infrequently and the books even less so. There’s a few other EA-related podcasts, including one for the EA Forum, but a substantial percentage have become dormant, as is far too common for channels because of the considerable amount of effort required to put out episodes.
There are a lot of listeners
The limited availability of audio is a shame because many people love to listen to content. For example, ever since the 80,000 Hours podcast came out, a common way for people to become more fully engaged in EA is to mainline all of their episodes. Many others got involved through binging the HPMOR audiobook, as Nick Lowry puts it in this meme. We are definitely a community of podcast listeners.
Why audio? Often, you can’t read with your eyes but you can with your ears. For example, when you’re working out, commuting, or doing chores. Sometimes it’s just for a change of pace. In addition, some people find listening to be easier than reading. Because it feels easier, they choose to spend time learning that might otherwise be spent on lower value things.
Regardless, if you like to listen to EA content, you’ll quickly run out of relevant podcasts—especially if you’re listening at 2-3x speed—and have to either use your own text-to-speech software or listen to topics that are less relevant to your interests.
Existing text-to-speech solutions are sub-optimal
We’ve experimented extensively with text-to-speech software over the years, and all of the dozens of programs we’ve tried have fairly substantial flaws. In fact, a huge inspiration for this project was our frustration with the existing solutions and thinking that there must be a better way. Here are some of the problems that often occur with these apps:
They’re glitchy, frequently crashing, losing your spot, failing at handling formatting edge cases, etc.
Their playlists don’t work or exist, so you’ll pause every 2-7 minutes to pick a new article to read, making it awkward to use during commutes, workouts, or chores. Or maybe you can’t change the order, like with Pocket, which makes it unusable for many.
They’re platform specific, forcing you to download yet another app, instead of, say, the podcast app you already use.
Pause buttons on headphones don’t work, making it exasperating to use when you’re being interrupted frequently.
Their UI is bad, requiring you to constantly fiddle around with the settings.
They don’t automatically add new posts. You have to do it manually, thus often missing important updates.
They use old, low-quality voices, instead of the newer, way better ones. Voices have improved a lot in the last year.
They cost money, creating yet another barrier to the content.
They limit you to 2x speed (at most), and their original voices are slower than most human speech, so it’s more like 1.75x. This is irritating if you’re used to faster speeds.
In the end, this leads to only the most motivated people using the services, leaving out a huge percentage of the potential audience.
How The Nonlinear Library fixes these problems
To make it as seamless as possible for EAs to use, we decided to release it as a podcast so you can use the podcast app you’re already familiar with. Additionally, podcast players tend to be reasonably well designed and offer great customizability of playlists and speeds.
We’re paying for some of the best AI voices because old voices suck. And we spent a bunch of time fixing weird formatting errors and mispronunciations and have a system to fix other recurring ones. If you spot any frequent mispronunciations or bugs, please report them in this form so we can continue improving the service.
Initially, as an MVP, we’re just posting each day’s top upvoted articles from the EA Forum, Alignment Forum, and LessWrong. (2) We are planning on increasing the size and quality of the library over time to make it a more thorough and helpful resource.
Why not have a human read the content?
The Astral Codex Ten podcast and other rationalist podcasts do this. We seriously considered this, but it’s just too time consuming, and there is a lot of written content. Given the value of EA time, both financially and counterfactually, this wasn’t a very appealing solution. We looked into hiring remote workers but that would still have ended up costing at least $30 an episode. This compared to approximately $1 an episode via text-to-speech software.
On top of the time costs leading to higher monetary costs, it also makes us able to make a far more complete library. If we did this with humans and we invested a ton of time and management, we might be able to convert seven articles a week. At that rate, we’d never be able to keep up with new posts, let alone include the historical posts that are so valuable. With text-to-speech software, we could have the possibility of keeping up with all new posts and converting the old ones, creating a much more complete repository of EA content. Just imagine being able to listen to over 80% of EA writing you’re interested in compared to less than 1%.
Additionally, the automaticity of text-to-speech fits with Nonlinear’s general strategy of looking for interventions that have “passive impact”. Passive impact is the altruistic equivalent of passive income, where you make an upfront investment and then generate income with little to no ongoing maintenance costs. If we used human readers, we’d have a constant ongoing cost of managing them and hiring replacements. With TTS, after setting it up, we can mostly let it run on its own, freeing up our time to do other high impact activities.
Finally, and least importantly, there is something delightfully ironic about having an AI talk to you about how to align future AI.
On a side note, if for whatever reason you would not like your content in The Nonlinear Library, just fill out this form. We can remove that particular article or add you to a list to never add your content to the library, whichever you prefer.
Future Playlists (“Bookshelves”)
There are a lot of sub-projects that we are considering doing or are currently working on. Here are some examples:
Top of all time playlists: a playlist of the top 300 upvoted posts of all time on the EA Forum, one for LessWrong, etc. This allows people to binge all of the best content EA has put out over the years. Depending on their popularity, we will also consider setting up top playlists by year or by topic. As the library grows we’ll have the potential to have even larger lists as well.
Playlists by topic (or tag): a playlist for biosecurity, one for animal welfare, one for community building, etc.
Playlists by forum: one for the EA Forum, one for LessWrong, etc.
Archives. Our current model focuses on turning new content into audio. However, there is a substantial backlog of posts that would be great to convert.
Org specific podcasts. We’d be happy to help EA organizations set up their own podcast version of their content. Just reach out to us.
Other? Let us know in the comments if there are other sources or topics you’d like covered.
Who we are
We’re Nonlinear, a meta longtermist organization focused on reducing existential and suffering risks. More about us.
Footnotes
(1) Sometimes the researcher is the same person as the person who puts the results into action, such as Charity Entrepreneurship’s model. Sometimes it’s a longer causal chain, where the research improves the conclusions of another researcher, which improves the conclusions of another researcher, and so forth, but eventually it ends in real world actions. Finally, there is often the intrinsic happiness of doing good research felt by the researcher themselves.
(2) The current upvote thresholds for which articles are converted are:
25 for the EA forum
30 for LessWrong
No threshold for the Alignment Forum due to low volume
This is based on the frequency of posts, relevance to EA, and quality at certain upvote levels.
- EA Houses: Live or Stay with EAs Around The World by 15 Apr 2022 18:51 UTC; 223 points) (
- The most important lesson I learned after ten years in EA by 3 Aug 2022 12:28 UTC; 191 points) (
- 2021 AI Alignment Literature Review and Charity Comparison by 23 Dec 2021 14:06 UTC; 176 points) (
- We all teach: here’s how to do it better by 30 Sep 2022 2:06 UTC; 170 points) (
- 2021 AI Alignment Literature Review and Charity Comparison by 23 Dec 2021 14:06 UTC; 168 points) (LessWrong;
- Apply to >50 AI safety funders in one application with the Nonlinear Network [Round Closed] by 12 Apr 2023 21:06 UTC; 157 points) (
- Announcing: EA Forum Podcast – Audio narrations of EA Forum posts by 5 Dec 2022 21:50 UTC; 154 points) (
- The Parable of the Boy Who Cried 5% Chance of Wolf by 15 Aug 2022 14:33 UTC; 140 points) (LessWrong;
- App and book recommendations for people who want to be happier and more productive by 6 Nov 2021 17:40 UTC; 140 points) (LessWrong;
- Why fun writing can save lives: the case for it being high impact to make EA writing entertaining by 11 Nov 2021 22:17 UTC; 125 points) (
- Why fun writing can save lives: the case for it being high impact to make EA writing entertaining by 11 Nov 2021 22:17 UTC; 125 points) (
- Announcing Amplify creative grants by 3 Oct 2022 10:31 UTC; 100 points) (
- What is it like doing AI safety work? by 21 Feb 2023 19:24 UTC; 99 points) (
- The Parable of the Boy Who Cried 5% Chance of Wolf by 15 Aug 2022 14:22 UTC; 77 points) (
- Two Podcast Opportunities by 29 Dec 2021 14:03 UTC; 68 points) (
- A list of EA-relevant business books I’ve read by 21 Feb 2023 20:48 UTC; 66 points) (
- Perhaps the highest leverage meta-skill: an EA guide to hiring by 22 Aug 2022 10:18 UTC; 65 points) (
- Apply to >30 AI safety funders in one application with the Nonlinear Network by 12 Apr 2023 21:23 UTC; 65 points) (LessWrong;
- 14 Techniques to Accelerate Your Learning by 19 May 2022 13:59 UTC; 63 points) (LessWrong;
- Three common mistakes when naming an org or project by 23 Jul 2022 13:22 UTC; 61 points) (
- Impostor syndrome: how I cured it with spreadsheets and meditation by 10 Feb 2023 15:05 UTC; 58 points) (
- New Forum License: Creative Commons by 25 Oct 2022 21:37 UTC; 57 points) (
- What is it like doing AI safety work? by 21 Feb 2023 20:12 UTC; 57 points) (LessWrong;
- Two reasons we might be closer to solving alignment than it seems by 24 Sep 2022 20:00 UTC; 57 points) (LessWrong;
- Vitalik: Cryptoeconomics and X-Risk Researchers Should Listen to Each Other More by 21 Nov 2021 18:50 UTC; 56 points) (
- How and why to turn everything into audio by 11 Aug 2022 8:49 UTC; 55 points) (
- How and why to turn everything into audio by 11 Aug 2022 8:55 UTC; 54 points) (LessWrong;
- Hiatus: EA and LW post summaries by 17 May 2023 17:17 UTC; 51 points) (
- Doing good easier: how to have passive impact by 2 May 2022 14:48 UTC; 51 points) (
- 14 Techniques to Accelerate Your Learning by 15 Apr 2022 12:09 UTC; 46 points) (
- Two reasons we might be closer to solving alignment than it seems by 24 Sep 2022 17:38 UTC; 44 points) (
- Effective giving subforum and other updates (bonus Forum update November 2022) by 28 Nov 2022 12:51 UTC; 42 points) (
- Four reasons I find AI safety emotionally compelling by 28 Jun 2022 14:10 UTC; 39 points) (LessWrong;
- Would you like to run the EARadio podcast? by 29 Jan 2022 4:39 UTC; 38 points) (
- The $100,000 Truman Prize: Rewarding Anonymous EA Work by 22 Sep 2022 21:07 UTC; 37 points) (
- Living Nomadically: My 80/20 Guide by 8 Feb 2023 1:31 UTC; 37 points) (LessWrong;
- EA Dating Spreadsheet: see EA dating profiles by 21 Sep 2022 18:34 UTC; 33 points) (
- 28 Oct 2022 16:00 UTC; 32 points) 's comment on New book on s-risks by (
- Four reasons I find AI safety emotionally compelling by 28 Jun 2022 14:01 UTC; 32 points) (
- Impostor syndrome: how to cure it with spreadsheets and meditation by 9 Feb 2023 21:04 UTC; 30 points) (LessWrong;
- EA & LW Forums Weekly Summary (5th Dec − 11th Dec 22′) by 13 Dec 2022 2:53 UTC; 27 points) (
- EA Updates for November 2021 by 1 Nov 2021 11:15 UTC; 24 points) (
- Why write more: improve your epistemics, self-care, & 28 other reasons by 17 Dec 2022 19:25 UTC; 22 points) (LessWrong;
- Best introductory overviews of AGI safety? by 13 Dec 2022 19:04 UTC; 21 points) (
- Best introductory overviews of AGI safety? by 13 Dec 2022 19:01 UTC; 21 points) (LessWrong;
- 16 Nov 2021 11:37 UTC; 18 points) 's comment on Ngo and Yudkowsky on alignment difficulty by (
- Writing can save lives and get you an EA job: a list of reasons to write more EA content by 17 Dec 2022 19:07 UTC; 17 points) (
- What are the best EA materials for busy people? by 9 Jul 2023 16:45 UTC; 16 points) (
- Hiatus: EA and LW post summaries by 17 May 2023 17:17 UTC; 14 points) (LessWrong;
- 4 Aug 2022 6:19 UTC; 12 points) 's comment on Another call for EA distillers by (
- 11 Feb 2022 1:53 UTC; 8 points) 's comment on Text-to-speechifying the EA Forum by (
- 7 Sep 2023 10:04 UTC; 7 points) 's comment on Sharing Information About Nonlinear by (
- 1 Dec 2021 12:34 UTC; 7 points) 's comment on A list of EA-related podcasts by (
- EA & LW Forums Weekly Summary (5th Dec − 11th Dec 22′) by 13 Dec 2022 2:53 UTC; 7 points) (LessWrong;
- Text-to-speechifying the EA Forum by 11 Feb 2022 1:43 UTC; 6 points) (
- 6 Jan 2022 0:51 UTC; 6 points) 's comment on The Unweaving of a Beautiful Thing by (
- 19 Dec 2021 22:12 UTC; 5 points) 's comment on Supporting Video, Audio, and other non-text media on the Forum by (
- 23 Feb 2023 5:11 UTC; 5 points) 's comment on A list of EA-relevant business books I’ve read by (
- What are your recommendations for technical AI alignment podcasts? by 11 May 2022 21:52 UTC; 5 points) (LessWrong;
- 28 Oct 2022 16:03 UTC; 4 points) 's comment on New book on s-risks by (
- 22 Nov 2021 23:09 UTC; 3 points) 's comment on The Case for Reducing EA Jargon & How to Do It by (
- 16 Nov 2021 13:01 UTC; 3 points) 's comment on Why fun writing can save lives: the case for it being high impact to make EA writing entertaining by (
- 27 Feb 2023 20:44 UTC; 2 points) 's comment on How can we improve discussions on the Forum? by (
I would love to be able to listen to Open Philanthropy research reports in this way.
Whatever happens with the discussions about copyright, I really hope this continues to exist. I listened to six forum posts at 5am today while walking a baby around to sleep… very good for parental mental health
3 months on, and this has become one of the most valuable EA/Alignment/Rationality dissemination innovations I’ve seen. Has replaced almost all my more vapid listening. Would get through an extra 10-20 hours of content a week. Thank you Nonlinear/Kat/Emerson
I’m so glad to hear that! Shared it with the team to much party-parrot emoji reactions and totally made my day. Thank you for letting us know!
Thank you for creating this resource. I listened to the audio version of this post and was impressed with the quality of the speech synthesis. Virtually indistinguishable from a human, and with a very pleasant voice tone.
Yeah, they’ve gotten way better in the last year or so. A huge jump in quality.
I think a lot of people who were turned off by the robot-iness of TTS in the past should try again because they’ve come such a long way.
It does lose some of the prosody, compared with a human, but I agree it’s good (and improving).
Kat—are you able to explain whether you’ll moderate the content at all (and if so, how)?
We’re not really planning on doing any moderation. In some rare cases we might decide to remove something if we consider it a very clear info hazard.
However, even if we convert pretty much all EA content, I doubt we’d ever remove a lower bar of at least 5 upvotes, which I think would allow the community itself to moderate as opposed to leaving it in our hands at Nonlinear.
Are you getting author’s consent before turning their work into a podcast?
It’s a bot so it does it automatically. We figured since the vast majority of people will want to have their content read by more people that it made more sense to by default convert and let people opt out by filling out this form.
If someone deletes their original post, do you auto-remove it from the podcast as well? That would seem important to me.
Good idea! I’ll add that to the list of things to do.
I do think that there’s an interesting fuzzy boundary here between “derivative work” and “interpretative tool”.
e.g. with the framing “turn it into a podcast” I feel kind of uncomfortable and gut-level wish I was consulted on that happening to any of my posts.
But here’s another framing: it’s pretty easy to imagine a near-future world where anyone who wants can have a browser extension which will read things to them at this quality level rather than having visual fonts. If I ask “am I in favour of people having access to that browser extension?”, I’m a fairly unambiguous yes. And then the current project can be seen as selectively providing early access to that technology. And that seems … pretty fine?
This actually makes me more favourable to the version with automated rather than human readers. Human readers would make it seem more like a derivative work, whereas the automation makes the current thing seem closer to an interpretative tool.
As far as I understand, text to speech browser extensions do exist, but I haven’t tested their quality relative to this one.
Gears level info about TTS:
I feel like I am writing an Elon Musk tweet but in case anyone is interested:
As the post mentions, basically the speech quality comes from using the newer TTS models. From the sound, I think this is using Amazon Polly, the voice “Matthew, Male”.
I am an imposter, but I know this because I “made an app” for myself that does general TTS for local docs and websites.
The relevant code that produces the voices is two lines long. I did not check but I would be surprised if there was not a browser extension already.
If EAs think that a browser extension would be valuable, or want really any of the permutations of forum/comment or other services, my guess is that a working quality project and full deployment could be made for $30,000 or maybe as little as $3,000 (the crux is operational like handling payment, accounts, as well as interpretation of the value of project management)
If there is interest I think we can just put this in EAIF or the future fund or something.
Hear the audio version of this comment in “Matthew”, voice.
Hear the audio version of this comment in “Kevin” voice.
You’re right that it’s Matthew!
And they do already exist, they just have the limitations I mentioned in the blog (glitchy, no playlists, not multi-platform, etc). Here are the ones I use for various purposes
For articles on desktop, Natural Reader
For ebooks on Android, Evie
For articles on Android, @Voice
That is not how copyright works Kat!
If we had to ask each person before converting their text to audio, it just wouldn’t feasibly happen.
The way we thought about it was that >99.9% of people will be thrilled to have their writing read by more people. For the <0.01% who won’t, we made it easy for them to opt out, either for a particular article or for all their work. This way the whole community and hundreds of EAs can have access to great content and the <0.01% can also keep their writing in only written format.
This way everybody wins. :) I think the utilitarian case for it is strong.
Especially once it’s more known, people will know that if their post gets enough upvotes it’ll be converted, so they can request to not have it converted beforehand if they want. The potential harm is small and mitigated and the potential upside is huge.
These responses do seem curiously blithe about the question of whether or not this is legal.
FWIW I think I endorse Kat’s reasoning here. I don’t think it matters if it is illegal if I’m correct in suspecting that the only people who could bring a copyright claim are the authors, and assuming the authors are happy with the system being used. This is analogous to the way it is illegal, by violating minimum wage laws, to do work for your own company without paying yourself, but the only person who has standing to sue you is AFAIK yourself.
Not a lawyer, not claiming to know the legal details of these cases, but I think this standing thing is real and an appropriate way to handle
I’ve seen this reasoning a lot, where EA organisations assume they won’t get sued because the only people they’re illegally using the data of are other EAs, and and as someone whose data has been misused with this reasoning, I don’t love it!
This is a reason to fix the system! My point is that it reduces to “make all the authors happy with how you are doing things”, there is not some spooky extra thing having to do with illegality
TBC I do not endorse using people’s content in a way they aren’t happy with, but I would still have that same belief if it wasn’t illegal at all to do so.
And there are various things one could probably do to make it not illegal but still messed up and the wrong thing to do! Like make it mandatory to check a box saying you waive your copyright for audio on a thing before you post on the forum. I think if, like some of the tech companies, you made this box really little and hard to find, most people would not change their posting behavior very much, and would now be totally legal (by assumption).
but it would still be a bad thing to do.
My more detailed response is:
I am generally pro more audio content, and sceptical of the goodness of current copyright laws.
<0.01% seems overconfident to me.
Once something is up on the internet, it’s up forever. Taking it down post-facto doesn’t actually undo the damage.
You only need one person to sue you for things to go quite badly wrong.
As such, flagrantly violating the law on a fairly large scale (and the scale is an important part of the pitch here) seems like a dubious idea. Especially if you also go on public record in a way that suggests you know it’s illegal and don’t care.
<0.01% is definitely overconfident given at that point I had already expressed misgivings and we do not have 10,000 authors on the EA Forum.
(I’m not against my writing being podcastified in principle but I want to check out any podcast services who broadcast my work in advance to decide if I’m happy to be associated with them. I’m strongly against someone else making that decision for me.)
I think this isn’t actually correct – I think it depends a lot on the type of content, how likely it is to get mirrored, the data format, etc. E.g. the old Leverage Research website is basically unavailable now (except for the front page I think), despite being text (which gets mirrored a lot more).
Whether it actually goes ‘badly wrong’ depends on the type of lawsuit, the severity of the violation, the PR effects, etc. It’s probably good to err on the side of not violating any laws, and worth looking into it a bit before doing it.
I otherwise agree with your points!
I’m focusing on the prudential angle here, but I’d also be quite sympathetic to an author who was mad that their work was used in this way without their permission, even if it was later taken down.
How would it not be a copyright violation? Seems better to require consent, presumably via a checkbox or an email. Consent also could improve the average quality a bit. Although then the question is whether the EAF/LW/AF designers can be bothered including that kind of email, or checkbox+API, etc as a feature.
It seems like an extremely good use of time for EA forum designers (and others) to implement this. I think this is a fantastic project and I really hope this goes ahead. I also think it’s valuable for the EA community to take less start-up style shortcuts as the movement gets older and more well-known. I think the risk of giving the impression of being low-integrity or messy is high enough that regardless of whether copyright laws are good or not, it’s worth following them.
Possibly, it is enough to just have a disclaimer like “by submitting, you agree to have this turned into an audio format” to satisfy copyright laws?
Seems like a good idea if it were easy
Very minor point, but 100%-99.9% = 0.1%
Which part isn’t feasible? If you have the skill and capacity on the team to write something which will scrape forum posts, check if they have karma above a threshold, convert them to speech, and post them as audio, it seems more likely than not that you’d have the skill/capacity to edit the tool such that it DMs the authors of the posts it wants to scrape, asking them to reply “Yes” or “OK”, and then only uploads posts where the author did respond to the DM with such a reply.
If you think that most authors wouldn’t reply, so this would make the tool much less useful, then this seems like a different claim. Especially if you’re only doing recent posts, it seems somewhat unlikely that the authors will not see their DMs, which would mean that a lack of reply is not that unlikely to indicate a preference against inclusion.
If I were to receive such messages, I would likely fail to respond (unintentionally) at least 20% of the time.
This seems roughly consistent with “somewhat unlikely”, I expect the fraction is similar for me.
(Note that I strong upvoted because I think your concern is really about fairness, control and privacy, which copyright law doesn’t directly support.)
I think this is how copyright law effectively works.
Basically, considerations such as “fair use”, low damages/profit, and the fact the copyright process is initiated by the owner, means that it’s at least a grey area and in this case, it’s very possible they are safe.
For more details, here is some logic about “fair use”, copying and pasting a forum post (heh).
It’s an obscure, poorly formatted forum post from 2005, but I think the content is correct. See the criteria mentioned here in this Stanford page.
Maybe another way of seeing this is that the poster is unlikely to suffer damages and there’s not much profit for the company.
I think catehall or other lawyers have been helpful, please stomp this post if this is wrong.
I don’t think this analysis is right. The character of use may be educational, but on the other hand, I’m not sure you’re transforming it—you’re simply reproducing the text as audio. The nature of work goes beyond presenting public domain things like facts and ideas, by quoting the text as a whole. The amount used is maximal. The effect on market is substantial, in that it prevents the author from selling their writing as audio.
As for precedents. Well… Righthaven v. Hoehn established that it is OK to present a full editorial article in the setting of a noncommercial online post that discusses it. But here you’re just presenting the work in its entirety, and not using it for comment. Google Books scanning was deemed legal. But that was because they don’t represent the whole book. Another relevant example, although it was settled out of court, is that Audible bought rights to some books, and then got sued for allowing the text to be read aloud by TTS AI. If you don’t even own the text, then having the text read is I think a larger infringement.
So I doubt that narrating a whole post is fair use—rather it looks like a copyright infringement.
I could be totally wrong, it’s honestly ridiculous for me to give IP/tort advice.
However, I think the issue at stake is EA forum posts being turned into podcasts. But for EA forum posts, there’s no standard commercial profit from making posts (besides that sweet, sweet clout, which if anything might be increased by podcasting it?). So I don’t get it.
After checking, I’m confused about Righthaven v. Hoehn. Righthaven had some commercial interest to their content (if malicious/rentseeking). Yet it was dismissed on lack of standing. So I guess...this seems orthogonal to this issue?
Would love if you could do this for the EA Intro Fellowship syllabus (e.g. here’s one syllabus, but note that the syllabus is continuously updating between semesters and different universities use different syllabi)
This is cool! I have some specific thoughts and questions on the TTS software that you list for personal use, and how other TTS / podcast options might compete with this Library:
What I personally do to listen to an article I’m on is I say “Read it” to the Google Assistant on my phone, and it reads aloud any article I view on my phone in a really nice AI voice. They have a few nice voices to choose from, with different accents. I think it’s even a bit more human-like than the current one used by The Nonlinear Library, which is already pretty human-like. So I think I’m likely to stick to using my Google Assistant than the Library, although this could be useful for when I want to download and listen to EA articles when I’m offline.
My phone is a Google Pixel 4a running Android 11, but I think this read aloud feature is probably available on any Android phone with Google Assistant. Have you tried using it? If yes, what made you not mention it in this article? I haven’t tried using Evie, and I think I tried @Voice before and wasn’t pleased with it.
My guess is the Google Assistant is a bit better (in terms of ease of use and voice quality) than Evie or @Voice. It has these 2 other cons though (aside from not being usable offline), so maybe that’s why it might still be valuable for Nonlinear to make the Nonlinear Library:
There’s no way to queue a next article to read out loud
Once you close the app, it doesn’t save where you are at
Have you or others you know tried Speechify as a TTS? I know the founder. I tried it before and they have nice premium voices, but I find it somewhat expensive at around $140/year.
Do you pay for Natural Reader? If yes, do you generally recommend the paid plan?
I guess there’s less need now for the EA Forum podcast now that we have the Nonlinear Library?
Thanks for the thoughtful reply!
Re: google assistant, the way I use TTS is I queue up a lot of posts and listen to it while I’m getting ready in the morning or working out, etc, so having to select a new one each time would stop me from using it. Additionally I often am listening to long articles, so it losing my spot if I stopped listening wouldn’t work. Pocket has that glitch for me about 30% of the time, which is a huge reason I don’t use it.
Re: Speechify, I haven’t tried it! If somebody does and finds it good, please say so in the comments!
Re: Natural Reader, I do indeed pay. I found the 20 minutes a day limit prohibitive. I’d recommend it to other people if the cost isn’t too high for them and they’ll use it regularly. ($10/month).
Re: EA Forum podcast, I think there still could be use for that because lots of people will continue to have a strong preference for human voices. It would especially be useful for posts that are heavy in math or graphs which they would be able to describe out loud, like how the Astral Codex Ten podcast does it.
I use speechify, its voices are quite good but has the same formatting issues as all the rest (reading junk text) which I think is the real bottleneck here
Got it re: Google assistant, EA Forum podcast, and Speechify!
Re: Natural reader:
do you use the Premium or the Plus plan? The Premium plan is only $5/month if paid annually, while the Plus one is $9/month if paid annually.
Have you tried using the mobile app? I downloaded it just now. The reviews aren’t great but maybe the latest update fixed some issues. I don’t think it has the ability to let you queue next articles though, so I might try out @Voice again.
I’ve found this super useful over the past several months—thanks!
Am I alone in finding the text-to-speech voice really unpleasant?
I can recommend, instead, Freya’s voice and pacing at:
https://www.naturalreaders.com/online/
There’s even text-to-speech with a really natural sounding Snoop Dogg voice by a different company. These are all way more engaging options.
Thank you for making this! I’ve been using this generously the past few days. One suggestion I’ll make is I think it would be nice if the episode descriptions were something like:
“Title, Date
Beginning of post (50 words?)
Link to post”
Having the post links immediately available would be super helpful for reading the comments while listening.
Thanks for doing this! I’ve found it useful, and I expect that it will increase my engagement with EA Forum/LW content going forward.
OMG! I am SO happy to see this! I was considering developing something like this myself, which is my way of saying how much I’m excited about this project
Do you accept small donations?
I probably started ~20 podcast episodes already, it’s a big deal!
Finally catching up on comments after EAG and a vacation.
I’m so glad to hear you like it! Even happier you like it enough to donate!
We’re still pretty new and figuring out how to accept small donations. I’ll send you more information in a private message. Anybody else who’d like to give, just DM me.
This is great and very helpful. Looking forward to listening.
Re ‘a human reader’, we are doing this for EA Forum posts on the EA Forum Podcast … but only for selected posts
I also read and discuss some other EA-ish content (esp. relevant academic work) on my Found in the Struce podcast.
What a cool project! I listen to the vast majority of my reading these days and am perpetually out of good things to read.
The linked audio is reasonably high quality, and more importantly, it doesn’t have some of the formatting artifacts that other TTS programs have. Well done.
Your story for why this is a potentially high impact project is plausible to me, especially given how much you’ve automated. I have independently been thinking about building something similar, but with a very different story for why it could be worth my time to do it. That means this could be a different story for why your thing was good :), which I thought I’d share
My story was that the top-performing people in a given cause area are large fraction of the valuable work, if you buy power law type arguments. By definition, their time is a lot more valuable than average. But it is also more valuable for them to be better informed, because the changes they make to their decisions by being better informed are leveraged by their high work output or its consequences.
If you buy this story, I think you wind up focusing on figuring out how to cater what is audio-fied to what would be useful to the most productive people in EA. So like, what do top AI safety researchers wish they had time to listen to. I’d bet that this is actually a very different set of things than Forum/ LW posts.
When I started to do my thing, I suspected that a lot of the researchers who are doing the best work would benefit from being able to hear more academic papers, from arxiv for example. But IMO the key problem is that these don’t get read well because of formatting issues. I think this is a solvable problem, and have a few leads, but it was too annoying for me to do as a side project. DM me if you’re interested in chatting about that
Side point: this view of why this is high impact also speaks to letting the top people in question choose what they listen to, which looks more like an app that does TTS on demand than a podcast feed. This happens to avoid copyright issues, if the existence of other TTS apps is any indication.
You might be able to hack together an equivalent solution (on both copyright and customization) without needing to develop your own app by having a simple website that lets people log in and makes them a private RSS feed (compatible with most podcast players I think, though not confident in any of this). Then if they input a link on the website its compiled and added to their RSS feed for use in the player. If you had an api for calling your TTS script (and had solved these formatting issues) I or someone else could probably hack something like this website together pretty fast
(Sorry, when I said your story for impact was “plausible”, in my head I was comparing it to my own idea for why this would be good, and I meant that it was plausibly better than my story. I actually buy your pitch as written, seems like a solidly good thing; apologies)
After several months of use, I’ve found this tool incredibly useful and time-saving.
It seems like this feed is now capturing only a small portion of EA Forum posts meeting a 25 karma theshold or a 25 total positive votes threshold. (Not sure which of those thresholds footnote 2 is indicating you intended to use. But I notice that this post was included in the feed despite having <25 individual votes, while having >25 karma.)
E.g., this post wasn’t included, and some of the recent “MIRI conversations” posts weren’t captured.
I have become surprisingly dependent on this feed surprisingly quickly, so (a) I’m wondering if you could return it to capturing everything meeting the intended threshold and (b) I’m grateful to you for making this!
Could you include AI governance newsletters? In particular, I’d like to be able to listen to ChinAI, Import.ai, Charlotte Stix’s newsletter, and CSET’s newsletter. (Caveat that I haven’t read the thread about copyright and that I have no idea if these authors would be happy with their newsletters being added to this feed.)
Suggestion:
Distinguish or isolate the intro and conclusion about the Nonlinear Library from the main content.
Ideally, use a different voice (eg, a human recording). A common solution is to insert a chime. Either of those would require splicing audio files. If you want to keep to just splicing text files, maybe there’s a way of inserting something to make the TTL pause?
Amazon Polly, the backend that provides these recordings, lets you use “SSML”, basically a trivial annotation of text, that you can easily use to add pauses (as well as emphasis, pitch, volume, etc.).
Basically, what you’re asking for is probably a 2-3 line pull request. (Nonlinear might be using SSML, if not, it’s pretty easy to add it.)
It seems like on Spotify and Apple Podcasts, only episodes since Oct 20 are available? (But I think last week older things were available.) Do older eps automatically get deleted? Is there a way to find them still?
(Apologies if this was already covered in the post/comments—I only skimmed.)
Thanks for bringing this to our attention! That shouldn’t be happening. We’ll look into it.
listennotes.com provides a nice index of podcasts, down to episodes.
Suggestion: Consider claiming ownership, and updating it with at least Spotify & Google Podcast links (they’re parsing the Apple feed).
https://www.listennotes.com/podcasts/the-nonlinear-library-the-nonlinear-fund-xuqI3sgU5DY/
Podcast audience nationality estimation might hold some interest. I’m not sure of their methods or certainty range. https://www.listennotes.com/podcasts/the-nonlinear-library-the-nonlinear-fund-xuqI3sgU5DY/#audience Amazing, 5791 Nonlinear Library episodes published in the last two years.
Thanks for this, I’ve found this very helpful for consuming more EA Forum content. Are there ways to only have EA Forum posts on our podcast provider’s “subscriber” feed, as opposed to LW and AF posts too? Eg I find that if I have a lot of podcast episodes in my podcast feed, I am less likely to listen to any of them, as it’s harder to find the podcasts that I really want to listen to.
I suspect this could look like different podcast “feed”/profile with only EA Forum posts (and maybe similarly for AF and LW posts, eg they could have their own feed/profile with only podcast episodes for posts on those platforms).
Zapier have a guide on using their service for filtering an RSS feed. In your case you should be able to filter based off the
link
field (ie/link contains forum.effectivealtruism.org
), to create a new RSS feed.Dm me if you attempt this and are having any issues. Or if sufficient people upvote your comment (say 5+ votes) I’m happy to make a public RSS feed for this.
That’s very nice of you 🙂
I would really like this option for the alignment forum
I’ve noticed some question posts get included in the feed. I did find these useful, but maybe it’d be good to have the narrator or title explicitly note that this is a question post, and maybe it’d be good to have the narrator read the top answer or two?
Did you consider creating an audiobook from a new EA handbook? I think that it would be very useful, especially if the handbook is used during introductory fellowships...
I see you’ve started including some text from the post in each episode description, which is useful! Could you also include the URL to the post, at the top of the episode description? I often want to check out comments on interesting posts.
Maybe posts that were published before the Nonlinear Library started uploading but which have (say) >100 karma, got Forum prizes, and/or got LW curation should also be included in the feed?
(Apologies if this was already suggested in the post or comments—I only skimmed)
Is there a way to have this feed sorted by some rough sense of value?
Likewise is there a way to trim the intro off the episodes.
Somewhat similar ideas:
Maybe the start of the episode descriptions should be the post’s current karma at the time of automated upload (to help people decide what to read) and then the start of the post’s text (for the same purpose), rather than always “Welcome to the Nonlinear Library...”
Maybe the start of the narration could include the karma?
I also think some very brief intro and/or outro seems worthwhile, in part so I know that one episode has stopped and another has begun (as the narrator’s voice is always the same and the cadence doesn’t signal stopping or starting). But a shorter intro/outro might be better, e.g. just “Welcome to the Nonlinear Library. This is [title] by [author], published on [site]”, or that but without the first sentence.
As per the thread with Pablo, I think the podcast sounds pretty good. Having said that, I do have one small suggested improvement. When I look at the logo (a Sierpinski triangle), and think about what it’s supposed to represent, it makes me think of a pyramid, or of growing replicators “one person recruits three, and so on”. In particular, although this may seem kind-of unfair, it kinda reminds of this scene from the Office. Given that movement building is a major project of the org, that’s probably not the connotation that you want. I realize that most people aren’t going to think of this connotation, but I’m very curious of others’ thoughts, because even a few seems too many...
Uhh, cool! :)
Did you consider including some content from or info about the comments to the respective article? E.g. I could imagine it would safe me some time if such a reader would also read out top level comments above say 10 karma and how many sub-comments are below it. Roughly half of the time I want to check out the comment section after listening on Pocket and maybe half of those times would be covered with the above readouts.
Could post authors also get to first listen to what their post would sound like?
(For some posts, it might be perhaps difficult to know in advance whether the automatic narration would cause too many misunderstandings to be a net positive in audio form. This might be especially relevant for posts that were never meant to be “speech-like”, such as those that make heavy use of nested bullet points.)
I agree some posts will come out better in TTS than others. One thing we’re looking into doing is setting up some simple rules that will lead to some articles not being posted (e.g. a critical mass of images or numbers), or maybe a little disclaimer at the front of such articles along the lines of “This article has a lot of numbers in it, so you might want to consider reading the original at [Source]”.
Regardless, if ever anybody feels like the article has come out too bad to be net positive, they can fill out this form (t.ly/G73f) and we’ll remove it. Or, if you think in advance it’s not going to be good, just let us know (through the form or our contact form t.ly/YBIr) and we can make sure it never goes up. :)
That disclaimer for technical articles sounds good. :) Also, yeah, perhaps the authors themselves can pre-test their potentially tricky articles with some (similar enough) TTS, and then decide whether to opt out of the library. (Perhaps disclaimers could also exist for articles with a sufficient amount of hyperlinks, since many people use those instead of only explicit in-text references.)
Interesting! Of course, the experience might be, in some ways, quite confusing compared to a human narration. For example, the automatic narration does not seem to separate headings or quotes from the main text. Could the AI be taught to identify headings and quotes, and make them stand out?
(E.g., headings might be ideally narrated with longer pauses, and quotes perhaps even in a different voice.)
Another difference between automatic vs. human narration: The automatic narration does not notify the listener whenever they might miss some meaningful hyperlink or figure in the text (unlike e.g. Holden Karnofsky’s self-narration of the Cold Takes Audio). Of course, this is part of the quantity over quality tradeoff; I’m just wondering what would be ideal and still possible to implement automatically. :)
Exciting! But where’s the podcast’s URL? All I can find is a Spotify link.
Edit: I was able to track it down, here it is https://spkt.io/f/8692/7888/read_8617d3aee53f3ab844a309d37895c143
Here it is! Spotify, Google Podcasts, Pocket Casts. Or just search for it in your preferred podcasting app. We put it on all the biggest ones. Just let us know if it’s not on one and it’ll be easy enough for us to add it.
Added clarification at the top of the post to make it easier to find. :)
I would love to see an index linking to initial text posts and to podcast link(s)!
As explanation, I listen to podcasts while commuting, and often numerous episodes of Nonlinear Library (NL) because they are so short. I sometimes want to engage further, by re-reading them, saving them in a reading manager, checking the author’s other work, or reading comments (or adding my own). It would be helpful to have somewhere that links to text posts.
If I remember correctly, NL made an intentional choice not to link from podcast episode descriptions to EA/LW/AF, in hopes of avoiding a deluge of engagement with posts, or communities as a whole, that do not uphold conversational and engagement norms that we so prize them for.
Should, or could, we request read-only links to posts?
might support sharing these posts more broadly beyond these communities, while minimizing the risk of unwelcome engagement (and supporting disclaimers to serve as an introduction to the community—”these posts are not endorsed,” or “this is a forum”).
Linking to the original post is the most requested feature! Unfortunately because of obscure technical reasons it’s quite difficult. We did to figure out a way to do it for our channels on Spotify and some of the sub channels across all podcast players, but not the main channel.
We definitely did not choose to not link to it due to worries about audience engagement. It’s just because podcast players are astoundingly bad. Also our particular setup makes it challenging
There are some recent posts, for example this one that are just the intro and outro (22 seconds long) and miss the main post. Would be great if this bug could be fixed.
Thank you for pointing that out! We’re working on it.
Please include links to originals in podcast episode details.
Content request:
https://www.glowfic.com/posts/4582?view=flat
Posted by Eliezer Yudkowsky a few days ago:
https://www.facebook.com/yudkowsky/posts/10160074795314228
I’d be happy to help formatting the text if you want
Will you tag each EA forum post that is added to the Nonlinear library with the ‘audio’ tag? I want to make sure we don’t duplicate these on the EA Forum Podcast or on Found in the Struce or elsewhere.
Of course sometimes a human reading might be worth doing on top of a machine-read one… but probably better to first focus on spreading the audio widely.
If these posts are going to be tagged, I suggest using a special tag besides, or instead of, the ‘audio’ tag.
Why, because it’s a nonhuman voice?
No, just because the tag would include a huge number of posts. It seems like a good heuristic to have dedicated tags in these cases. Moreover, the tag can be removed if we eventually don’t find it useful, whereas the reverse seems more complicated. (I emphasize that we can also tag these posts with ‘audio’; I’m only proposing that we use a special tag regardless of what other tags we use.)
I’m quite liking this tool, and found myself consuming more content in both EA and LW.
Quick suggestion that I think could be helpful. I noticed many of the posts are quite short, less than 1 min long, might contain a few links, or be a quick question for commenters, making them less useful for listeners. I’m not sure what filter could be additionally applied, and considering it could be valuable
This is awesome! I did something similar for Astral Codex Ten (feed, post) a while back. The human version is also good, if you like that kind of thing.
Why have a digital version if there is already an AC10 podcast with a human reader?
Originally I made a digital SSC podcast (feed) so that I could listen through the back catalog of posts (the human reader version didn’t start until ~2017). I ended up getting used to the robot narrator, so I just kept it running on ACX. One small upside is that the digital versions get created within minutes of new posts.
This is great and you should make a LW post; these are in a really nice format for shunting around.
As a small nit: any idea why the first few essays of the Codex (https://www.lesswrong.com/codex) are not here?
That makes sense. Still, I’m wondering if it might not be better to just continue the digital versions and put a link to the human reader ones, to clutter things up a little bit less. Or you might keep it up but put a link to the human reader version for people who might not know about it. It is certainly possible that some people actually prefer hearing a machine than a (particular) human. (Although his production value is very high, higher than mine was).
By the way, as you probably know there’s a reason why he doesn’t record these until the day after the blogs come out: they are often adjusted/adapted and he wants to give some space for that to happen. On the other hand you could argue that people will just choose the version they prefer and we can trust the consumer here