80k hrs #88 - Response to criticism
Iâm a regular listener of the 80k hrs podcast. Our paper came up on the episode with Tristan Harris, and Rob encouraged responses on this forum so here I go.
Update: apologies, but this post will only make sense if you listen to this episode of the 80k hrs podcast.
Conflict can be an effective tactic for good
I have a mini Nassim Taleb inside me that I let out for special occasions đ . Iâm sometimes rude to Tristan, Kevin Roose and others. Itâs not just because Tristan is worried about the possible negative impacts of social media (iâm not against that at all). It is because he has been one of the most influential people in building a white hot moral panic, and frequently bends truth for the cause.
One part he gets right is that, trigger a high-reach person into conflict with you, serves to give your message more attention. Even if they donât reply, you are more likely to be boosted by their detractors as well. This underdog advantage isnât âhateâ, and the small advantage is massively outweighed by institutional status, finances and social proof. To play by gentlemans rules is to their advantageâcurtailing the tools in at my disposal to makes bullshit as costly as possible.
I acknowledge there are some negative costs to this (e.g. polluting the information commons with avoidable conflict), and good people can disagree about if the tradeoff is worth it. But I believe it is.
Was Tristan & tech media successful in improving the YouTubeâs recommendation algorithm?
Iâll give this win to Tristan and Roose. I believe YouTube did respond to this pressure when in early 2019 they reduced recommendations to conspiracies and borderline content and this was better overall, but not great.
But YouTube was probably never as they describedâa recommendation rabbit hole to radicalization. If it was, there was never strong evidence to support it.
The YouTube recommendation has always boosted recent, highly watched videos, and has been through 3 main phases:
Clickbait Phase: Favoured high click-through rates on video thumbnails. This meant that clickbait thumbnails, were very âtabloidyâ and edgy, and frequently misrepresented the content of the video. But no-one ever showed that this influenced users down an extremist rabbit holeâthey just asserted it, or used very week attempts at evidence.
View-Neutral Phase: Favoured videos that people watch more of, and rated highly after watching. This was a big improvement for quality recommendations. They hadnât started putting their thumb on the scales, so the recommendations largely matched the portion of views for a video.
Authoritative Phase: Favours traditional media, especially highly partisan cable news. Very little recommendations to conspiracies and borderline content. This was announced early 2019, and deployed in April 2019.
Tristan regularly represents todayâs algorithm as a radicalization rabbit hole. His defence that critics are unfair because the algorithm changed after he made the critique is wrong. He didnât make any effort to clarify on the Social Dilemma (Released Jan 2020), or in his appearances about it, and hasnât updated his talking points. For example, speaking on the Joe Rogan Podcast in October 2020 he said: âno matter what I start with, what is it going to recommend next. So if you start with a WW2 video, YouTube recommends a bunch of holocaust denial videosâ.
Whatâs the problem with scapegoating the algorithm and encouraging draconean platform moderation ?
Tristanâs hyperbole sets the stage for drastic action. Draconian solutions for misdiagnosed problems will probably have unintended consequences that are worse than doing nothing. I wrote about this in regards to the the QAnon crackdown:
Demand for partisan conspiracy content is strong, which will be supplied by the internet in one way or another. There is big movement towards free speech platforms due to moderation, which (due to selection effects) are intense bubbles of far right and conspiracy content
Content moderation is building grievances that will not be easily placated. YouTubeâs moderation removes more than 1000x right vs left videos. With the current trajectory political content may end up largely separated into tribal platforms.
Scapegoating the scary algorithm, or adopting a puritanical approach to moderation will work against more practical and effective actions.
The anonymous user limitation of YouTube studies
Itâs technically quite difficult to analyse the YouTube algorithm that includes personalization. Our study was the most rigorous and comprehensive look at the recommendations political influence at the time, despite the limitation of collecting non-personalized recommendations. To take the results at face value, you need to assume this will âaverage outâ to about the same influence once aggregated. I think itâs an open question, but itâs reasonable to assume the results will bein the same ballpark.
My experience with critics who point to this as a flaw or reason to ignore the results, are inconsistent with their skepticism. The metrics that Tristan uses in this podcast (e.g. ârecommended flat Earth videos hundreds of millions of timesâ) are based on Gualliam Chaslots data, which is also based on anonymous recommendations. I also am skeptical about these results:
- These figures are much higher that what we see and Chaslot is not transparent about how these numbers have been calculated
- Chaslots data is based on the API, which gives distorted recommendations compared to our method of scraping the website (much closer to real-world).
This quality of the research is quickly improving in this space. This most recent study uses real-world user traffic to estimate people following recommendations from videos. A very promising approach once they fix some issues.
We have been collecting personalized recommendations since november. We are analysing the results and will present there on transparency.tube and a paper in the coming months. I hope tristan and other prominent people will start update the way they talk about YouTube based on the best and latest research. If they continue to misdiagnose problems, the fervor for solutions they whip up will be misdirected.
What are the most effective ways we can address problems from social media
I have a narrow focus on the mechanics of YouTubeâs platform, but Iâll give my intuitional grab bag of ideas that are the most promising to reduce the bad things about social media:
Building, or popularising apps/âextensions that can be used by people to control their own future behaviour towards their higher order desires. The types of apps that Rob suggested are great and some are new to me (i.e. Inbox when ready, todobook, freedom, news feed eradicator).
Platform nudges like adjusting recommendations and information banners on misinformation with research into the effectiveness of these interventions
Grassroots efforts to cool down partisanship, like BraverAngels, with research into measuring the impact of these.
And for the easy one đ . Addressing of corruption and decay in institutions (e.g. academia, media, politics), lack of state capacity, low growth, negative polarization and income inequality. Just fix those things and social media will be much less toxic.
Robâs did some really good background research and gently pushed back in the right places. The best interview with Tristan I have listened to.
- 2 Aug 2021 14:33 UTC; 33 points) 's comment on roÂhinÂmshahâs Shortform by (LessWrong;
Thanks for writing this. I havenât (yet) listened to the podcast and thatâs perhaps why your post here felt like I was joining in the middle of a discussion. Could I suggest that at the top of your post you very briefly say who you are and what your main claim is, just so these are clear? I take it the claim is that YouTubeâs recommendations engine does not (contrary to recent popular opinion) push people towards polarisation and conspiracy theories. If that is your main claim, Iâd like you to say why YouTube doesnât have that feature and why people who claim it does are mistaken.
(FWIW, Iâm an old forum hand and Iâve learnt you canât expect people to read papers you link to. If you want people to discuss them, you need to make your main claims in the post here itself.)
In general I agree, but the forum guidelines do state âPolish: Weâd rather see an idea presented imperfectly than not see it at all.â, and this is a post explicitly billed as âresponseâ that were invited by Rob. So if this is all the time Mark wants to spend on it, I feel it is perfectly fine to have a post that is only for people who have listened to the podcast/âare aware of the debate.
Oh, what I said wasnât a criticism, so much as a suggestion to how more people might get up to speed on whatâs under debate!
Thanks for posting this! I thought it was interesting, and I would support more people writing up responses to 80 K podcasts.
Minor: you have a typo in your link to transparency.tube
Thankyou đ Fixed.
Thanks for writing this and for your research in this area. Based on my own read of the literature, it seems broadly correct to me, and I wish that more people had an accurate impression of polarization on social media vs mainstream news and their relative effects.
While I think your position is much more correct than the conventional one, I did want to point to an interesting paper by Roâee Levy, which has some very good descriptive and casual statistics on polarization on Facebook: https://ââpapers.ssrn.com/ââsol3/ââpapers.cfm?abstract_id=3653388. It suggests (among many other interesting findings) that Facebook probably is somewhat more slanted than mainstream news and that this may drive a small but meaningful increase in affective polarization. That being said, itâs unlikely to be the primary driver of US trends.
Hi Mark, thanks for writing this post. I only had a cursory reading of your linked paper and the 80k episode transcript, but my impression is that Tristanâs main worry (as I understand it) and your analysis are not incompatible:
Tristan and parts of broader society fear that through the recommendation algorithm, users discover radicalizing content. According to your paper, the algorithm does not favour and might even actively be biased against e.g conspiracy content.
Again, I am not terribly familiar with the whole discussion, but so far I have not yet seen the point made clearly (enough), that both these claims can be true: The algorithm could show less âradicalizingâ content than an unbiased algorithm would, but even these fewer recommendations could be enough to radicalize the viewers compared to a baseline where the algorithm would recommend no such content. Thus, YouTube could be accused of not âdoing enoughâ.
Your own paper cites this paper arguing that there is a clear pattern of viewership migration from moderate âIntellectual Dark Webâ channels to alt-right content based on an analysis of user comments. Despite the limitation of using only user comments that your paper mentions, I think that commenting users are still a valid subset of all users and their movement towards more radical content needs to be explained, and that the recommendation algorithm is certainly a plausible explanation. Since you have doubts about this hypothesis, may I ask if you think there are likelier ways these users have radicalized?
A way to test the role of the recommendation algorithm could be to redo the analysis of the user movement data for comments left after the change of the recommendation algorithm. If the movement is basically the same despite less recommendations for radical content, that is evidence that the recommendations never played a role like you argue in this post. If however the movement towards alt-right or radical content is lessened, it is reasonable to conclude that recommendations have played a role in the past, and by extension could still play a (smaller) role now.
I agree you can still criticize YouTube, even if they are recommending conspiracy content less than âview-neuralâ. My main disagreement is with the factsâTristan us representing YouTube is as a radicalization pipeline caused be the influence of recommendations. Letâs say that YouTube is more radicalizing than a no-recommendation system all-things-considered because users were sure to click on radical content whenever it appeared. In this case you would describe radicalization as a demand from users, rather than a radialization rabbit hole caused by a manipulative algorithm. Iâm open to this possibility, I wouldnât give this much pushback if this is what is being described.
The algorithmic âAuditing Radicalization Pathways on YouTubeâ paper is clever in the way they use comments to get at real-world movement. But that paper doesnât tell us much given that a) they didnât analyse movement form right to left (one way movement tell you churn, but nothing directional) and b) thy didnât share their data.
The best I have seen is this study, which uses real-world web usage forma representative sample of users to get the real behaviour of users who are clicking on recommendations. They are currently re-doing analysis with better classifications so we will see what happens.
Still doesnât fully answer your question tho. To get at the real influence of recommendation you will need to do actual experiments, something only YouTube can really do right now. Or if a third party was allowed to provide a youtube recsys somehow.
My suspicions about radicalization that leads to real word violence is mainly to do with things outside influence of analythims. Disillusionment, Experience of malevolence, Grooming by terrorist ideologically violent religious/âpolitical groups.
Thank you for this post, Mark! I appreciate that you included the graph, though Iâm not sure how to interpret it. Do you mind explaining what the ârecommendation impression advantageâ is? (Iâm sure you explain this in great detail in your paper, so feel free to ignore me or say âgo read the paperâ :D).
The main question that pops out for me is âadvantage relative to what?â I imagine a lot of people would say âeven if YouTubeâs algorithm is less likely to recommend [conspiracy videos/âpropaganda/âfake news] than [traditional media/âvideos about cats], then itâs still a problem! Any amount of recommending [bad stuff that is harmful/âdangerous/âinaccurate] should not be tolerated!â
What would you say to those people?
Recommendation advantage is the ratio of impressions sent vs received. https://ââgithub.com/ââmarkledwich2/âârecfluence#calculations-and-considerations
Yeas, I agree with that. Definitely a lot of room for criticism and different points of view about what should be removed, or sans-recommended. My main effort here is to make sure people know what is happening.
I downvoted this post. Some of our writing guidelines here are to approach disagreements with curiosity as well as trying to be kind. You are clearly deciding against both of these.
I wasnât aware of the writing guidelines. But also, I think this was kind and curious given the context, so I downvoted your comment.
I did not realise you are a new user and probably would have framed my comment differently if I had, I am sorry about that!
To familiarise yourself with our writing guidelines, you can find them on the left bar under âAbout the Forumâ, or just click.
In the past, other users have stated they prefer when people who downvote give explanations for their downvotes. This does seem particularly helpful if you are new and donât know the ins and outs of our forum guidelines and norms yet.
It is great to see you engage with your expertise, and I think it would be a shame if users are put off from engaging with your writing because your content is framed antagonistically.
I read this post before I encountered this comment. I didnât recall seeing anything unkind or uncivil. I then re-read the post to see if I missed anything.
I still havenât been able to find anything problematic. In fact, I notice a few things that I really appreciate from Mark. Some of these include:
Acknowledging explicitly that heâs sometimes rude to his opponents (and explaining why)
Acknowledging certain successes of those he disagrees with (e.g., âIâll give this win to Tristan and Roose.â)
Citing specific actions/âquotes when criticizing others (e.g., the quote from the Joe Rogan podcast)
Acknowledging criticisms of his own work
Overall, I found the piece to be thoughtfully written & in alignment with the community guidelines. Iâm also relatively new to the forum, though, so please point out if Iâm misinterpreting the guidelines.
Iâll also add that I appreciate/âsupport the guideline of âapproaching disagreements with curiosityâ and âaim to explain, not persuade.â But I also think that it would be a mistake to overapply these. In some contexts, it makes sense for a writer to âaim to persuadeâ and approach a disagreement from the standpoint of expertise rather than curiosity.
Like any post, Iâm sure this post could have been written in a way that was more kind/âcurious/âcommunity-normsy. But Iâm struggling to see any areas in which this post falls short. I also think âover-correctingâ could have harms (e.g., causing people to worry excessively about how to phrase things, deterring people from posting, reducing the clarity of posts, making writers feel like they have to pretend to be super curious when theyâre actually trying to persuade).
Denise, do you mind pointing out some parts of the post that violate the writing guidelines? (Itâs not your responsibility, of course, and I fully understand if you donât have time to articulate it. If you do, though, I think Iâd find it helpful & it might help me understand the guidelines better.)
Sure. I am pretty baffled by the response to my comments. I agree the first was insufficiently careful about the fact that Mark is a new user, but even the second got downvotes.
In the past, users of the forum have said many times that posting on the EA Forum gives them anxiety as they are afraid of hostile criticism. So I think it is good to be on the lookout for posts and comments that might have this effect. Being âkindâ and âapproaching disagreements with curiosityâ should protect against this risk. But I ask the question: Is Tristan going to feel comfortable engaging in the Forum, in particular as a response to this post? I donât think so.
Quotes I thought were problematic in that I think they would upset Tristan or put him off responding (or others who might work with him or agree with him):
I read this as Mark proudly announcing that he likes to violate good discourse norms.
Others which I think will make feel Tristan accused and unwelcome (not âkindâ and not âapproaching disagreements with curiosityâ):
Generally hostile:
If the âConflict can be an effective tactic for goodâ section had not been written, I would not have downvoted, as it seems to add little to the content, while making Tristan likely feel very unwelcome.
There was a post which was similar in style to Markâs post arguing against Will here and the response to that was pretty negative, so I am surprised that Markâs post is being perceived so differently.
I only rarely downvote. There have been frequent requests in the past that it would be good if users generally explained why they downvoted. This has not come up before, but I took from that that the next time I downvote, it would be good if I explained why. So I did. And then got heavily downvoted myself for it. I am not sure what to make of thisâare the people requesting for downvoters to generally explain themselves just different people than the ones who downvoted my comment (apparently so, otherwise they would have explained themselves)? Whatever is the reason, I doubt I will explain my downvotes again in the future.
Thank you, Denise! I think this gives me a much better sense of some specific parts of the post that may be problematic. I still donât think this post, on balance, is particularly âbadâ discourse (my judgment might be too affected by what I see on other online discussion platformsâand maybe as I spend more time on the EA forum, Iâll raise my standards!). Nonetheless, your comment helped me see where youâre coming from.
Iâll add that I appreciated that you explained why you downvoted, and it seems like a good norm to me. I think some of the downvotes might just be people who disagree with you. However, I also think some people may be reacting to the way you articulated your explanation. Iâll explain what I mean below:
In the first comment, it seemed to me (and others) like you assumed Mark intentionally violated the norms. You also accused him of being unkind and uncurious without offering additional details.
In the second comment, you linked to the guidelines, but you didnât engage with Markâs claim (âI think this was kind and curious given the context.â). This seemed a bit dismissive to me (akin to when people assume that a genuine disagreement is simply due to a lack of information/âeducation on the part of the person they disagree with).
In the third comment (which I upvoted), you explained some specific parts of the post that you found excessively unkind/âuncivil. This was the first comment where I started to understand why you downvoted this post.
To me, this might explain why your most recent post has received a lot of upvotes. In terms of âwhat to make of this,â I hope you donât conclude âusers should not explain why they downvote.â Rather, I wonder if a conclusion like âusers should explain why they downvote comments, and they should do so in ways that are kind & curious, ideally supported by specific examples when possibleâ would be accurate. Of course, the higher the bar to justify a downvote, the fewer people will do it, and I donât think we should always expect downvote-explainers to write up a thorough essay on why theyâre downvoting.
Finally, Iâll briefly add that upvotes/âdownvotes are useful metrics, but I wouldnât place too much value in them. Iâm guessing that upvotes/âdownvotes often correspond to âdo I agree with this?â rather than âdo I think this is a valuable contribution?â Even if your most recent comment had 99 downvotes, I would still find it helpful and appreciate it!
My reaction was similar to Akashâs.
I wished that the initial comment had been more specific given the userâs status and the tone of the criticism (when I put myself in the authorâs shoes, I could imagine being baffled, since the tone of âmyâ post was relatively tame by the standards of most online discussion spaces).
I downvoted that comment, because I didnât see the explanation as helpful to the author and I want to discourage comments that attack an authorâs motivations without evidence (âyou are clearly deciding against both of theseââI wouldnât call the post âkindâ, but it seemed reasonably curious to me in that it closely engaged with Tristanâs work and acknowledged that he had achieved some of his aims, with plausibly good results).
I thought the third comment was really helpful, and is exactly what I hoped to see from the first comment. I upvoted it. Highlighting specific passages is great; it was also nice to see language like âI read X as the author intending Yâ rather than âby X, the author intended Yâ.
As for the post itself, I chose not to vote, as I was caught between upvoting and downvoting. I also objected to elements of the authorâs tone, but I thought the content was a useful counterpoint to a widely-experienced piece of EA content and provided enough specific arguments for commentators to engage productively.
I expect that people interpreted the âYou are clearly deciding against both of theseâ as an unkind/âuncharitable phrase, since it reads like an accusation of deliberate wrongdoing. I expect that, if youâd instead said something like âParts of your post seem unnecessarily inflammatoryâ, then it wouldnât have received such a negative response.
I also personally tend to interpret the kindness guidelines as being primarily about how to engage with people who are on the forum, or who are likely to read forum posts. Of course we shouldnât be rude in general, but it seems significantly less bad to critique external literature harshly than to directly critique people harshly.
I agree that the kindness guidelines are largely related to community management. I also think they apply more weakly to public figures than to other people who arenât active on the Forum. When someone who has a Netflix special and influence over millions of listeners is making ostensibly bad/âdeceptive arguments, the stakes are higher than usual, and Iâm more likely to think that criticism is valuable enough that even âunkindâ responses are net-valuable.
That said, all of this is contextual; if people began to violate the norm more often, moderation would crack down more to arrest the slide. I havenât seen this happening.
I also really appreciate your comments. I didnât downvote your initial comment, but my first reaction upon seeing it was something like âHey, I felt really positive about a researcher coming to the forum and explaining why he disagrees with Tristan. I donât want someone to discourage this from happening!â Iâve initially read the parts you cited partly as tongue in cheek and maybe as a little unnecessary, but far from wanting to signal that the overall contribution was not welcome.
I appreciate that you explained your negative reaction a lot, especially given how rarely people do it. I did read over the parts you cited not even wondering much how Tristan would react to it and I think itâs great someone brought it up as I now think that new users of our forum should strive to communicate disagreements less confrontationally than is common on other platforms. So I think itâd be unfortunate if you feel discouraged from this experience.
Thank you all for your responses, I really appreciated them. Your perspectives make more sense to me now, though I have to say I still feel really confused.
[Following comment not exhaustively responding to everything you said.]
I hadnât intended to communicate in my first comment that Mark deliberately intended to violate the forum guidelines, but that he deliberately decided against being kind and curious. (Thank you for pointing that out, I did not think of the alternative reading.) I didnât provide any evidence for this because I thought Mark said this very explicitly at the start of his post:
Gentlemanâs rules usually include things like being kind and curious I would guess, and Mark says explicitly that he ignores them because the tradeoff is worth it to him. I donât understand how these lines can be interpreted in any other way, this seems like the literal reading to me.
I have to admit that even after all your kind elaborate explanations I struggle to understand how anything in the section âConflict can be an effective tactic for goodâ could be read as tongue-in-cheek, as it reads very openly hostile to me (...itâs right there in the title?) .
I donât think it is that unlikely that interviewees on the 80k podcast would respond to a kind thoughtful critique on the EA Forum. That said, this is not just about Tristan, but everyone who might disagree with Mark, as the âConflict can be an effective tactic for goodâ section made me doubt they would be treated with curiosity and kindness.
I will take from this that people can have very different interpretations of the same content, even if I think the content is is very explicit and straightforward.