I overall feel I learnt nothing new from the generated answers and could recognize existing inspiration. ChatGPT is valuable at coming up with a bunch of stuff fast, but I’m not impressed by the quality itself.
Specifically in the first 20 examples, I’d say over half of them are mostly false (ie. I would not follow its advice and think there are good reasons to not follow the advice) : 2 3 4 6 7 8 10 11 13 16
Others are uninteresting.
Just 2 I found both mostly true and mostly novel (“If you’ve never had a cold, you’re not exposing yourself to enough germs” and “If you’ve never received a parking ticket, you’re not driving in enough unfamiliar places.”)
I found the Robin Hanson versions mostly uninteresting.
I found the historical accident versions mostly uninteresting.
What problem are you trying to solve by recommending to not date within EA?
If it’s conflicts of interest, it seems like you’ll get more mileage directly promoting norms of avoiding conflict of interest by disclosing what would bias judgement and avoiding being a decision maker in that situation.
As one anecdote, I worked in a typical enough non EA startup in which multiple coworkers had romantic relationships with each other, and multiple coworkers had strong friend relationships with each other. In my experience management decisions were more highly influenced and biased by friend relationships than by romantic relationships. Many companies and institutions have cliques and friend networks that try to gain power together, and I do think it makes sense to have strong norms on disclosing that and reducing those conflicts of interest.
On one hand I agree that avoiding conflicts of interest is important, on the other I think you’re approaching it too narrowly if you focus on romantic/sexual relationships. But I wouldn’t bite the bullet of saying one shouldn’t have friendly or romantic/sexual relationships in the EA community, as that just seems too high a cost to pay.